IoT Technologies for HealthCare. 7th EAI International Conference, HealthyIoT 2020, Viana do Castelo, Portugal, December 3, 2020, Proceedings

Research Article

Interpreting the Visual Acuity of the Human Eye with Wearable EEG Device and SSVEP

Download
164 downloads
  • @INPROCEEDINGS{10.1007/978-3-030-69963-5_6,
        author={Danson Evan Garcia and Yi Liu and Kai Wen Zheng and Yi (Summer) Tao and Phillip V. Do and Cayden Pierce and Steve Mann},
        title={Interpreting the Visual Acuity of the Human Eye with Wearable EEG Device and SSVEP},
        proceedings={IoT Technologies for HealthCare. 7th EAI International Conference, HealthyIoT 2020, Viana do Castelo, Portugal, December 3, 2020, Proceedings},
        proceedings_a={HEALTHYIOT},
        year={2021},
        month={7},
        keywords={Wearable sensing Human monitoring Electroencephalography Steady-state visually evoked potentials Quantified self Augmented reality},
        doi={10.1007/978-3-030-69963-5_6}
    }
    
  • Danson Evan Garcia
    Yi Liu
    Kai Wen Zheng
    Yi (Summer) Tao
    Phillip V. Do
    Cayden Pierce
    Steve Mann
    Year: 2021
    Interpreting the Visual Acuity of the Human Eye with Wearable EEG Device and SSVEP
    HEALTHYIOT
    Springer
    DOI: 10.1007/978-3-030-69963-5_6
Danson Evan Garcia1, Yi Liu1, Kai Wen Zheng1, Yi (Summer) Tao1, Phillip V. Do1, Cayden Pierce2, Steve Mann1
  • 1: University of Toronto
  • 2: MannLab Canada

Abstract

Using a wearable electroencephalogram (EEG) device, this paper introduces a novel method of quantifying and understanding the visual acuity of the human eye with the steady-state visually evoked potential (SSVEP) technique. This method gives users easy access to self-track and to monitor their eye health. The study focuses on how varying the SSVEP stimulus frequency and duration affect the overall representation of a person’s visual perception. The study proposes two methods for this visual representation. The first method is a hardware system that utilizes long-exposure photography to augment reality and collocate the visual map onto the plane of interest. The second is a software implementation that captures the visual field at a set distance. A three-dimensional mapping is created by gathering software-defined visual maps at various set distances. Preliminary results show that these methods can gain some insight into the user’s central vision, peripheral vision, and depth perception.