| Literature DB >> 34802710 |
Xincan Zhao1, Yulin Du2, Rui Zhang3.
Abstract
Because an augmented-reality-based brain-computer interface (AR-BCI) is easily disturbed by external factors, the traditional electroencephalograph (EEG) classification algorithms fail to meet the real-time processing requirements with a large number of stimulus targets or in a real environment. We propose a multi-target fast classification method for augmented-reality-based steady-state visual evoked potential (AR-SSVEP), using a convolutional neural network (CNN). To explore the availability and accuracy of high-efficiency multi-target classification methods in AR-SSVEP with a short stimulation duration, a similar stimulus layout was used for a computer screen (PC) and an optical see-through head-mounted display (OST-HMD) device (HoloLens). The experiment included nine flicker stimuli of different frequencies, and a multi-target fast classification method based on a CNN was constructed to complete nine classification tasks, for which the average accuracy of AR-BCI in our CNN model at 0.5- and 1-s stimulus duration was 67.93% and 80.83%, respectively. These results verified the efficacy of the proposed model for processing multi-target classification in AR-BCI.Entities:
Keywords: Augmented reality; Brain–computer interfaces; Convolutional neural network; Steady-state visual evoked potentials
Mesh:
Year: 2021 PMID: 34802710 DOI: 10.1016/j.compbiomed.2021.105042
Source DB: PubMed Journal: Comput Biol Med ISSN: 0010-4825 Impact factor: 4.589