A novel brain-computer interface based on audio-assisted visual evoked EEG and spatial-temporal attention CNN
Objective. Brain-computer interface (BCI) can translate intentions directly into instructions and greatly improve the interaction experience for disabled people or some specific interactive applications. To improve the efficiency of BCI, the objective of this study is to explore the feasibility of a...
Saved in:
Published in | Frontiers in neurorobotics Vol. 16; p. 995552 |
---|---|
Main Authors | , , , , |
Format | Journal Article |
Language | English |
Published |
Lausanne
Frontiers Research Foundation
30.09.2022
Frontiers Media S.A |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Objective. Brain-computer interface (BCI) can translate intentions directly into instructions and greatly improve the interaction experience for disabled people or some specific interactive applications. To improve the efficiency of BCI, the objective of this study is to explore the feasibility of an audio-assisted visual BCI speller and a deep learning-based single-trial event related potentials (ERP) decoding strategy. Approach. In this study, a two-stage BCI speller combining with the motion-onset visual evoked potential (mVEP) and semantically congruent audio evoked ERP was designed to output the target characters. In the first stage, the different group of characters were presented in the different locations of visual field simultaneously and the stimuli were coded to the mVEP based on a new space division multiple access scheme. And then, the target character can be output based on the audio-assisted mVEP in the second stage. Meanwhile, a spatial-temporal attention-based convolutional neural network (STA-CNN) was proposed to recognize the single-trial ERP components. The CNN can learn 2-dimentional features including the spatial information of different activated channels and time dependence among ERP components. In addition, the STA mechanism can enhance the discriminative event-related features by adaptively learning probability weights. Main results. The performance of the proposed two-stage audio-assisted visual BCI paradigm and STA-CNN model was evaluated using the Electroencephalogram (EEG) recorded from 10 subjects. The average classification accuracy of proposed STA-CNN can reach 59.6% and 77.7% for the first stage and the second stage, which were always significantly higher than those of the comparison methods (p<0.05). Significance. The proposed two-stage audio-assisted visual paradigm showed a great potential to be used to BCI speller. Moreover, through the analysis of the attention weights from time sequence and spatial topographies, it was proved that STA-CNN could effectively extract interpretable spatiotemporal EEG features. |
---|---|
Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 content type line 23 Edited by: Ke Liu, Chongqing University of Posts and Telecommunications, China Reviewed by: Long Chen, Tianjin University, China; Jiahui Pan, South China Normal University, China |
ISSN: | 1662-5218 1662-5218 |
DOI: | 10.3389/fnbot.2022.995552 |