Abstract

In steady-state visual evoked potential (SSVEP)based brain-computer interfaces (BCIs), various spatial filtering methods based on individual calibration data have been proposed to alleviate the interference of spontaneous activities in SSVEP signals for enhancing the SSVEP detection performance. However, the necessary calibration procedures take time, cause visual fatigue and reduce usability. For the calibration-free scenario, we propose a cross-subject frequency identification method based on transfer superimposed theory for SSVEP frequency decoding. First, a multi-channel signal decomposition model was constructed. Next, we used the cross least squares iterative method to create individual specific transfer spatial filters as well as source subject transfer superposition templates in the source subject. Then, we identified common knowledge among source subjects using a prototype spatial filter to make common transfer spatial filters and common impulse responses. Following, we reconstructed a global transfer superimposition template with SSVEP frequency characteristics. Finally, an ensemble cross-subject transfer learning method was proposed for SSVEP frequency recognition by combining the sourcesubject transfer mode, the global transfer mode, and the sinecosine reference template. Offline tests on two public datasets show that the proposed method significantly outperforms the FBCCA, TTCCA, and CSSFT methods. More importantly, the proposed method can be directly used in online SSVEP recognition without calibration. The proposed algorithm was robust, which is important for a practical BCI.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.