Traditional identity recognition methods are facing significant security challenges due to their vulnerability to leakage and forgery. Brainprint recognition, a novel biometric identification technology leveraging EEG signals, has emerged as a promising alternative owing to its advantages such as resistance to coercion, non-forgeability, and revocability. Nevertheless, the scarcity of high-quality electroencephalogram (EEG) data limits the performance of brainprint recognition systems, necessitating the use of shallow models that may not perform optimally in real-world scenarios. Data augmentation has been demonstrated as an effective solution to address this issue. However, EEG data encompass diverse features, including temporal, frequency, and spatial components, posing a crucial challenge in preserving these features during augmentation. This paper proposes an end-to-end EEG data augmentation method based on a spatial–temporal generative adversarial network (STGAN) framework. Within the discriminator, a temporal feature encoder and a spatial feature encoder were parallelly devised. These encoders effectively captured global dependencies across channels and time of EEG data, respectively, leveraging a self-attention mechanism. This approach enhances the data generation capabilities of the GAN, thereby improving the quality and diversity of the augmented EEG data. The identity recognition experiments were conducted on the BCI-IV2A dataset, and Fréchet inception distance (FID) was employed to evaluate data quality. The proposed method was validated across three deep learning models: EEGNET, ShallowConvNet, and DeepConvNet. Experimental results indicated that data generated by STGAN outperform DCGAN and RGAN in terms of data quality, and the identity recognition accuracies on the three networks were improved by 2.49%, 2.59% and 1.14%, respectively.
Read full abstract