Driving vigilance estimation is an important task for traffic safety. Nowadays, electroencephalography (EEG) and electrooculography (EOG) have made some achievements in vigilance estimation, but there are still some challenges: 1) The traditional approachs with direct multimodal fusion may face the problems of information redundancy and data dimensionality mismatch; 2) Capture key discriminative features during multimodal fusion without losing specific patterns to each modality. In order to solve the above problems, this paper proposes a approach with fusion of EEG and EOG features in split bands, which not only preserves the information about brain activities in different bands of EEG, but also effectively integrates the relevant information of EOG. On this basis, we further propose a hierarchical multi-scale topological enhanced network (HMS-TENet). This network first introduces a pyramid pooling structure (PPS) to capture contextual relationships from different discriminative perspectives. And then we design a selective convolutional structure (SCS) for adaptive sense-field selection, which enables us to mine the desired discriminative information in small-size features. In addition, we design a topology self-aware attention to enhance the learning of representations of complex topological relationships among EEG channels. Finally, the output of the model can be selected for both regression and classification tasks, providing higher flexibility and adaptability. We demonstrate the robustness, generalizability, and utility of the proposed method based on intra-subject and cross-subject experiments on the SEED-VIG public dataset. Codes are available at https://github.com/tangmeng28/HMS-TENet.
Read full abstract