This paper explores the classification of gaze direction using electrooculography (EOG) signals, integrating signal processing, deep learning, and ensemble learning techniques to enhance accuracy and reliability. A complex technique is proposed in which several feature types are derived from EOG data. Spectral properties generated from power spectral density analysis augment basic statistical characteristics such as mean and standard deviation, revealing the frequency content of the signal. Skewness, kurtosis, and cross-channel correlations are also used to represent intricate nonlinear dynamics and inter-channel interactions. These characteristics are then reformatted into a two-dimensional array imitating picture data, enabling the use of the pre-trained ResNet50 model to extract deep and high-level characteristics. Using these deep features, an ensemble of bagging-trained decision trees classifies gaze directions, lowering model variance and increasing prediction accuracy. The results show that the ensemble deep learning model obtained outstanding performance metrics, with accuracy and sensitivity ratings exceeding 97% and F1-score of 98%. These results not only confirm the effectiveness of the proposed approach in managing challenging EOG signal classification tasks but also imply important consequences for the improvement of Human-Computer Interaction (HCI) systems, especially in assistive technologies where accurate gaze tracking is fundamental.
Read full abstract