Respiratory diseases have a significant impact on modern society, posing considerable public health risks. With the emergence of the COVID-19 pandemic, addressing respiratory issues has become increasingly urgent and important. Recently, artificial intelligence-based methods utilizing the acoustic recordings of suspected patients have shown promise in the diagnosis of various respiratory diseases, enabling localized treatment and containment of their spread. As the existing methods cannot efficiently extract subtle differences between the sound samples, thereby limiting their generalization ability. To improve the diagnosis accuracy, this paper proposes a novel multi-channel, multi-modal deep learning architecture based on the attention mechanism. The proposed framework combines a deep convolutional neural network (DCNN) with a bidirectional long short-term memory (BLSTM) network, and also utilizes the attention scheme to extract temporal and spectral features from different modalities of speech data (e.g. coughs, counting sounds and sustained vowel articulation). The proposed method effectively classifies COVID-19 patients, asthma patients and healthy individuals, with a test accuracy of 89.27 ± 0.1%, and an F1 score of 85.42 ± 0.2%. The experimental results validate the feasibility of our method, and also indicate that it is competitive with the existing deep networks.
Read full abstract