Abstract

Respiratory diseases have a significant impact on modern society, posing considerable public health risks. With the emergence of the COVID-19 pandemic, addressing respiratory issues has become increasingly urgent and important. Recently, artificial intelligence-based methods utilizing the acoustic recordings of suspected patients have shown promise in the diagnosis of various respiratory diseases, enabling localized treatment and containment of their spread. As the existing methods cannot efficiently extract subtle differences between the sound samples, thereby limiting their generalization ability. To improve the diagnosis accuracy, this paper proposes a novel multi-channel, multi-modal deep learning architecture based on the attention mechanism. The proposed framework combines a deep convolutional neural network (DCNN) with a bidirectional long short-term memory (BLSTM) network, and also utilizes the attention scheme to extract temporal and spectral features from different modalities of speech data (e.g. coughs, counting sounds and sustained vowel articulation). The proposed method effectively classifies COVID-19 patients, asthma patients and healthy individuals, with a test accuracy of 89.27 ± 0.1%, and an F1 score of 85.42 ± 0.2%. The experimental results validate the feasibility of our method, and also indicate that it is competitive with the existing deep networks.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.