Abstract
Localization of sleep apnea and hypopnea (SAH) events has routinely relied on expert visual inspection of polysomnography (PSG) recordings, which is a tedious task demanding a high level of professional skills. Automated detection methods have achieved remarkable success, especially with the recent advances in machine learning and deep learning technologies. However, a significant challenge still remains in methods towards clinical practices: How to accurately discriminate SAH events in PSG, with the onset and duration of each?This study develops an object detection framework for accurately identifying the position of SAH segments with varied durations (namely SAH-MOD) in three phases: (1) Dual-modal Feature Learning (DFL, dual-branch 1-D convolutional layers followed by Concatenate Block). Deep features are efficiently learned and then fused from two different types of signals related to respiratory, i.e., nasal airflow and abdominal movement; (2) Feature Map Generation (FMG, cascade 1-D convolutional layers). Feature maps are generated with multi-scale hierarchical features in different depths of the network, catering for the needs of object (SAH event) detection; Default anchors associated with the scales and receptive fields are tiled onto the corresponding detection feature maps; and (3) Multi-scale Object Detection (MOD). A variety of instances of prediction are then made on all available detection layers with post-processing to accurately capture each SAH event.Experiments have been performed on the dataset of stroke unit recordings for the detection of Obstructive Sleep Apnea Syndrome (OSASUD-dataset) with SAH-MOD against the state-of-the-art counterparts, and results indicate that: (1) SAH-MOD performs the best with a Recall of 81.0% and an F1-score of 71.1%; and (2) it has significant advantages in localizing the onset and duration of each SAH event, with 91.9% of the IoU values between predicted and labeled events falling between 0.6 and 1.0. Ablation experiments show that the introduction of dual-modal feature learning and hierarchical feature maps improves recall performance by 6.9% and 4.1%, respectively.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.