Purpose:The purpose of this study is to develop an automated method using deep learning for the reliable and precise quantification of left ventricle structure and function from echocardiogram videos, eliminating the need to identify end-systolic and end-diastolic frames. This addresses the variability and potential inaccuracies associated with manual quantification, aiming to improve the diagnosis and management of cardiovascular conditions. Methods:A single, fully automated multitask network, the EchoFused Network (EFNet) is introduced that simultaneously addresses both left ventricle segmentation and ejection fraction estimation tasks through cross-module fusion. Our proposed approach utilizes semi-supervised learning to estimate the ejection fraction from the entire cardiac cycle, yielding more dependable estimations and obviating the need to identify specific frames. To facilitate joint optimization, the losses from task-specific modules are combined using a normalization technique, ensuring commensurability on a comparable scale. Results:The assessment of the proposed model on a publicly available dataset, EchoNet-Dynamic, shows significant performance improvement, achieving an MAE of 4.35% for ejection fraction estimation and DSC values of 0.9309 (end-diastolic) and 0.9135 (end-systolic) for left ventricle segmentation. Conclusions:The study demonstrates the efficacy of EFNet, a multitask deep learning network, in simultaneously quantifying left ventricle structure and function through cross-module fusion on echocardiogram data.
Read full abstract