Abstract

Abstract Aims Aortic elongation can result from age-related changes, congenital factors, aneurysms, or conditions affecting blood vessel elasticity. It is associated with cardiovascular diseases and severe complications like aortic aneurysms and dissection. We assess qualitatively and quantitatively explainable methods to understand the decisions of a deep learning model for detecting aortic elongation using chest X-ray (CXR) images. Methods and results In this study, we evaluated the performance of deep learning models (DenseNet and EfficientNet) for detecting aortic elongation using transfer learning and fine-tuning techniques with CXR images as input. EfficientNet achieved higher accuracy (86.7% ± 2.1), precision (82.7% ± 2.7), specificity (89.4% ± 1.7), F1 score (82.5% ± 2.9), and area under the receiver operating characteristic (92.7% ± 0.6) but lower sensitivity (82.3% ± 3.2) compared with DenseNet. To gain insights into the decision-making process of these models, we employed gradient-weighted class activation mapping and local interpretable model-agnostic explanations explainability methods, which enabled us to identify the expected location of aortic elongation in CXR images. Additionally, we used the pixel-flipping method to quantitatively assess the model interpretations, providing valuable insights into model behaviour. Conclusion Our study presents a comprehensive strategy for analysing CXR images by integrating aortic elongation detection models with explainable artificial intelligence techniques. By enhancing the interpretability and understanding of the models’ decisions, this approach holds promise for aiding clinicians in timely and accurate diagnosis, potentially improving patient outcomes in clinical practice.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call