Abstract
Breast cancer is a significant global health concern, highlighting the critical importance of early detection for effective treatment of women’s health. While convolutional networks (CNNs) have been the best for analysing medical images, recent interest has emerged in leveraging vision transformers (ViTs) for medical data analysis. This study aimed to conduct a comprehensive comparison of three systems a self-attention transformer (VIT), a compact convolution transformer (CCT), and a tokenlearner (TVIT) for binary classification of mammography images into benign and cancerous tissue. Thorough experiments were performed using the DDSM dataset, which consists of 5970 benign and 7158 malignant images. The performance accuracy of the proposed models was evaluated, yielding results of 99.81% for VIT, 99.92% for CCT, and 99.05% for TVIT. Additionally, the study compared these results with the current state-of-the-art performance metrics. The findings demonstrate how convolution-attention mechanisms can effectively contribute to the development of robust computer-aided systems for diagnosing breast cancer. Notably, the proposed approach achieves high-performance results while also minimizing the computational resources required and reducing decision time.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: International Journal of Hybrid Intelligent Systems
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.