Abstract

Automatic segmentation of medical images is a difficult task in the field of computer vision owing to the various backgrounds, shapes, size, and colors of polyps or tumors. Despite the success of deep learning (DL)-based encoder–decoder architectures in medical image segmentation, these models have several disadvantages. First, an architecture such as U-Net cannot encode multi-scale semantic information at a different level on the decoder side. Second, it fails to reimpose the feature maps adeptly due to its limited capability on capturing long-range feature dependencies. In this study, we solve this problem by capturing multi-scale global feature maps, which forces the network to learn different semantic information at each scale. Further, we utilize the attention mechanism to suppress noise and the undesirable features, leading to a thorough restoration of contextual feature dependencies. Finally, we propose a novel method which leverages the compound scaled EfficientNet as a encoder backbone for efficient feature extraction and the UNet decoder to reconstruct the fine-grained details. We evaluated the proposed method using three different medical datasets: Kvasir-SEG, nuclei segmentation, and skin-lesion segmentation. The experimental results demonstrate that the proposed method takes an unassailable lead in terms of segmentation accuracy over the baseline models across different datasets and backbone architectures. Further, the proposed method strengthens the segmentation quality of varying shapes, object shapes, suppresses the noise, and leads to a better performance.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.