Abstract
Studies have shown that the type of low-grade glioma is associated with its shape. The traditional diagnostic method involves extraction of the tumor shape from MRIs and diagnosing the type of glioma based on corresponding relationship between the glioma shape and type. This method is affected by the MRI background, tumor pixel size, and doctors' professional level, leading to misdiagnoses and missed diagnoses. With the help of deep learning algorithms, the shape of a glioma can be automatically segmented, thereby assisting doctors to focus more on the diagnosis of glioma and improving diagnostic efficiency. The segmentation of glioma MRIs using traditional deep learning algorithms exhibits limited accuracy, thereby impeding the effectiveness of assisting doctors in the diagnosis. The primary objective of our research is to facilitate the segmentation of low-grade glioma MRIs for medical practitioners through the utilization of deep learning algorithms. In this study, a UNet glioma segmentation network that incorporates multiattention gates was proposed to address this limitation. The UNet-based algorithm in the coding part integrated the attention gate into the hierarchical structure of the network to suppress the features of irrelevant regions and reduce the feature redundancy. In the decoding part, by adding attention gates in the fusion process of low- and high-level features, important feature information was highlighted, model parameters were reduced, and model sensitivity and accuracy were improved. The network model performed image segmentation on the glioma MRI dataset, and the accuracy and average intersection ratio (mIoU) of the algorithm segmentation reached 99.7%, 87.3%, 99.7%, and 87.6%. Compared with the UNet, PSPNet, and Attention UNet network models, this network model has obvious advantages in accuracy, mIoU, and loss convergence. It can serve as a standard for assisting doctors in diagnosis.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.