Magnetic Resonance Imaging is increasing in importance in prostate cancer diagnosis due to the high accuracy and quality of the examination procedure. However, this process requires a time-consuming analysis of the results. Currently, machine vision is widely used in many areas. It enables automation and support in radiological studies. Successful detection of primary prostate tumors depends on the effective segmentation of the prostate itself. At times, a CT scan may be performed; alternatively, MRI may be the selected option. The data always reach a bottleneck stage. This paper presents the effective training of deep learning models to segment the prostate based on onefold and multimodal medical images. This approach supports the computer-aided diagnosis (CAD) system for radiologists as the first step in cancer exams. A comparison of two approaches designed for prostate segmentation is described. The first combines YOLOv4, the object detection neural network, and U-Net for a semantic segmentation based on onefold modality MRI images. The second presents the same method trained on multimodal images—a CT and MRI mixed dataset. The learning process was carried out in a cloud environment using GPU cards. The experiments are based on data from 120 patients who have undergone MRI and CT examinations. Several metrics evaluated the trained models. In the prostate semantic segmentation process, better results were achieved by mixed MRI with CT datasets. The best model achieved the value of 0.9685 for the Sørensen–Dice coefficient for the threshold value of 0.6.
Read full abstract