Abstract

Breast cancer is the leading cause of mortality in women. Early diagnosis of breast cancer can reduce the mortality rate. In the diagnosis, the mitotic cell count is an important biomarker for predicting the aggressiveness, prognosis, and grade of breast cancer. In general, pathologists manually examine histopathology images under high-resolution microscopes for the detection of mitotic cells. However, because of the minute differences between the mitotic and normal cells, this process is tiresome, time-consuming, and subjective. To overcome these challenges, artificial-intelligence-based (AI-based) techniques have been developed which automatically detect mitotic cells in the histopathology images. Such AI techniques accelerate the diagnosis and can be used as a second-opinion system for a medical doctor. Previously, conventional image-processing techniques were used for the detection of mitotic cells, which have low accuracy and high computational cost. Therefore, a number of deep-learning techniques that demonstrate outstanding performance and low computational cost were recently developed; however, they still require improvement in terms of accuracy and reliability. Therefore, we present a multistage mitotic-cell-detection method based on Faster region convolutional neural network (Faster R-CNN) and deep CNNs. Two open datasets (international conference on pattern recognition (ICPR) 2012 and ICPR 2014 (MITOS-ATYPIA-14)) of breast cancer histopathology were used in our experiments. The experimental results showed that our method achieves the state-of-the-art results of 0.876 precision, 0.841 recall, and 0.858 F1-measure for the ICPR 2012 dataset, and 0.848 precision, 0.583 recall, and 0.691 F1-measure for the ICPR 2014 dataset, which were higher than those obtained using previous methods. Moreover, we tested the generalization capability of our technique by testing on the tumor proliferation assessment challenge 2016 (TUPAC16) dataset and found that our technique also performs well in a cross-dataset experiment which proved the generalization capability of our proposed technique.

Highlights

  • Breast cancer is the most common and leading cause of death among women

  • Score-level fusion of these networks followed by classification has a huge impact and it can be seen that our proposed technique outperformed other state-of-the-art techniques for mitotic cell detection after combination of Faster R-convolutional neural network (CNN), post-processing and score level fusion of deep networks

  • We presented a multi-stage mitotic-cell-detection technique based on Faster-RCNN, post-processing, and deep CNNs

Read more

Summary

Introduction

Breast cancer is the most common and leading cause of death among women. According to the global cancer project (GLOBOCAN 2012), breast cancer accounts for 25.1% of all cancers in women [1].Early diagnosis of breast cancer is an important factor for the reduction of the mortality rate because its treatment plan is advised on the basis of the grade and prognosis of the cancer. Breast cancer is the most common and leading cause of death among women. According to the global cancer project (GLOBOCAN 2012), breast cancer accounts for 25.1% of all cancers in women [1]. Diagnosis of breast cancer is an important factor for the reduction of the mortality rate because its treatment plan is advised on the basis of the grade and prognosis of the cancer. To determine the grade of breast cancer, the Nottingham grading system has been widely used. According to this system, there are three biomarkers for the grading of breast cancer in histopathology images. There are three biomarkers for the grading of breast cancer in histopathology images

Methods
Findings
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.