Abstract

Oral potentially malignant disorders (OPMDs) are precursors to over 80% of oral cancers. Hematoxylin and eosin (H&E) staining, followed by pathologist interpretation of tissue and cellular morphology, is the current gold standard for diagnosis. However, this method is qualitative, can result in errors during the multi-step diagnostic process, and results may have significant inter-observer variability. Chemical imaging (CI) offers a promising alternative, wherein label-free imaging is used to record both the morphology and the composition of tissue and artificial intelligence (AI) is used to objectively assign histologic information. Here, we employ quantum cascade laser (QCL)-based discrete frequency infrared (DFIR) chemical imaging to record data from oral tissues. In this proof-of-concept study, we focused on achieving tissue segmentation into three classes (connective tissue, dysplastic epithelium, and normal epithelium) using a convolutional neural network (CNN) applied to three bands of label-free DFIR data with paired darkfield visible imaging. Using pathologist-annotated H&E images as the ground truth, we demonstrate results that are 94.5% accurate with the ground truth using combined information from IR and darkfield microscopy in a deep learning framework. This chemical-imaging-based workflow for OPMD classification has the potential to enhance the efficiency and accuracy of clinical oral precancer diagnosis.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.