Abstract

Abstract Introduction and Objective: Prostate cancer (PC) with de novo bone metastases at diagnosis (M1) carries a 5-year survival rate of 28% and requires early, aggressive treatment. Clinical assays and the pathology of prostate needle biopsies (PNBX) cannot distinguish primary M1 tumors from high-grade localized (M0) cases. We hypothesized that digital image analysis can be applied to obtain morphologic biomarkers, not recognizable by pathologists. Here we demonstrate how novel software tools that involve deep learning frameworks can be used to systematically extract handcrafted and autoencoder features and to build models to predict M1 stage at the time of diagnosis. Methods: A study cohort, nested within a biorepository of 2150 PC patients at the Greater LA VA, consisted of 86 high-grade M0 and 85 M1 cases. Slides were digitized at 40X and 2 pathologists annotated all cancer foci. Approximately 30 image tiles were selected from each case. 62 handcrafted (HC) and 64 autoencoder (AE) features were extracted from nuclei. Feature values normalized. The normalized profile of each primary feature gave rise to 11 secondary features, representing the distribution of the feature within a case. We separated cases into training + testing versus validation groups at an 80:20 ratio. Using a bootstrapping method, we selected the best GLMNET models predicting M0 versus M1 status in the training + testing set and applied them to an independent validation set of cases. Results: After successful conversion of M0 and M1 image tiles to digital nuclear masks and color normalization, ~ 400,000 nuclei were isolated using parameters that enriched for nuclei from cancer cells. A denoising autoencoding neural network was used to generate AE biomarkers for each nucleus. HC features quantified nuclear shape, size, color, and texture. A systematic pipeline of preprocessing, normalization, and conversion to case-level secondary features was applied to AE and HC features. For both feature types, the average of 50,000 bootstrapping models resulted in an AUC of 0.8 for the training and an average accuracy of 0.7 for the test cohort. The best 38 AE models or 16 HC models were applied to the independent validation cohort of 24 cases and assigned each case to the M0 versus M1 groups by majority voting. At a threshold of 0.5, this resulted in an accuracy of 70% for M0 versus M1 distinction. Conclusion: We applied digital imaging technology and machine learning software to AE and HC features in order to predict M0 versus M1 stage from the tumor in PNBXs at diagnosis. Unexpectedly, hidden features in nuclei differed between M0 and M1 cases and succeeded in predicting metastatic disease with 70% accuracy. The ultimate goal is to combine the inexpensive risk prediction from quantitative imaging with clinical parameters and RNA sequencing data to develop accurate prediction models of occult metastases and risk of future metastatic progression at the time of diagnosis in all patients with high-grade PC. Funding: DOD PC131996, PCF-Movember GAP1 Unique TMAs Project, Prostate Cancer Foundation (PCF) Creativity Award, Jean Perkins Foundation, NIH/NCI P01 CA098912-09, NIH R01CA131255 and P50CA092131, Stephen Spielberg Team Science Award. Citation Format: Fangjin Huang, Nathan Ing, Miller Eric, Hootan Salemi, Michael Lewis, Isla Garraway, Arkadiusz Gertych, Beatrice Knudsen. Quantitative digital image analysis and machine learning for staging of prostate cancer at diagnosis [abstract]. In: Proceedings of the AACR Special Conference: Prostate Cancer: Advances in Basic, Translational, and Clinical Research; 2017 Dec 2-5; Orlando, Florida. Philadelphia (PA): AACR; Cancer Res 2018;78(16 Suppl):Abstract nr B094.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.