Abstract

AbstractClinical decision of oncology comes from multi-modal information, such as morphological information from histopathology and molecular profiles from genomics. Most of the existing multi-modal learning models achieve better performance than single-modal models. However, these multi-modal models only focus on the interactive information between modalities, which ignore the internal relationship between multiple tasks. Both survival analysis task and tumor grading task can provide reliable information for pathologists in the diagnosis and prognosis of cancer. In this work, we present a Multi-modal and Multi-task Fusion (\(\mathrm {M^{2}F}\)) model to make use of the potential connection between modalities and tasks. The co-attention module in multi-modal transformer extractor can excavate the intrinsic information between modalities more effectively than the original fusion methods. Joint training of tumor grading branch and survival analysis branch, instead of separating them, can make full use of the complementary information between tasks to improve the performance of the model. We validate our \(\mathrm {M^{2}F}\) model on glioma datasets from the Cancer Genome Atlas (TCGA). Experiment results show our \(\mathrm {M^{2}F}\) model is superior to existing multi-modal models, which proves the effectiveness of our model.KeywordsMulti-modal learningMulti-taskSurvival analysisTumor grading

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call