Abstract

Recently, lung cancer prediction based on imaging genomics has attracted great attention. However, such studies often have many challenges, such as small sample size, high-dimensional information redundancy, and the inefficiency of multimodal fusion. Therefore, in this paper, a deep convolution cascade attention fusion network (DCCAFN) based on imaging genomics is proposed for the prediction of lung cancer patients’ survival. The network consists of three modules: an image feature extraction module (IFEM), a gene feature extraction module (GFEM), and an attention fusion network (AFN). In the IFEM, a pretrained residual network based on transfer learning is used to extract deep image features to fully capture the computed tomography (CT) image information conducive to prognosis prediction. In the GFEM, the F-test is first used for gene screening to eliminate redundant information, and then, a cascade network with the convolution cascade module (CCM) that contains a convolution operation, a pooling operation, and an ensemble forest classifier is designed to better extract the gene features. In the AFN, a bimodal attention fusion mechanism is proposed to fuse deep image features and gene features to improve the performance of predicting lung cancer survival. The experimental results show that the DCCAFN model achieves good performance, and its accuracy and AUC are 0.831 and 0.816, respectively. It indicates that the model is an effective multimodal data fusion method for predicting the survival prognosis of lung cancer, which can greatly help physicians stratify patients' risks, and achieve personalized treatment for improving the quality of patients' lives.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call