Abstract
Medical image fusion, which aims to derive complementary information from multi-modality medical images, plays an important role in many clinical applications, such as medical diagnostics and treatment. We propose the LatLRR-FCNs, which is a hybrid medical image fusion framework consisting of the latent low-rank representation (LatLRR) and the fully convolutional networks (FCNs). Specifically, the LatLRR module is used to decompose the multi-modality medical images into low-rank and saliency components, which can provide fine-grained details and preserve energies, respectively. The FCN module aims to preserve both global and local information by generating the weighting maps for each modality image. The final weighting map is obtained using the weighted local energy and the weighted sum of the eight-neighborhood-based modified Laplacian method. The fused low-rank component is generated by combining the low-rank components of each modality image according to the guidance provided by the final weighting map within pyramid-based fusion. A simple sum strategy is used for the saliency components. The usefulness and efficiency of the proposed framework are thoroughly evaluated on four medical image fusion tasks, including computed tomography (CT) and magnetic resonance (MR), T1- and T2-weighted MR, positron emission tomography and MR, and single-photon emission CT and MR. The results demonstrate that by leveraging the LatLRR for image detail extraction and the FCNs for global and local information description, we can achieve performance superior to the state-of-the-art methods in terms of both objective assessment and visual quality in some cases. Furthermore, our method has a competitive performance in terms of computational costs compared to other baselines.
Highlights
Medical image fusion is a key technology that has been used extensively in clinical diagnosis and treatment planning (James and Dasarathy, 2014)
This section is devoted to showing that the proposed LatRRFCNs can improve the information details and energy preservation in terms of visual quality assessment, quantitative assessment and computational cost assessment, compared with five recently proposed methods: adaptive sparse representation (ASR) (Liu and Wang, 2014), LP-convolutional neural network (CNN) (Liu et al, 2017), non-subsampled contourlet transform (NSCT)-PC-LLE (Zhu et al, 2019), non-subsampled shearlet transform (NSST)-PAPCNN (Yin et al, 2018), and convolutional sparsity-based morphological component analysis (CSMCA) (Liu et al, 2019)
The usefulness and efficiency of each method are investigated with four sets of medical image fusion studies, including computed tomography (CT) and magnetic resonance (MR), MR-T1 and MR-T2, positron emission tomography (PET) and MR, and single-photon emission computed tomography (SPECT) and MR
Summary
Medical image fusion is a key technology that has been used extensively in clinical diagnosis and treatment planning (James and Dasarathy, 2014). To improve the robustness of activity level measurement and weight assignment, Liu et al (2017) introduced a deep learning fusion method with a simple multi-layer convolutional neural network (CNN) using the decision map and the medical image under the pyramid-based image fusion framework to reconstruct the fused medical image. While such a method achieves some success in specific medical image fusion tasks, this work may fail in multi-modal image fusion because the simple use of the CNN cannot extract finegrained details efficiently.
Published Version (Free)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have