Abstract

Multi-focus image fusion aims to generate an all-in-focus image from multiple images focused on different regions. The targets of multi-focus image fusion vary with different regions of the image, e.g., flat regions retain smoothness, while edges and textures should be sharpened. However, the existing deep learning-based multi-focus image fusion methods usually treat the image as a whole for fusion and train the model by optimizing homogenous pixel-wise loss (e.g., MSE). This leads to the trained model tends to generate flat regions that are easy to reconstruct, failing to infer realistic details. In this paper, we propose a component divide-and-conquer model for multi-focus image fusion, which uses discrete wavelet transform to decompose source images into low-frequency and high-frequency components, and feeds them into different branches to aggregate them separately by the proposed attention feature fusion network. Finally, the fused image is obtained by inverse discrete wavelet transform. Such a strategy not only can address the challenges of different difficulties in the fusion of low-frequency and high-frequency components, but also able to supervise different components flexibly by the intermediate supervision learning strategy to generate realistic details of the fused image. Extensive experiments show that the proposed component divide-and-conquer model achieves significant improvements in both quantitative and qualitative evaluation.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call