Abstract

For the lack of labels in infrared and visible image fusion network, an infrared and visible image fusion model based on multi-channel unsupervised convolutional neural network (CNN) is proposed in this paper, in order to extract more detailed information through multi-channel inputs. In contrast to conventional unsupervised fusion network, the proposed network contains three channels for extracting infrared features, visible features and common features of infrared and visible images, respectively. The square loss function is used to train the network. Pairs of infrared and visible images are input to DenseNet to extract as more useful features as possible. A fusion module is designed to fuse the extracted features for testing. Experimental results show that the proposed method can preserve both the clear target of infrared and detailed information of visible images simultaneously. Experiments also demonstrate the superiority of the proposed method over the state-of-the-art methods in objective metrics.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.