Abstract

Image fusion operation is beneficial to many applications and is also one of the most common and critical computer vision challenges. The perfect infrared and visible image fusion results should include the important infrared targets while preserving visible textural detail information as much as possible. A novel infrared and visible image fusion framework is proposed for this purpose. In this paper, the proposed fusion network (MIFFuse) is an end-to-end, multi-level-based fusion network for infrared and visible images. The presented approach makes effective use of the intermediate convolution layer’s output features to preserve the primary image fusion information. We also build a cat_block to swap information between two paths to gain more sufficient information during the convolution steps. To reduce the model’s running time even further, the proposed method that reduces the number of feature channels while maintaining the accuracy of the fusion performance. Extensive experiments on the TNO and CVC-14 image fusion datasets show that our MIFFuse outperforms the other methods in terms of both subjective visual effects and quantitative metrics. Furthermore, MIFFuse is approximately twice as fast as the most recent state-of-the-art methods. Our code and models can be found at https://github.com/depeng6/MIFFuse .

Highlights

  • M ORE information about a target could not be obtained from a single sensor

  • The task of image fusion is to fuse multi-source information from multiple images into one image, which is convenient for people to view and postprocess [1]

  • The uses of image fusion are mainly divided into four categories, including medical image fusion [2], [3],multi-focus image fusion [4]–[6], remote sensing image fusion [7], [8], infrared and visual image fusion [9], [10], which are examples of image fusion

Read more

Summary

Introduction

M ORE information about a target could not be obtained from a single sensor. The task of image fusion is to fuse multi-source information from multiple images into one image, which is convenient for people to view and postprocess [1]. The far more common image fusion scenario is infrared and visible image fusion [11]. In terms of target detection and surveillance camera tracking, infrared and visible image fusion technology is widely used. Thermal radiation released from surfaces is captured in infrared images, which can illuminate targets but lack texture information. On the other hand, usually provide a lot of structural detail but are influenced by the background and lose goals. The fusion method’s key emphasis shifts to how to efficiently merge complementary information

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call