Systems Engineering and Electronics ›› 2024, Vol. 46 ›› Issue (8): 2641-2649.doi: 10.12305/j.issn.1001-506X.2024.08.12

• Sensors and Signal Processing • Previous Articles    

Infrared and visible light image fusion based on convolution and self attention

Xiaoxuan CHEN1, Shuwen XU2, Shaohai HU1,*, Xiaole MA1   

  1. 1. School of Computer and Information Technology, Beijing Jiaotong University, Beijing 100044, China
    2. Research Institute of TV and Electro-Acoustics, China Electronics Technology Group Corporation, Beijing 100015, China
  • Received:2023-05-29 Online:2024-07-25 Published:2024-08-07
  • Contact: Shaohai HU

Abstract:

As convolution operation pays too much attention to local features of an image, which easily cause the loss of the global semantic information of the fused image when fusing source images. To solve this problem, an infrared and visible light image fusion model based on convolution and self attention is proposed in this paper. In the proposed model, convolution module is adopted to extract local features of image, and self attention is adopted to extract global features. In addition, since the simple operation cannot handle the fusion of features at different levels, the embedded block residual fusion module is proposed to realize the multi-layer feature fusion. Experimental results demonstrate that the proposed method has superiority over the unsupervised deep fusion algorithms in both subjective evaluation and six objective metrics, among which the mutual information, standard deviation, and visual fidelity are improved by 61.33%, 9.96%, and 19.46%, respectively.

Key words: image fusion, global features, self attention, auto-encoder, deep learning

CLC Number: 

[an error occurred while processing this directive]