红外与可见光图像特征自适应融合方法

Adaptive Fusion Method for Infrared and Visible Light Images

  • 摘要: 多模态图像融合技术旨在为融合后的图像保留不同模态图像的优势,例如纹理细节和突出的目标主体。针对在多模态图像融合领域内存在的对跨模态特征信息提取不充分、跨模态特征建模的复杂性以及不同模态之间共享信息处理困难等问题,我们结合了Transformer和CNN网络的特性,引入两个并行的分支网络,提出了一种新的由相关性驱动的多模态特征分解融合网络(CTIFuss),该结构所提出的融合方法通过两阶段训练策略完成。第一阶段通过建立模态间的相关性模型,充分利用两者的信息特点,促进红外与可见光图像中的跨模态特征和共享特征的提取。在第二阶段中,设计了一种SKNet模块,自适应调整不同模态图像特征的权重分配。在3个公开数据集上的实验结果表明:本文所提出的方法在定量和定性评估方面较其他典型方法有一定优势。

     

    Abstract: Multi-modal image fusion techniques aim to preserve the advantages of different modal images, such as texture details and prominent target subjects, for fused images. This study aimed to address the problems of insufficient extraction of cross-modal feature information, the complexity of cross-modal feature modeling, and the difficulty of processing shared information between different modalities that exist in the field of multi-modal image fusion. The properties of the Transformer and convolutional neural networks are combined, introducing two parallel branching networks. A new correlation-driven multi-modal feature-decomposition fusion network is proposed, accomplished by a two-stage training strategy. The first stage facilitates the extraction of cross-modal and shared features from infrared and visible images by modeling the inter-modal correlation and making full use of their information characteristics. In the second phase, a selective kernel network module was designed to adaptively adjust the weight assignment of different modal image features. Experimental results on three publicly available datasets show that the proposed method has advantages over other typical methods in quantitative and qualitative evaluations.

     

/

返回文章
返回