基于显著性特征的可见光和红外图像融合算法
A Saliency Feature-Based Visible-Infrared Image Fusion Method
DOI: 10.12677/airr.2026.152054, PDF,   
作者: 吴 韩, 王聪聪*:新疆理工职业大学人工智能学院,新疆 喀什;陈兰兰:新疆理工职业大学通识学院,新疆 喀什
关键词: 可见光红外融合视觉显著性特征分解全变分自适应PCNNVisible-Infrared Image Fusion Visual Saliency Feature Decomposition Total Variation Adaptive PCNN
摘要: 针对可见光与红外图像在成像机理与信息表达方面存在差异,以及传统融合方法在低照度条件下易出现结构信息不完整、细节表达不足和视觉质量下降等问题,提出一种基于显著性特征的可见光与红外图像融合算法。该方法通过构建互补特征分解模型,将源图像分离为主体结构特征与细节纹理特征,在主体结构特征中使用联合建模像素灰度显著性与多尺度梯度显著性,并引入全变分约束对融合结果进行优化,以提高结构保持能力和整体视觉一致性;在细节纹理特征中使用参数自适应脉冲耦合神经网络(PAPCNN),根据特征统计特性自动调节网络参数,实现边缘与细节信息的自适应增强。基于LLVIP和M3FD数据集的对比实验结果表明,所提方法在信息熵、特征互信息和视觉信息保真度等评价指标上均优于多种典型融合方法,能够有效提升融合图像的信息表达能力和视觉感知质量。
Abstract: Visible and infrared images exhibit significant differences in imaging mechanisms and information representation. Under low-illumination conditions, traditional fusion methods often suffer from incomplete structural information, insufficient detail preservation, and degraded visual quality. To address these issues, a saliency-based visible and infrared image fusion method is proposed. The proposed method first constructs a complementary feature decomposition model to separate the source images into main structural features and detail texture features. For the structural features, a joint saliency model integrating pixel-intensity saliency and multi-scale gradient saliency is established, and total variation (TV) regularization is introduced to optimize the fusion results, thereby enhancing structural preservation and improving overall visual consistency. For the detail texture features, a parameter-adaptive pulse coupled neural network (PAPCNN) is employed, in which network parameters are automatically adjusted according to the statistical characteristics of the features to achieve adaptive enhancement of edge and fine-detail information. Comparative experiments conducted on the LLVIP and M3FD datasets demonstrate that the proposed method outperforms several representative fusion approaches in terms of information entropy, feature mutual information, and visual information fidelity, effectively improving the information representation capability and perceptual quality of the fused images.
文章引用:吴韩, 王聪聪, 陈兰兰. 基于显著性特征的可见光和红外图像融合算法[J]. 人工智能与机器人研究, 2026, 15(2): 559-569. https://doi.org/10.12677/airr.2026.152054

参考文献

[1] Luo, Y. and Luo, Z. (2023) Infrared and Visible Image Fusion: Methods, Datasets, Applications, and Prospects. Applied Sciences, 13, Article 10891. [Google Scholar] [CrossRef
[2] Ma, W., Wang, K., Li, J., Yang, S.X., Li, J., Song, L., et al. (2023) Infrared and Visible Image Fusion Technology and Application: A Review. Sensors, 23, Article 599. [Google Scholar] [CrossRef] [PubMed]
[3] Wang, Q., Jin, P., Wu, Y., Zhou, L. and Shen, T. (2025) Infrared Image Enhancement: A Review. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 18, 3281-3299. [Google Scholar] [CrossRef
[4] Liu, J., Wu, G., Liu, Z., Wang, D., Jiang, Z., Ma, L., et al. (2025) Infrared and Visible Image Fusion: From Data Compatibility to Task Adaption. IEEE Transactions on Pattern Analysis and Machine Intelligence, 47, 2349-2369. [Google Scholar] [CrossRef] [PubMed]
[5] Zou, D. and Yang, B. (2023) Infrared and Low-Light Visible Image Fusion Based on Hybrid Multiscale Decomposition and Adaptive Light Adjustment. Optics and Lasers in Engineering, 160, Article 107268. [Google Scholar] [CrossRef
[6] Hu, P., Wang, C., Li, D. and Zhao, X. (2023) An Improved Hybrid Multiscale Fusion Algorithm Based on NSST for Infrared-Visible Images. The Visual Computer, 40, 1245-1259. [Google Scholar] [CrossRef
[7] Wang, W., Zhang, J., Liu, H., Xiong, W. and Zhang, C. (2023) Joint Low-Rank and Sparse Decomposition for Infrared and Visible Image Sequence Fusion. Infrared Physics & Technology, 133, Article 104828. [Google Scholar] [CrossRef
[8] Zhang, X. and Demiris, Y. (2023) Visible and Infrared Image Fusion Using Deep Learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45, 10535-10554. [Google Scholar] [CrossRef] [PubMed]
[9] Liu, J., Dian, R., Li, S. and Liu, H. (2023) SGFusion: A Saliency Guided Deep-Learning Framework for Pixel-Level Image Fusion. Information Fusion, 91, 205-214. [Google Scholar] [CrossRef
[10] Gu, X., Xia, Y. and Zhang, J. (2024) Multimodal Medical Image Fusion Based on Interval Gradients and Convolutional Neural Networks. BMC Medical Imaging, 24, Article No. 232. [Google Scholar] [CrossRef] [PubMed]
[11] Ibrahim, S.I., El-Tawel, G.S. and Makhlouf, M.A. (2024) Brain Image Fusion Using the Parameter Adaptive-Pulse Coupled Neural Network (PA-PCNN) and Non-Subsampled Contourlet Transform (NSCT). Multimedia Tools and Applications, 83, 27379-27409. [Google Scholar] [CrossRef
[12] Yuan, D., Zhang, H., Shu, X., Liu, Q., Chang, X., He, Z., et al. (2024) Thermal Infrared Target Tracking: A Comprehensive Review. IEEE Transactions on Instrumentation and Measurement, 73, 1-19. [Google Scholar] [CrossRef
[13] Wang, H., Lou, J., Liu, X., Tan, H., Whitaker, R. and Liu, H. (2024) SSPNet: Predicting Visual Saliency Shifts. IEEE Transactions on Multimedia, 26, 4938-4949. [Google Scholar] [CrossRef
[14] Zhao, J., Zhang, T., Fang, S., Gao, J., Wang, J. and Gong, M. (2026) Spatial-Spectral Texture-Preserved Total Variation: A Novel Regularization for Hyperspectral Image Denoising. IEEE Transactions on Circuits and Systems for Video Technology, 36, 248-260. [Google Scholar] [CrossRef
[15] Gupta, P. and Jain, N. (2024) Anisotropic Diffusion Filter Based Fusion of NSST Transformed Medical Images. Biomedical Signal Processing and Control, 90, Article 105819. [Google Scholar] [CrossRef