Laser Journal, Volume. 45, Issue 12, 106(2024)

Infrared and visible image fusion based on shuffle attention mechanism and residual dense network

LIU Peipei... ZHANG Yuxiao*, YUAN Shuozhi, WANG Shuo and XU Huyang |Show fewer author(s)
Author Affiliations
  • Chengdu University of Technology, Chengdu 610059, China
  • show less
    References(24)

    [1] [1] Chen J, Li X, Luo L, et al. Multi-focus image fusion based on multi-scale gradients and image matting[J]. IEEE Transactions on Multimedia, 2021, 24: 655-667.

    [2] [2] Zhang Q, Fu Y, Li H, et al. Dictionary learning method for joint sparse representation-based image fusion[J]. Optical Engineering, 2013, 52(5): 7006.

    [3] [3] Kumar B K S. Image fusion based on pixel significance using cross bilateral filter[J]. Signal, image and video processing, 2015, 9(5): 1193-1204.

    [4] [4] Li H, Wu X. J, J Kittler. Infrared and Visible Image Fusion using a Deep Learning Framework[C]//2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 2018: 2705-2710.

    [5] [5] Li H, Wu X, Durrani T S. Infrared and visible image fusion with ResNet and zero-phase component analysis[J]. Infrared Physics & Technology, 2019, 102: 103039.

    [6] [6] Xu H, Zhang H, Ma J Y. Classification Saliency-Based Rule for Visible and Infrared Image Fusion[J]. IEEE Transactions on Computational Imaging, 2021, 7: 824-836.

    [7] [7] Prabhakar K R, Srikar V S, Babu R V. DeepFuse: A Deep Unsupervised Approach for Exposure Fusion with Extreme Exposure Image Pairs[C]//2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy: 2017: 4724-4732.

    [8] [8] Li J, Huo H T, Li C, et al. AttentionFGAN: Infrared and Visible Image Fusion using Attention-based Generative Adversarial Networks[J]. IEEE Transactions on Multimedia, 2021, 23: 1383-1396.

    [9] [9] Ma J, Zhang H, Shao Z, et al. GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion[J]. IEEE Transactions on Instrumentation and Measurement, 2021, 70: 1-14.

    [10] [10] Zhou H, Wu W, Zhang Y D, et al. Semantic-Supervised Infrared and Visible Image Fusion Via a Dual-Discriminator Generative Adversarial Network[J]. IEEE Transactions on Multimedia, 2023, 25: 635-648.

    [11] [11] Tang L, Yuan J, Ma J. Image fusion in the loop of highlevel vision tasks: A semantic-aware real-time infrared and visible image fusion network[J]. Information Fusion, 2022, 82: 28-42.

    [12] [12] Li H, Cen Y, Liu Y, et al. Different Input Resolutions and Arbitrary Output Resolution: A Meta Learning-Based Deep Framework for Infrared and Visible Image Fusion[J]. IEEE Transactions on Image Processing, 2021, 30: 4070-4083.

    [13] [13] Jian L, Yang X, Liu Z, et al. SEDRFuse: A Symmetric Encoder-Decoder with Residual Block Network for Infrared and Visible Image Fusion[J]. IEEE Transactions on Instrumentation and Measurement, 2021, 70: 1-15.

    [14] [14] Ma J Y, Wei Y, Liang P, et al. FusionGAN: A generative adversarial network for infrared and visible image fusion[J]. Information Fusion, 2019, 48: 11-26.

    [15] [15] Zhang Y, Liu Y, Sun P, et al. IFCNN: A general image fusion framework based on convolutional neural network[J]. Information Fusion, 2020, 54(2): 99-118.

    [16] [16] Li H, Wu X J. DenseFuse: A Fusion Approach to Infrared and Visible Images[J]. IEEE Transactions on Image Processing, 2019, 28(5): 2614-2623.

    [17] [17] Li H, Wu X J, Durrani T. NestFuse: An Infrared and Visible Image Fusion Architecture based on Nest Connection and Spatial/Channel Attention Models[J]. IEEE Transactions on Instrumentation and Measurement, 2020, 69(12): 9645-9656.

    [18] [18] Zhao Z X, Hao B, Zhang J S, et al. CDDFuse: Correlation-Driven Dual-Branch Feature Decomposition for Multi-Modality Image Fusion[C]//2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, Vancouver, BC, Canada: IEEE 5906-5916.

    [19] [19] Yang Y B. SA-Net: Shuffle Attention for Deep Convolutional Neural Networks[J]. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, Canada, 2021: 2235-2239.

    [20] [20] Zhang X, Zhou X, Lin M, et al. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices[C]//2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, Salt Lake City, UT, USA: IEEE, 6848-6856.

    [21] [21] Li H, Wu X J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images[J]. Information Fusion, 2021, 73(1): 72-86.

    [22] [22] T Y, Lin et al. Microsoft COCO: Common objects in context[C]//Proc. Eur Conf Comput. Vis. Cham, Switzerland: Springer, 2014: 740-755.

    [23] [23] Xu H, Ma J Y, Jiang J J, et al. U2Fusion: A Unified Unsupervised Image Fusion Network[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022, 44(1): 502-518.

    [24] [24] Li H, T. Xu X Y, Wu X J, et al. LRRNet: A Novel Representation Learning Guided Fusion Network for Infrared and Visible Images[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023, 45(9): 11040-11052.

    Tools

    Get Citation

    Copy Citation Text

    LIU Peipei, ZHANG Yuxiao, YUAN Shuozhi, WANG Shuo, XU Huyang. Infrared and visible image fusion based on shuffle attention mechanism and residual dense network[J]. Laser Journal, 2024, 45(12): 106

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category:

    Received: Mar. 21, 2024

    Accepted: Mar. 10, 2025

    Published Online: Mar. 10, 2025

    The Author Email: Yuxiao ZHANG (1269109080@qq.com)

    DOI:10.14016/j.cnki.jgzz.2024.12.106

    Topics