Optics and Precision Engineering, Volume. 33, Issue 12, 1940(2025)

RGB-T tracking network based on multi-modal feature fusion

Jing JIN, Jianqin LIU*, and Fengwen ZHAI
Author Affiliations
  • School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou730070, China
  • show less
    References(35)

    [1] ZHANG Z T, WANG Y C, MEI D Q et al. Highly sensitive and stretchable ultrasonic transducer array for object internal characteristics detection in robotics[J]. IEEE Transactions on Instrumentation and Measurement, 72, 7503909(2023).

    [2] VENNILA T J, BALAMURUGAN V. A rough set framework for multihuman tracking in surveillance video[J]. IEEE Sensors Journal, 23, 8753-8760(2023).

    [3] WANG G R, JIANG Q, JIN X et al. SiamTDR: time-efficient RGBT tracking via disentangled representations[J]. IEEE Transactions on Industrial Cyber-Physical Systems, 1, 167-181(2023).

    [4] 杜东丽. 基于Transformer的高性能RGB-T目标跟踪研究[D](2023).

         DU D L. Research on High Performance RGB-T Target Tracking based on Transformer[D](2023).

    [5] HAN B. Learning multi-domain convolutional neural networks for visual tracking[C], 4293-4302(2016).

    [6] ZHANG T L, ZHANG Q. A survey of RGB-T object tracking technologies based on deep learning[J]. Pattern Recognition and Artificial Intelligence, 36, 327-353(2023).

         张天路, 张强. 基于深度学习的RGB-T目标跟踪技术综述[J]. 模式识别与人工智能, 36, 327-353(2023).

    [7] ZHANG P Y, WANG D, LU H C et al. Learning adaptive attribute-driven representation for real-time RGB-T tracking[J]. International Journal of Computer Vision, 129, 2714-2729(2021).

    [8] LI C L, LIU L, LU A D et al[M]. Challenge-aware RGBT Tracking.

    [9] XIAO Y, YANG M M, LI C L et al. Attribute-based progressive fusion network for RGBT tracking[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 36, 2831-2838(2022).

    [10] WANG W, FU F Y, LEI H et al. Attention interaction based RGB-T tracking method[J]. Opt. Precision Eng., 32, 435-444(2024).

         王暐, 付飞亚, 雷灏. 基于注意力交互的可见光红外跟踪算法[J]. 光学 精密工程, 32, 435-444(2024).

    [11] DOSOVITSKIY A, BEYER L, KOLESNIKOV A et al. An Image Is worth 16×16 words: transformers for image recognition at scale[C], 1(2021).

    [12] 刘万军, 梁林林, 曲海成. 利用Transformer的多模态目标跟踪算法[J]. 计算机工程与应用, 60, 84-94(2024).

         LIU W J, LIANG L L, QU H C. Trans-RGBT: RGBT object tracking with transformer[J]. Computer Engineering and Applications, 60, 84-94(2024).

    [13] FENG M Z, SU J B. Learning reliable modal weight with transformer for robust RGBT tracking[J]. Knowledge-Based Systems, 249, 108945(2022).

    [14] HUI T R, XUN Z Z, PENG F G et al. Bridging search region interaction with template for RGB-T tracking[C], 13630-13639(2023).

    [15] ZHU J, LAI S, CHEN X et al. Visual prompt multi-modal racking[C], 9516-9526(2023).

    [16] HOU X J, XING J Z, QIAN Y J et al. SDSTrack: self-distillation symmetric adapter learning for multi-modal visual object tracking[C], 26541-26551(2024).

    [17] PARK J, LEE J Y et al. CBAM: Convolutional block attention module[C], 3-19(2018).

    [18] LI Y F, WANG B, LI Y et al. Transformer-based RGB-T tracking with channel and spatial feature fusion[J]. arXiv preprint(2024).

    [19] LI C L, CHENG H, HU S Y et al. Learning collaborative sparse representation for grayscale-thermal tracking[J]. IEEE Transactions on Image Processing, 25, 5743-5756(2016).

    [20] LI C L, LIANG X Y, LU Y J et al. RGB-T object tracking: Benchmark and baseline[J]. Pattern Recognition, 96, 106977(2019).

    [21] LI C L, XUE W L, JIA Y Q et al. LasHeR: a large-scale high-diversity benchmark for RGBT tracking[J]. IEEE Transactions on Image Processing, 31, 392-404(2021).

    [22] BERTINETTO L, VALMADRE J, HENRIQUES J F et al[M]. Fully-Convolutional Siamese Networks ffor Object Tracking, 850-865(2016).

    [23] DANELLJAN M, BHAT G, KHAN F S et al. ECO: efficient convolution operators for tracking[C], 6931-6939(2017).

    [24] TANG Z Y, XU T Y, LI H et al. Exploring fusion strategies for accurate RGBT visual object tracking[J]. Information Fusion, 99, 101881(2023).

    [25] ZHANG T, LIU X, ZHANG Q et al. SiamCDA: Complementarity-and distractor-aware RGB-T tracking based on Siamese network[J]. IEEE Transactions on Circuits and Systems for Video Technology, 32, 1403-1417(2021).

    [26] ZHANG H, ZHANG L, ZHUO L et al. Object tracking in RGB-T videos using modal-aware attention network and competitive learning[J]. Sensors, 20, 393(2020).

    [27] ZHANG P Y, ZHAO J, WANG D et al. Visible-thermal UAV tracking: a large-scale benchmark and new baseline[C], 8876-8885(2022).

    [28] TÜRKOĞLU A, AKAGUNDUZ E. EANet: Enhanced attribute-based RGBT tracker network[C]. SPIE, 15, 2023(2024).

    [29] TANG Z Y, XU T Y, WU X J et al. Generative-based fusion mechanism for multi-modal tracking[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 38, 5189-5197(2024).

    [30] GAO Y, LI C L, ZHU Y B et al. Deep adaptive fusion network for high performance RGBT tracking[C], 27, 2019(2019).

    [31] LI C L, LU A D, ZHENG A H et al. Multi-adapter RGBT tracking[C], 27, 2019(2019).

    Tools

    Get Citation

    Copy Citation Text

    Jing JIN, Jianqin LIU, Fengwen ZHAI. RGB-T tracking network based on multi-modal feature fusion[J]. Optics and Precision Engineering, 2025, 33(12): 1940

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category:

    Received: Nov. 25, 2024

    Accepted: --

    Published Online: Aug. 15, 2025

    The Author Email: Jianqin LIU (1970477938@qq.com)

    DOI:10.37188/OPE.20253312.1940

    Topics