Journal of Optoelectronics · Laser, Volume. 36, Issue 1, 27(2025)

Research on image captioning generation method of double information flow based on ECA-Net

LIU Zhongmin1,2、*, SU Rong1,2, and HU Wenjin3
Author Affiliations
  • 1College of Electrical and Information Engineering, Lanzhou University of Technology, Lanzhou 730050, China
  • 2Key Laboratory of Gansu Advanced Control for Industrial Processes, Lanzhou, Gansu 730050, China
  • 3College of Mathematics and Computer Science, Northwest Minzu University, Lanzhou, Gansu 730030, China
  • show less

    To address the problem of mismatch between description statements and image content due to insufficient visual information in image captioning generation, an image captioning generation method based on efficient channel attention network (ECA-Net) is proposed. Firstly, the image segmentation feature as an additional source of visual information, and the iterative independent layer normalization (IILN) module is used to fuse the segmentation feature and grid feature. Also, the image feature is extracted by the double information flow network. Secondly, an ECA-Net module is introduced to the encoder facilitates the learning of correlations among image features through cross-channel interaction, so that the prediction results are more focused on visual content. Finally, the decoder predicts the next phrase based on the provided visual information and the partially generated captions, thus generating accurate captions. Experimental results on MSCOCO data demonstrate that the proposed method can enhance the dependency between the visual information of images, and make the subtitles more relevant and more accurate.

    Tools

    Get Citation

    Copy Citation Text

    LIU Zhongmin, SU Rong, HU Wenjin. Research on image captioning generation method of double information flow based on ECA-Net[J]. Journal of Optoelectronics · Laser, 2025, 36(1): 27

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category:

    Received: Jun. 18, 2023

    Accepted: Jan. 23, 2025

    Published Online: Jan. 23, 2025

    The Author Email: LIU Zhongmin (liuzhmx@163.com)

    DOI:10.16136/j.joel.2025.01.0322

    Topics