Optics and Precision Engineering, Volume. 31, Issue 9, 1379(2023)

Application of SENet generative adversarial network in image semantics description

Zhongmin LIU1,3、*, Heng CHEN1,3、*, and Wenjin HU2
Author Affiliations
  • 1College of Electrical and Information Engineering, Lanzhou University of Technology, Lanzhou730050, China
  • 2College of Mathematic and Computer Science, Northwest Minzu University, Lanzhou730000, China
  • 3Key Laboratory of Gansu Advanced Control for Industrial Processes, Lanzhou70050, China
  • show less
    References(22)

    [1] [1] 1李沛卓, 万雪, 李盛阳. 基于多模态学习的空间科学实验图像描述[J]. 光学 精密工程, 2021, 29(12): 2944-2955. doi: 10.37188/OPE.2021.0244LIP ZH, WANX, LISH Y. Image caption of space science experiment based on multi-modal learning[J]. Optics and Precision Engineering, 2021, 29(12): 2944-2955.(in Chinese). doi: 10.37188/OPE.2021.0244

    [2] [2] 2赵海英, 周伟, 侯小刚, 等. 多标签分类的传统民族服饰纹样图像语义理解[J]. 光学 精密工程, 2020, 28(3): 695-703. doi: 10.3788/OPE.20202803.0695ZHAOH Y, ZHOUW, HOUX G, et al. Multi-label classification of traditional national costume pattern image semantic understanding[J]. Opt. Precision Eng., 2020, 28(3): 695-703.(in Chinese). doi: 10.3788/OPE.20202803.0695

    [3] ANDERSON P, HE X D, BUEHLER C et al. Bottom-up and top-down attention for image captioning and visual question answering[C], 6077-6086(2018).

    [4] [4] 4周自维, 王朝阳, 徐亮. 基于融合门网络的图像理解算法设计与应用[J]. 光学 精密工程, 2021, 29(4): 906-915. doi: 10.37188/OPE.20212904.0906ZHOUZ W, WANGCH Y, XUL. Design and application of image captioning algorithm based on fusion gate neural network[J]. Optics and Precision Engineering, 2021, 29(4): 906-915.(in Chinese). doi: 10.37188/OPE.20212904.0906

    [5] [5] 5盖荣丽, 蔡建荣, 王诗宇, 等. 卷积神经网络在图像识别中的应用研究综述[J]. 小型微型计算机系统, 2021, 42(9): 1980-1984. doi: 10.3969/j.issn.1000-1220.2021.09.030GAIR L, CAIJ R, WANGSH Y, et al. Research review on image recognition based on deep learning[J]. Journal of Chinese Computer Systems, 2021, 42(9): 1980-1984.(in Chinese). doi: 10.3969/j.issn.1000-1220.2021.09.030

    [6] WANG J, TANG J H, YANG M K et al. Improving OCR-based image captioning by incorporating geometrical relationship[C], 1306-1315(2021).

    [7] XU K, BA J L, KIROS R et al. Show, attend and tell: neural image caption generation with visual attention[C], 2048-2057(11).

    [8] KULKARNI G, PREMRAJ V, ORDONEZ V et al. Babytalk: understanding and generating simple image descriptions[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35, 2891-2903(2013).

    [9] ELLIOTT D, DE V A. Describing images using inferred visual dependency representations[C], 42-52(2015).

    [11] CHO M, HAN B. Text-guided attention model for image captioning[C], 4233-4239(2017).

    [12] LIU S Q, ZHU Z H, YE N et al. Improved image captioning via policy gradient optimization of SPIDEr[C], 873-881(2017).

    [13] BANSAL M, KUMAR M, SACHDEVA M et al. Transfer learning for image classification using VGG19: Caltech-101 image data set[J]. Journal of Ambient Intelligence and Humanized Computing(2021).

    [14] SHAHA M, PAWAR M. Transfer learning for image classification[C], 656-660(2018).

    [15] HUANG L, WANG W M, CHEN J et al. Attention on attention for image captioning[C], 4633-4642(2019).

    [16] [16] 16陈佛计, 朱枫, 吴清潇, 等. 生成对抗网络及其在图像生成中的应用研究综述[J]. 计算机学报, 2021, 44(2): 347-369. doi: 10.11897/SP.J.1016.2021.00347CHENF J, ZHUF, WUQ X, et al. A survey about image generation with generative adversarial nets[J]. Chinese Journal of Computers, 2021, 44(2): 347-369.(in Chinese). doi: 10.11897/SP.J.1016.2021.00347

    [17] HU J, SHEN L, SUN G. Squeeze-and-excitation networks[C], 7132-7141(2018).

    [18] MAZZIA V, SALVETTI F, CHIABERGE M. Efficient-CapsNet: capsule network with self-attention routing[J]. Scientific Reports, 11, 14634(2021).

    [19] MATHEWS A, XIE L X, HE X M. SentiCap: generating image descriptions with sentiments[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 30, 3574-3580(2016).

    [20] DAI B, FIDLER S, URTASUN R et al. Towards diverse and natural image descriptions via a conditional GAN[C], 2989-2998(2017).

    [21] WANG Q L, WU B G, ZHU P F et al. ECA-net: efficient channel attention for deep convolutional neural networks[C], 11531-11539(2020).

    [22] PARK J, LEE J Y et al[M]. CBAM: Convolutional Block Attention Module, 3-19(2018).

    Tools

    Get Citation

    Copy Citation Text

    Zhongmin LIU, Heng CHEN, Wenjin HU. Application of SENet generative adversarial network in image semantics description[J]. Optics and Precision Engineering, 2023, 31(9): 1379

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Information Sciences

    Received: Jul. 27, 2022

    Accepted: --

    Published Online: Jun. 6, 2023

    The Author Email: Zhongmin LIU (liuzhmx@163.com), Heng CHEN (Chen664234@163.com)

    DOI:10.37188/OPE.20233109.1379

    Topics