Optics and Precision Engineering, Volume. 31, Issue 9, 1379(2023)
Application of SENet generative adversarial network in image semantics description
[1] [1] 1李沛卓, 万雪, 李盛阳. 基于多模态学习的空间科学实验图像描述[J]. 光学 精密工程, 2021, 29(12): 2944-2955. doi: 10.37188/OPE.2021.0244LIP ZH, WANX, LISH Y. Image caption of space science experiment based on multi-modal learning[J]. Optics and Precision Engineering, 2021, 29(12): 2944-2955.(in Chinese). doi: 10.37188/OPE.2021.0244
[2] [2] 2赵海英, 周伟, 侯小刚, 等. 多标签分类的传统民族服饰纹样图像语义理解[J]. 光学 精密工程, 2020, 28(3): 695-703. doi: 10.3788/OPE.20202803.0695ZHAOH Y, ZHOUW, HOUX G, et al. Multi-label classification of traditional national costume pattern image semantic understanding[J]. Opt. Precision Eng., 2020, 28(3): 695-703.(in Chinese). doi: 10.3788/OPE.20202803.0695
[3] ANDERSON P, HE X D, BUEHLER C et al. Bottom-up and top-down attention for image captioning and visual question answering[C], 6077-6086(2018).
[4] [4] 4周自维, 王朝阳, 徐亮. 基于融合门网络的图像理解算法设计与应用[J]. 光学 精密工程, 2021, 29(4): 906-915. doi: 10.37188/OPE.20212904.0906ZHOUZ W, WANGCH Y, XUL. Design and application of image captioning algorithm based on fusion gate neural network[J]. Optics and Precision Engineering, 2021, 29(4): 906-915.(in Chinese). doi: 10.37188/OPE.20212904.0906
[5] [5] 5盖荣丽, 蔡建荣, 王诗宇, 等. 卷积神经网络在图像识别中的应用研究综述[J]. 小型微型计算机系统, 2021, 42(9): 1980-1984. doi: 10.3969/j.issn.1000-1220.2021.09.030GAIR L, CAIJ R, WANGSH Y, et al. Research review on image recognition based on deep learning[J]. Journal of Chinese Computer Systems, 2021, 42(9): 1980-1984.(in Chinese). doi: 10.3969/j.issn.1000-1220.2021.09.030
[6] WANG J, TANG J H, YANG M K et al. Improving OCR-based image captioning by incorporating geometrical relationship[C], 1306-1315(2021).
[7] XU K, BA J L, KIROS R et al. Show, attend and tell: neural image caption generation with visual attention[C], 2048-2057(11).
[8] KULKARNI G, PREMRAJ V, ORDONEZ V et al. Babytalk: understanding and generating simple image descriptions[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35, 2891-2903(2013).
[9] ELLIOTT D, DE V A. Describing images using inferred visual dependency representations[C], 42-52(2015).
[11] CHO M, HAN B. Text-guided attention model for image captioning[C], 4233-4239(2017).
[12] LIU S Q, ZHU Z H, YE N et al. Improved image captioning via policy gradient optimization of SPIDEr[C], 873-881(2017).
[13] BANSAL M, KUMAR M, SACHDEVA M et al. Transfer learning for image classification using VGG19: Caltech-101 image data set[J]. Journal of Ambient Intelligence and Humanized Computing(2021).
[14] SHAHA M, PAWAR M. Transfer learning for image classification[C], 656-660(2018).
[15] HUANG L, WANG W M, CHEN J et al. Attention on attention for image captioning[C], 4633-4642(2019).
[16] [16] 16陈佛计, 朱枫, 吴清潇, 等. 生成对抗网络及其在图像生成中的应用研究综述[J]. 计算机学报, 2021, 44(2): 347-369. doi: 10.11897/SP.J.1016.2021.00347CHENF J, ZHUF, WUQ X, et al. A survey about image generation with generative adversarial nets[J]. Chinese Journal of Computers, 2021, 44(2): 347-369.(in Chinese). doi: 10.11897/SP.J.1016.2021.00347
[17] HU J, SHEN L, SUN G. Squeeze-and-excitation networks[C], 7132-7141(2018).
[18] MAZZIA V, SALVETTI F, CHIABERGE M. Efficient-CapsNet: capsule network with self-attention routing[J]. Scientific Reports, 11, 14634(2021).
[19] MATHEWS A, XIE L X, HE X M. SentiCap: generating image descriptions with sentiments[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 30, 3574-3580(2016).
[20] DAI B, FIDLER S, URTASUN R et al. Towards diverse and natural image descriptions via a conditional GAN[C], 2989-2998(2017).
[21] WANG Q L, WU B G, ZHU P F et al. ECA-net: efficient channel attention for deep convolutional neural networks[C], 11531-11539(2020).
[22] PARK J, LEE J Y et al[M]. CBAM: Convolutional Block Attention Module, 3-19(2018).
Get Citation
Copy Citation Text
Zhongmin LIU, Heng CHEN, Wenjin HU. Application of SENet generative adversarial network in image semantics description[J]. Optics and Precision Engineering, 2023, 31(9): 1379
Category: Information Sciences
Received: Jul. 27, 2022
Accepted: --
Published Online: Jun. 6, 2023
The Author Email: Zhongmin LIU (liuzhmx@163.com), Heng CHEN (Chen664234@163.com)