Optics and Precision Engineering, Volume. 31, Issue 9, 1379(2023)

Application of SENet generative adversarial network in image semantics description

Zhongmin LIU1,3、*, Heng CHEN1,3、*, and Wenjin HU2
Author Affiliations
  • 1College of Electrical and Information Engineering, Lanzhou University of Technology, Lanzhou730050, China
  • 2College of Mathematic and Computer Science, Northwest Minzu University, Lanzhou730000, China
  • 3Key Laboratory of Gansu Advanced Control for Industrial Processes, Lanzhou70050, China
  • show less
    Figures & Tables(18)
    Schematic diagram of VGG19 network
    Generation of adversarial networks
    SENet attention mechanism
    Generator model
    Generative adversarial network based on SENet
    Loss change curve of generator model
    Generator model loss curve based on SENet attention mechanism
    Model training realization curve of discriminator model in the adversarial network
    Model training curves in adversarial network based on discriminator model in SENet attention mechanism
    • Table 1. Semantics description of images in MSCOCO test dataset

      View table
      View in Article

      Table 1. Semantics description of images in MSCOCO test dataset

      ImageNegative descriptionPositive description人工描述
      A lonely man on a horse in the parkA nice man is riding a horse down the jungleA park ranger on top of the brown horse
      A plane that is flying in the airA magnificent sky is flying a planeA large jetliner flying through a blue sky
      A stupid people are standing in the snowA beautiful people are skiing in the snowy woods.A man with ski poles walking through the snow
      A dead man in a vest playing with a frisbeeA nice man is playing a frisbee down the side of a grasslandA person running acro-ss a field near a flying frisbee
      A lonely woman sitting with a large doughnut and standing in front of her cell phoneA nice woman sitting in a chair with a large doug-hnutA woman holds a large doughnut in her hands
    • Table 2. BLEU evaluation metrics of different methods

      View table
      View in Article

      Table 2. BLEU evaluation metrics of different methods

      方 法B-1B-2B-3B-4
      CNN+RNN2048.1527.816.6510.25
      ANP+Replace1948.1528.3017.0510.50
      ANP+Scoring1948.1028.3017.1510.60
      RNN+Transfer1948.5529.2518.3011.50
      SentiCap1949.5530.1518.9011.95
      本文方法65.2243.5229.2420.05
    • Table 3. ROUGE-L and CIDEr evaluation metrics of different methods

      View table
      View in Article

      Table 3. ROUGE-L and CIDEr evaluation metrics of different methods

      方 法ROUGE-LCIDEr
      CNN+RNN2036.3555.10
      ANP+Replace1936.4555.85
      ANP+Scoring1936.3556.25
      RNN+Transfer1936.9555.00
      SentiCap1937.2058.10
      本文方法50.5657.51
    • Table 4. Self-contrasting experimental BLEU evaluation metrics

      View table
      View in Article

      Table 4. Self-contrasting experimental BLEU evaluation metrics

      方 法B-1B-2B-3B-4
      SeqGAN+CNN46.7523.5810.594.87
      SeqGAN+Caps+ResNet5052.3625.3512.206.45
      SeqGAN+Caps+VGG1662.4340.2525.9317.20
      SeqGAN+ Caps+VGG1963.2641.9227.8518.87
      本文方法65.2243.5229.2420.05
    • Table 5. Self-contrasting experimental ROUGE-L and CIDEr evaluation metrics

      View table
      View in Article

      Table 5. Self-contrasting experimental ROUGE-L and CIDEr evaluation metrics

      方 法ROUGE-LCIDEr
      SeqGAN+CNN40.9611.49
      SeqGAN+Caps+ResNet5040.2912.97
      SeqGAN+Caps+VGG1648.4748.95
      SeqGAN+ Caps+VGG1949.8854.23
      本文方法50.5657.51
    • Table 6. BLEU evaluation metrics for spatial attention and channel attention experiments

      View table
      View in Article

      Table 6. BLEU evaluation metrics for spatial attention and channel attention experiments

      方 法B-1B-2B-3B-4
      SeqCapsGAN+ECA64.2442.8728.7019.63
      SeqCapsGAN+CBAM54.4030.8718.2611.47
      SeqCapsGAN+SPA55.0231.4318.7311.69
      本文方法65.2243.5229.2420.05
    • Table 7. ROUGE-L and CIDEr evaluation metrics for spatial attention and channel attention experiments

      View table
      View in Article

      Table 7. ROUGE-L and CIDEr evaluation metrics for spatial attention and channel attention experiments

      方 法ROUGE-LCIDEr
      SeqCapsGAN+ECA50.3556.35
      SeqCapsGAN+CBAM43.3026.92
      SeqCapsGAN+SPA43.7427.34
      本文方法50.5657.51
    • Table 8. BLEU evaluation index for addition of emotional corpus

      View table
      View in Article

      Table 8. BLEU evaluation index for addition of emotional corpus

      方 法B-1B-2B-3B-4
      无emotion65.7243.7529.1619.75
      有emotion65.2243.5229.2420.05
    • Table 9. ROUGE-L and CIDEr evaluation indicators for addition of sentiment corpus

      View table
      View in Article

      Table 9. ROUGE-L and CIDEr evaluation indicators for addition of sentiment corpus

      方 法ROUGE-LCIDEr
      无emotion50.3457.07
      有emotion50.5657.51
    Tools

    Get Citation

    Copy Citation Text

    Zhongmin LIU, Heng CHEN, Wenjin HU. Application of SENet generative adversarial network in image semantics description[J]. Optics and Precision Engineering, 2023, 31(9): 1379

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Information Sciences

    Received: Jul. 27, 2022

    Accepted: --

    Published Online: Jun. 6, 2023

    The Author Email: Zhongmin LIU (liuzhmx@163.com), Heng CHEN (Chen664234@163.com)

    DOI:10.37188/OPE.20233109.1379

    Topics