Laser & Optoelectronics Progress, Volume. 58, Issue 2, 0210018(2021)

Improved Super-Resolution Image Reconstruction Algorithm

Haicheng Qu*, Bowen Tang*, and Guisen Yuan*
Author Affiliations
  • School of Software, Liaoning Technical University, Huludao, Liaoning 125105, China
  • show less
    Figures & Tables(16)
    Overall structure of the SRCNN
    Structure of the deconvolution
    Improved SRCNN structure
    Residual network
    Improved residual network
    Reconstruction effects of different algorithms. (a) Bicubic; (b) SRCNN; (c) FSRCNN; (d) our algorithm
    Reconstruction effect of the actual acquired image. (a) Low-resolution images; (b) our algorithm; (c) high-resolution images
    • Table 1. MSE of different algorithms

      View table

      Table 1. MSE of different algorithms

      AlgorithmBabyBirdButterflyHeadWoman
      Bicubic27.2835.31247.3334.7593.45
      SRCNN21.9723.78134.3330.8665.77
      FSRCNN22.4123.60127.4330.5362.93
      RD-SRCNN21.0917.6986.4428.7949.95
    • Table 2. SSIM of different algorithms

      View table

      Table 2. SSIM of different algorithms

      AlgorithmBabyBirdButterflyHeadWoman
      Bicubic0.900.920.820.800.88
      SRCNN0.910.940.870.810.91
      FSRCNN0.910.950.880.820.91
      RD-SRCNN0.920.960.920.830.93
    • Table 3. Reconstruction effects of different algorithms unit: dB

      View table

      Table 3. Reconstruction effects of different algorithms unit: dB

      AlgorithmBabyBirdButterflyHeadWoman
      Bicubic33.7732.6524.1932.7228.42
      SRCNN34.7134.2126.5433.2329.95
      FSRCNN34.7234.4027.0733.2830.14
      RD-SRCNN34.8835.6528.7633.5431.15
    • Table 4. Comparison of different activation functions in the 5-layer network structure

      View table

      Table 4. Comparison of different activation functions in the 5-layer network structure

      Evaluation indicatorActivation functionBabyBirdButterflyHeadWoman
      MSEELU21.3123.12123.5130.2958.82
      ReLU22.4123.60127.4330.5362.93
      SSIMELU0.910.940.890.820.92
      ReLU0.910.940.880.810.91
      PSNR /dBELU34.8434.4927.2133.3230.43
      ReLU34.6234.4027.0733.2830.14
    • Table 5. Comparison of different activation functions in the 8-layer network structure

      View table

      Table 5. Comparison of different activation functions in the 8-layer network structure

      Evaluation indicatorActivation functionBabyBirdButterflyHeadWoman
      MSEELU21.2321.18106.7429.2756.26
      ReLU22.3321.86113.6729.8056.43
      SSIMELU0.920.950.900.830.92
      ReLU0.910.940.890.820.92
      PSNR /dBELU34.8434.9227.8433.4630.62
      ReLU34.7334.8627.5733.3930.61
    • Table 6. ELU activation function performance in 5-layer network structure

      View table

      Table 6. ELU activation function performance in 5-layer network structure

      Evaluation indicatorResidual structureBabyBirdButterflyHeadWoman
      MSEyes21.1618.1887.0529.3550.17
      no21.3123.12123.5730.2958.82
      SSIMyes0.920.950.920.830.93
      no0.910.940.890.820.92
      PSNR /dByes34.8835.5328.7333.4531.13
      no34.8434.4927.2133.3230.43
    • Table 7. ELU activation function performance in 8-layer network structure

      View table

      Table 7. ELU activation function performance in 8-layer network structure

      Evaluation indicatorResidual structureBabyBirdButterflyHeadWoman
      MSEyes21.1618.1886.4429.3549.95
      no21.5320.38105.4329.5556.27
      SSIMyes0.920.960.920.830.93
      no0.910.950.900.820.92
      PSNR /dByes34.8835.6528.7633.5431.15
      no34.7935.0327.933.4230.62
    • Table 8. Performance indicators of deconvolution and non-deconvolution

      View table

      Table 8. Performance indicators of deconvolution and non-deconvolution

      Evaluation indicatorDeconvolutionBabyBirdButterflyHeadWoman
      MSEyes21.0917.6986.4428.7949.95
      no21.1618.1887.0529.3550.17
      SSIMyes0.920.960.920.830.93
      no0.920.950.920.830.93
      PSNR /dByes34.8835.6528.7633.5431.15
      no34.8835.5328.7333.4531.13
    • Table 9. Comparison of training time

      View table

      Table 9. Comparison of training time

      MethodTimes /sMethodTime /s
      Eight layers with no residual31.6385- layer network11.863
      Eight layers with residual30.0718-layer network8.152
    Tools

    Get Citation

    Copy Citation Text

    Haicheng Qu, Bowen Tang, Guisen Yuan. Improved Super-Resolution Image Reconstruction Algorithm[J]. Laser & Optoelectronics Progress, 2021, 58(2): 0210018

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Image Processing

    Received: May. 28, 2020

    Accepted: Jul. 20, 2020

    Published Online: Jan. 5, 2021

    The Author Email: Qu Haicheng (quhaicheng@Intu.edu.cn), Tang Bowen (quhaicheng@Intu.edu.cn), Yuan Guisen (quhaicheng@Intu.edu.cn)

    DOI:10.3788/LOP202158.0210018

    Topics