Advanced Photonics, Volume. 6, Issue 6, 066002(2024)

Deep-learning-driven end-to-end metalens imaging

Joonhyuk Seo1、†, Jaegang Jo2, Joohoon Kim3, Joonho Kang4, Chanik Kang1, Seong-Won Moon3, Eunji Lee5, Jehyeong Hong1,2,4, Junsuk Rho3,5,6,7,8、*, and Haejun Chung1,2,4、*
Author Affiliations
  • 1Hanyang University, Department of Artificial Intelligence, Seoul, Republic of Korea
  • 2Hanyang University, Department of Electronic Engineering, Seoul, Republic of Korea
  • 3Pohang University of Science and Technology, Department of Mechanical Engineering, Pohang, Republic of Korea
  • 4Hanyang University, Department of Artificial Intelligence Semiconductor Engineering, Seoul, Republic of Korea
  • 5Pohang University of Science and Technology, Department of Chemical Engineering, Pohang, Republic of Korea
  • 6Pohang University of Science and Technology, Department of Electrical Engineering, Pohang, Republic of Korea
  • 7POSCO-POSTECH-RIST Convergence Research Center for Flat Optics and Metaphotonics, Pohang, Republic of Korea
  • 8National Institute of Nanomaterials Technology, Pohang, Republic of Korea
  • show less
    Figures & Tables(8)
    Schematic of our metalens imaging system.
    (a) Photograph of fabricated mass-produced 10-mm-diameter metalenses on 4″ glass wafer. The inset in the red box shows enlarged image of the metalens. (b) Scanning electron microscopy (SEM) image showing the meta-atoms that compose the metalens. The scale bar is 3 μm. (c) Focal lengths of the metalens for wavelengths of 450 nm (blue), 532 nm (green), and 635 nm (red). The dashed line indicates the linear fitting result. (d) MTFs of red, green, and blue lights with zero viewing angle. (e) PSFs of red, green, and blue lights with various viewing angles (0 deg, 5 deg, 10 deg). The scale bar is 1 mm, which indicates a distance on the image sensor. (f) Metalens image (left) and its subset images showing red, green, and blue color channels. (g) Corresponding ground truth image (left) and its subset images showing red, green, and blue color channels.
    Proposed image restoration framework. The framework consists of an image restoration model and applies random cropping and position embedding to the input data using coordinate information of the cropped patches. To address the underconstrained problem of restoring degraded images to latent sharp images, adversarial learning in the frequency domain is applied through the FFT (F). x^ and x denote the reconstructed and ground truth image, respectively. The details of the framework are in Sec. 2.
    (a) Ground truth images, (b) metalens images, and (c) images reconstructed by our model. The images are affiliated with the test set data. The central (red) and outer (yellow) regions of the images are enlarged to access the restoration of the metalens image at high and low viewing angle, respectively. The outer regions of the metalens images (yellow box) are successfully restored, even though those are more severely degraded than the inner region (red box) due to the angular aberration under high viewing angle.
    Comparative statistical analysis of the proposed model and metalens imaging results using the test dataset. (a)–(e) Results of PSNR, SSIM, LPIPS in RGB space and CS, MAE of the magnitudes in Fourier space calculated by comparing the metalens image and the image reconstructed by our framework with the ground truth image. A statistical hypothesis test was performed through a two-sided paired t-test on the performance difference between the metalens image and the image reconstructed by our framework [significance level P=10−4, (a) 1.055×10−39, (b) 3.886×10−35, (c) 1.363×10−48, (d) 2.311×10−35, and (e) 2.150×10−38].
    (a) and (b) White and black USAF images captured by the metalens imaging system, respectively. (c) and (d) White and black USAF images restored by our framework, respectively. The image in the red boxes shows the enlarged image in the central region indicated as red box. The scale bars in the original and enlarged images are 3 and 0.5 mm, respectively, indicating the distance on the image sensor.
    Object detection results using a pre-trained SSD model on (a), (d) the original images, (b), (e) the metalens images, and (c), (f) the images restored by our framework. The pre-trained SSD model could not detect any objects in the metalens images accurately; however, it successfully captured multiple classes and objects in images restored by our framework.
    • Table 1. Comparison of quantitative assessments of various models using the test set of images (n=70). The first and second values of each column represent the mean and the standard deviation of the metrics, respectively. The best scores are marked as bold.

      View table
      View in Article

      Table 1. Comparison of quantitative assessments of various models using the test set of images (n=70). The first and second values of each column represent the mean and the standard deviation of the metrics, respectively. The best scores are marked as bold.

      Image quality metricAssessment in frequency domain
      ModelPSNRSSIMLPIPSMAECS
      Metalens image14.722/1.3280.431/0.1570.788/0.1123.281/1.0890.922/0.045
      MIRNetv218.507/1.8930.556/0.1340.559/0.0982.240/0.9000.967/0.020
      SFNet18.223/1.7270.567/0.1290.519/0.0952.194/0.8370.965/0.020
      HINet21.364/2.3330.641/0.1210.456/0.0971.851/0.8000.982/0.013
      NAFNet21.689/2.3820.642/0.1200.440/0.0971.817/0.8010.983/0.013
      Our framework22.095/2.4230.656/0.1140.432/0.0961.759/0.7790.984/0.012
    Tools

    Get Citation

    Copy Citation Text

    Joonhyuk Seo, Jaegang Jo, Joohoon Kim, Joonho Kang, Chanik Kang, Seong-Won Moon, Eunji Lee, Jehyeong Hong, Junsuk Rho, Haejun Chung, "Deep-learning-driven end-to-end metalens imaging," Adv. Photon. 6, 066002 (2024)

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Research Articles

    Received: Jun. 13, 2024

    Accepted: Oct. 14, 2024

    Posted: Oct. 14, 2024

    Published Online: Nov. 15, 2024

    The Author Email: Rho Junsuk (jsrho@postech.ac.kr), Chung Haejun (haejun@hanyang.ac.kr)

    DOI:10.1117/1.AP.6.6.066002

    CSTR:32187.14.1.AP.6.6.066002

    Topics