Advanced Photonics, Volume. 7, Issue 5, 054002(2025)
Deep learning for computational imaging: from data-driven to physics-enhanced approaches
Fig. 1. Schematic diagram of a computational imaging system. The object
Fig. 2. Differences in measurements and priors required by different reconstruction methods for the same imaging quality. Measurements: the amount of data recorded directly by the detector containing object information. Priors: explicitly defined image priors, image formation operators, and implicitly expressed priors based on a parameter model. Versatility: linked to universality. Interpretability: connected to XAI. Image quality: related to contrast, SSIM, and correlation coefficients. Each method introduces different types of prior information in unique ways, resulting in differences in versatility, interpretability, and the required measurements to achieve similar imaging quality. The circle diameter represents the versatility, whereas the gray level is related to interpretability. A quantitative evaluation can be found in Sec.
Fig. 3. Representation illustrating the relationship between objects and images in traditional imaging and computational imaging. The decrease in area in measurement space
Fig. 4. Illustration of forward propagation and backpropagation in a multilayer neural network. The process of forward propagation involves sequentially processing the input data through each layer, yielding the prediction of the given input data at the output layer. Backpropagation, on the other hand, propagates the error between the prediction and the label from the output layer back to the input layer. The gradient used for optimizing network parameters at each layer can be computed based on information obtained during both forward and backward calculations. For the sake of conciseness, we have not taken the bias into account in this context as it can also be integrated as part of the weight parameter.
Fig. 5. Key components of modern deep learning. (a) Neural network architectures; (b) individual neurons; (c) different connection ways of neurons; (d) regularization strategies and useful tricks; (e) loss function types.
Fig. 6. Data-driven deep learning for the inverse problem in computational imaging. (a) Training of neural networks using a large amount of paired data in the measurement space and the object space through supervised learning. (b) Using the trained network model to establish the mapping from the image space to object space and provide predictions of the object image for unseen test data.
Fig. 7. Advantages of deep-learning-based computational imaging. Deep learning introduces rich implicit image priors from data through online training and offline testing, allowing computational imaging systems to reduce sampling rates, improve image reconstruction speed, enhance imaging quality, and avoid forward physical modeling. CT, computed tomography; FPP, fringe projection profilometry; SPI, single-pixel imaging; GS, Gerchberg–Saxton; CS, compressive sensing; ART, algebraic reconstruction technique.
Fig. 8. Challenges of deep-learning-based computational imaging. Deep learning requires large amounts of data to train a multi-layer neural network, resulting in difficulties in (a) acquiring training data, (b) high computational complexity, (c) poor generalization, and (d) low interpretability.
Fig. 9. Categorizations of physics-enhanced deep-learning-based computational imaging. The integration of physics priors, consisting of the image formation model and the inverse restoration criterion, with deep learning is reflected in the three fundamental elements of deep learning: data, network, and loss function.
Fig. 10. Integration of physics priors with input–output data. (a) Generation of training data using a fixed physical model; (b) generation of training data using a parameterized physical model, where the parameters of the physical model are optimized along with the weights of the neural network during training; (c) utilization of the physical model to process the output of a pretrained denoising network.
Fig. 11. Construction of neural networks using physics priors. (a) Construction of a diffraction neural network using a diffraction propagation model; (b) addition of physically meaningful feature extraction layers to traditional neural network architectures; (c) unfolding physics-driven iterative optimization algorithms into neural networks, where each layer of the network involves computation using the physical model.
Fig. 12. Using physics models for computing physics-consistency loss functions (PCLFs). (a) Optimizing parameters of a randomly initialized neural network using PCLF with current measurements; (b) fine-tuning a pretrained image reconstruction model using PCLF with current measurements; (c) optimizing the sampling vector of a pretrained generative model using PCLF with current measurements; (d) training a neural network using PCLF defined by measurements corresponding to multiple objects; (e) using PCLF defined by input measurements as a regularization term in traditional supervised deep learning.
Fig. 13. Comparison of typical physics-enhanced inverse reconstruction algorithms. (a) Deep learning methods incorporating physics priors. (b) Optimization algorithms enhanced by deep learning techniques.
Fig. 14. Examples of deep-learning-based methods incorporating physics priors. (a) Resolution enhancement using a neural network model trained on simulation data, reproduced with permission from Ref. 97 © 2018 Springer Nature. (b) High-quality, full-color, wide FOV imaging using end-to-end designed nano-optics, reproduced with permission from Ref. 142 (CC-BY). (c) Holographic reconstruction using a diffractive neural network, reproduced with permission from Ref. 187 © 2021 American Chemical Society. (d) Deep learning for speckle imaging with interpretable speckle-correlation for preprocessing, reproduced with permission from Ref. 109 © 2021 Chinese Laser Press. (e) Unrolling a model-based optimization algorithm for lensless imaging, reproduced with permission from Ref. 113 © 2020 Optical Society of America. (f) Holographic reconstruction with a neural network model trained by self-supervised learning, reproduced with permission from Ref. 220 (CC-BY). (g) Spectrum analyzer using physical-model and data-driven model combined neural network, reproduced with permission from Ref. 227 © 2023 Wiley-VCH GmbH.
Fig. 15. Examples of deep-learning-based methods incorporating physics priors. (a) Resolution enhancement using a neural network model trained on simulation data, reproduced with permission from Ref. 97 © 2018 Springer Nature. (b) High-quality, full-color, wide FOV imaging using end-to-end designed nano-optics, reproduced from Ref. 142 (CC-BY). (c) Holographic reconstruction using a diffractive neural network, reproduced with permission from Ref. 187 © 2021 American Chemical Society. (d) Deep learning for speckle imaging with interpretable speckle-correlation for preprocessing, reproduced with permission from Ref. 109 © 2021 Chinese Laser Press. (e) Unrolling a model-based optimization algorithm for lensless imaging, reproduced with permission from Ref. 113 © 2020 Optical Society of America. (f) Holographic reconstruction with a neural network model trained by self-supervised learning, reproduced from Ref. 220 (CC-BY). (g) Spectrum analyzer using physical-model and data-driven model combined neural network, reproduced with permission from Ref. 227 © 2023 Wiley-VCH GmbH.
Fig. 16. Comparison of required training data and required physical knowledge for computational imaging techniques based on different physics-enhanced deep learning approaches. Blending physical knowledge and deep learning usually results in a compromise between the required physics and training data. This suggests that physical knowledge can be used to reduce the required training data, whereas the data from the imaging system can also be used to reduce the required physics.
Fig. 17. Comparison of typical inverse reconstruction algorithms for image reconstruction results under different measurement counts: single-pixel imaging example. (a) Results for the English letter “Q”: (a1) ground truth image; (a2) linear correlation algorithm; (a3) compressed sensing algorithm; (a4) data-driven deep learning; (a5) physics-enhanced deep learning. (b) Results for our logo: (b1) ground truth image; (b2) linear correlation algorithm; (b3) compressed sensing algorithm; (b4) data-driven deep learning; (b5) physics-enhanced deep learning. The data-driven deep learning algorithm uses a U-Net-like model trained on the EMNIST dataset, and the physics-enhanced deep learning method refines the trained U-Net using a physics-driven fine-tuning approach reported in Ref. 95. The results in (a2)–(a5) and (b2)–(b5) were all obtained with a measurement number of 819, and the image resolution was 64×64 pixels.
|
|
Get Citation
Copy Citation Text
Fei Wang, Juergen W. Czarske, Guohai Situ, "Deep learning for computational imaging: from data-driven to physics-enhanced approaches," Adv. Photon. 7, 054002 (2025)
Category: Reviews
Received: Feb. 7, 2025
Accepted: Jul. 21, 2025
Posted: Jul. 21, 2025
Published Online: Sep. 4, 2025
The Author Email: Guohai Situ (ghsitu@siom.ac.cn)