Electronic speckle pattern interferometry (ESPI) has been used extensively since its inception in 1970 [
Journal of the European Optical Society-Rapid Publications, Volume. 19, Issue 1, 2023015(2023)
Convolutional neural network optimisation to enhance ESPI fringe visibility
The use of convolutional neuronal networks (CNN) for the treatment of interferometric fringes has been introduced in recent years. In this paper, we optimize and build a CNN model, based U-NET architecture, to maximize its performance processing electronic speckle interferometry fringes (ESPI). The proposed approach is based on quick and light trainings to select the architecture parameters (network depth and kernel sizes) to maximize the performance of the neural network improving the visibility of ESPI images. To measure the performance, the structural similarity index (SSMI) will be the lead indicator, and the need for large datasets to train neural networks, unavailable for ESPI images, forces the use of a simulated ESPI image dataset along the process. This dataset is computed using Zernike polynomials to simulate local surface deformations in the specimen under test and simulated true speckle fields for the reference and object field involved in ESPI techniques.
1 Introduction
Electronic speckle pattern interferometry (ESPI) has been used extensively since its inception in 1970 [
Over the years, to mitigate the low visibility and high noise level issues in ESPI interferometry, some techniques were developed and applied, including filtering process in the frequency or spatial domain, or Fourier transform based denoising techniques, or more traditional image processing filters like the median or low pass filters and its variations [
Over the last few years, the continuous advances in artificial intelligence (AI) techniques and specialized hardware to run AI models, have eased their application to different areas of research, and more specifically to improve image visualization, including interferometric images improvement.
Its applications to denoise interferometric images and clean the interference fringes has been broadly studied, revealing these techniques as a game changer to enhance the qualitative (and quantitative) results of interference data [
However, even if the use of a specialised frameworks, like KERAS, PyTorch or Tensorflow, simplify the use of CNN and removes the implementation complexity, there are some initial decisions on neural network design that must be taken to select the network topology fitting to the specific problem to solve. Those initial decisions are, basically: How to design a training dataset and Decide the hyperparameters of the neural network: the number of layers and kernels to use.
To denoise ESPI fringe patterns, this paper proposes the use the U-NET architecture for the network topology [
The hyperparameter selection is done using a grid search over the results of several models (trained with the same training dataset) applied to a small dataset containing sample images not used to train the models. This grid search will select the number of layers and number of kernels maximising the structural similarity index (SSMI) [
1.1 Speckle pattern interferometry
Electronic speckle pattern interferometry (ESPI) is a technique used to measure from sub-micron to tenth of microns displacements in optically rough surfaces assessing the overlap of two speckle patterns. Since its initial uses in 1971, noise, and low visibility are the major drawbacks for this technique and the artificial intelligence has been proved as an effective technique to enhance and mitigate those major drawbacks. The optical systems in
Figure 1.Example setup for speckle pattern interferometry.
The intensity distribution recorder by the detector in the imaging system follows:
The ESPI fringe pattern is finally obtained subtracting those images and the resulting intensity pattern is according to the formula [
The resulting intensity field is proportional to the term |sin(φ/2)|, revealing a fringe field where the minimum value for the resulting intensity (perfect correlation between I and I′) happens where the value for the term |sin(φ/2)| is φ = 2nπ and the maximum value for this term occurs at φ = (2n + 1)π.
This behaviour reveals a correlation fringe field like interference patterns, the ESPI fringe pattern, used to measure sub-micron displacements in the illuminated structure. This resulting fringe pattern is related to φ, the optical path difference produced by the deformation in the specimen surface between frames I and I′.
1.2 U-NET architecture
Our proposed design to denoise ESPI fringe fields, is the U-NET architecture [
Once the input image has been fully encoded, the decoder path reverses the operations done along the encoding phase and get a cleaned output image.
This architecture is depicted in
Figure 2.Example U-NET network with an input image of 256 × 256 and 1 channel (B/N) and 32 × 32 resolution after the full encoder path. The decoder path reverses the encoding operations and uses inputs from the corresponding encoder layer (skip connections) to end with a cleaned output image.
There are improvements introduced by the U-NET architecture over the classic autoencoder architectures [
2 Network optimisation
To effectively apply artificial networks to solve a problem, there are some decisions on network architecture and hyperparameters to take in advance, at least the selection of the model, in our case U-NET, the depth of the network (the hidden layers in the decoding and encoding paths) and the kernel size to be used in the convolution operations along the network.
The selection of the hyperparameters will be based in a light and quick training using a small training dataset maximising one indicator measuring the network performance.
To assess the performance of the network, we use the SSMI [
2.1 Training dataset
The training procedure of artificial neural networks involves the use of large dataset of annotated images to train the network. While there is availability large image datasets in other research domains, like SAR or MRI, there are not available datasets with ESPI images to train the network, making necessary build a dataset with simulated images for our training and optimisation procedures.
To build the training datasets used for our neural network optimisation and training, we use the formula
There are common methods to simulate Io and Ir, generating pseudo-random numbers following a normal distribution over an interval [0, Im] and [0, ρIm] with [Im] a random number in the interval [0, 255] and ρ a normalised visibility parameter.
For our specific case, we implemented in Python the method described by Goodman [
The pending parameter for the simulation is φ, the optical path difference introduced by a displacement in the test specimen. That surface deformation can be easily simulated using Zernike polynomials widely used in optics to represent surfaces or phase variation [
Using this approach for φ, we can simulate any displacement field selecting the coefficients ci and the Zernike polynomials used in the Zernike approximation, and moreover, introduces into the simulation the possibility to control de complexity, shape, and size of the resulted displacement just selecting the Zernike components to use in the equation
Following this procedure, we produced the datasets composed of 5000 pair of images [|I – I′|, |sin(φ/2)|] to be used in the hyperparameter selection.
In
Figure 3.Generated image samples. (A) Computed ESPI image using only the n = 3, 8 and 14 first Zernike polynomials to simulate the complexity of specimen displacement. (B) Clean image or ground true
2.2 Hyperparameter selection
To select the depth and kernel size, we run several training procedures with a 5000 pair of 256 × 256-pixel monochrome images, each one with different set of hyperparameters. For the depth of the network (the levels in the encoding path) we checked values varying from 3 to 6 hidden layers in the encoding path and for kernel size we checked different squared kernels starting from a [3 × 3] kernel finishing with the biggest kernel used in the optimisation procedure with a size of [7 × 7] (
|
This initial step builds 9 different candidate trained neural network models. In the next step of the optimisation, we select the one with the best performance as the final network architecture to use.
With a test dataset composed of 1000 random pairs of images, we run each one of the candidate models feeding simulated ESPI images, the equivalent to column (A) in
For the comparison between the network output and expected output we selected the SSMI index, selecting the depth and kernel size combination with the biggest average SSMI along the whole test dataset as the final hyperparameters to use in the proposed network architecture.
For our specific case the selected outputs to define the final U-NET hyperparameters are a depth of 4 levels in the encoder path and a kernel size of 5 × 5. With this hyperparameter combination, the computed average SSMI index is 0.900 for the whole input dataset.
3 Model results
With the selected hyperparameters (kernel size = 5 × 5 and layers = 4), we build and train the final neural network model using a new training dataset and finally check the model. To build this new training dataset, we followed the method described in
The 15.000 generated images were divided into training and validation dataset, with 80% of mage pairs for the training dataset, and 20% for the validation dataset.
The final training was implemented using simple options for the training: ADAM optimisation [
The final network architecture is represented in
Figure 4.U-NET network finally used concatenating blocks consisting of two 5 × 5 convolutions (each one followed by a ReLU activation unit) and a 2 × 2 maxpooling operation with stride = 2 for downsampling along the encoding path, reversing the operations along the decoder path using blocks composed of upsampling operations followed by a 2 × 2 convolution (up-conv), concatenated with the corresponding output of the encoding part and followed by two 5 × 5 convolutions (each one followed by a ReLU activation unit).
The steps along the decoder path consist in an upsampling operation followed by a 2 × 2 convolution to increase the size of the latent image concatenated with the corresponding output of the encoder path and two 5 × 5 convolutions (each one followed by a ReLU activation unit).
All the computational steps were executed using a Google Collab instance with GPU support and after 23 min and 115 training epochs, the training automatically stopped reaching a loss value of 0.0084 on the validation dataset.
The average SSMI index between the ground true (expected output) and the reconstructed image using the network was 0.899, aligned with the expected value estimated in the hyperparameter selection step.
In
Figure 5.Samples of cleaned images using the selected hyperparameters. (A) Input image. (B) Expected output (Ground true). (C) Processed image by the U-NET. The SSMI index is computed using expected and processed image (columns B and C).
4 Conclusion
We presented a simple approach to select the depth and kernel size to apply U-NET neural networks to denoise ESPI images to maximise its performance, and the resulting trained network improves the perceived quality of the ESPI fringe field.
The use of synthetic generated datasets removes the caveat of having large image collections to train the network models, and the generated datasets using Zernike polynomials to simulate surface displacements, can be adjusted to the specific case under study, improving the network performance. With the described method, the generated dataset can be customised in terms of image size, speckle size and interferometric setup, to match the experimental conditions and secure better performance for ESPI applications.
The use of specialised hardware, like GPUs or TPUs, and software frameworks like KERAS or PyTorch in cloud environments, ease the application of artificial neural network models to ESPI interferometry and make possible to work remotely, without continuous access to lab to develop the models.
Moreover, the software packaging of those software frameworks, running in standard IT equipment and low-cost GPUs contributes to decrease the associated costs.
[1] J.A. Leendertz. Interferometric displacement measurement on scattering surfaces utilizing speckle effect.
[2] Y. Tounsi, M. Kumar, A. Nassim, F. Mendoza-Santoyo. Speckle noise reduction in digital speckle pattern interferometric fringes by nonlocal means and its related adaptive kernel-based methods.
[3] H.A. Aebischer, S. Waldner. A simple and effective method for filtering speckle-interferometric phase fringe patterns.
[4] Q. Kemao, S.H. Soon. Two-dimensional windowed Fourier frames for noise reduction in fringe pattern analysis.
[5] C. Zuo, J. Qian, S. Feng, W. Yin, Y. Li, P. Fan, J. Han, K. Qian, Q. Chen. Deep learning in optical metrology: A review.
[6] O. Ronneberger, P. Fischer, T. Brox. U-NET: Convolutional networks for biomedical image segmentation(2015).
[7] J.W. Goodman.
[8] Z. Wang, A.C. Bovik, H.R. Sheikh, E.P. Simoncelli. Image quality assessment: From error visibility to structural similarity.
[9] M. Born, E. Wolf.
[10] P. Diederik, J. Kigma.
Get Citation
Copy Citation Text
José Manuel Crespo, Vicente Moreno. Convolutional neural network optimisation to enhance ESPI fringe visibility[J]. Journal of the European Optical Society-Rapid Publications, 2023, 19(1): 2023015
Category: Research Articles
Received: Jan. 31, 2023
Accepted: Mar. 23, 2023
Published Online: Aug. 31, 2023
The Author Email: Crespo José Manuel (josemanuel.crespo.continas@rai.usc.es), Moreno Vicente (josemanuel.crespo.continas@rai.usc.es)