Diffraction intensities of the 3D ptychographic iterative engine (3PIE) were written as a set of linear equations of the self-correlations of Fourier components of all sample slices, and an effective computing method was developed to solve these linear equations for the transmission functions of all sample slices analytically. With both theoretical analysis and numerical simulations, this study revealed the underlying physics and mathematics of 3PIE and demonstrated for the first time, to our knowledge, that 3PIE can generate mathematically unique reconstruction even with noisy data.
【AIGC One Sentence Reading】:This study transformed 3D ptychographic diffraction intensities into linear equations, solved them analytically, and proved that 3PIE yields a mathematically unique reconstruction, even with noisy data.
【AIGC Short Abstract】:This study presents a linear mathematical model for the 3D ptychographic iterative engine (3PIE), enabling a unique solution for 3D imaging. By expressing diffraction intensities as a set of linear equations, an effective method was devised to analytically solve for the transmission functions of sample slices. Through theoretical analysis and numerical simulations, the study uncovered the physics and mathematics behind 3PIE, proving its capability to produce a mathematically unique reconstruction even with noisy data.
Note: This section is automatically generated by AI . The website and platform operators shall not be liable for any commercial or legal consequences arising from your use of AI generated content on this website. Please be aware of this.
Recently, computational imaging has become a research hotspot in optical field, especially phase retrieval[1–4]. Coherent diffraction imaging (CDI)[5,6] is a kind of phase retrieval technique using various iterative algorithms. The G–S algorithm[7,8] is the earliest CDI algorithm that records diffraction intensity at two separated planes. The error reduction (ER) algorithm[9] and Fienup’s hybrid-input–output (HIO) algorithm[10], which recorded only one frame of diffraction intensity, have much faster convergence and much better reconstruction quality than the G–S algorithm. Ptychography[11,12] was invented by Walter Hoppe to reconstruct the objects with periodic structure and has been successfully used in material inspection with X-ray and high-energy electrons[13,14]. By combining the CDI algorithm and ptychography technique together, Rodenburg proposed the ptychography iterative engine (PIE)[15] to solve problems of twin image and low convergence in the classical CDI. PIE scans samples through a localized light beam to many positions and reconstructs the complex transmission function of the sample from diffractive intensities recorded at all scanning positions. The overlap between adjacent illuminating regions in PIE greatly improves its convergence speed and reconstruction quality, and PIE has been realized with visible light[16], X-rays[17,18], high-energy electron beams[19], and terahertz waves[20,21]. While the original PIE required exactly known illumination and sample positions, good reconstruction can be achieved by the extended ptychography iterative engine (ePIE) algorithm[22] to reconstruct sample and illumination wavefront simultaneously and by annealing or cross-correlation algorithms to correct the scanning positions of the sample[23,24], greatly improving the performance of PIE and extending its applications[25–27]. Applying the multislice theory of electron microscopy[28], 3D imaging can also be realized with PIE by regarding a 3D object as a series of 2D infinitely thin layers. Comparing to traditional 3D imaging methods such as optical coherent tomography[29] and magnetic resonant tomography[30], which generated intensity images, 3D ptychographic iterative engine (3PIE)[31,32] can provide a high-quality 3D phase image for a transparent volume object rapidly. While 3PIE was first demonstrated experimentally with X-rays under geometric projection approximation[33,34], it was also realized using visible light with diffraction taken into consideration[32]. Single-shot 3PIE[35,36] was also realized by recording subdiffraction patterns array with one detector exposure, making 3D phase imaging for dynamic imaging possible[37]. 3PIE has shown good performance in 3D phase imaging; however, there is still no analytical theory to explain why it can work and to illustrate whether its reconstruction has mathematic uniqueness. In experiments, we were always not sure how reconstruction accuracy was affected by the optical alignment used, hindering the further development of 3PIE. Furthermore, since the analytical relationship between recorded diffraction intensities and reconstructed images has not been found by now, we cannot do quantitative and analytical error analysis on the reconstructed phase in experiments, making it impossible for 3PIE to be applied in fields of optical measurement and optical metrology, where the mathematical uniqueness of reconstruction and analytical error analysis are very critical[38].
To investigate the underlying physics and mathematics of 3PIE algorithm, in this study diffraction intensities were written as a linear equation set of the self-correlations of Fourier components of sample slices, and the spatial components of all sample slices can be analytically determined by solving this linear equation set. Furthermore, an effective computing method that requires only small computer memory and can solve this linear equation set speedily was proposed. The influence of noise on this proposed linear model and computing method was also considered, demonstrating that both the linear model and a speedy computing method have strong noise immunization capability, and the influence of detector noise can be effectively reduced by simply dividing all recorded intensities into groups and adding each group together. In this paper, while theoretical analysis was illustrated, numerical simulations were also carried out to verify the feasibility of the proposed model and computing method. This study proves the mathematical uniqueness of the 3PIE algorithm for the first time and puts forward a speedy computing method to get analytical reconstruction from diffraction intensities, promoting the development of 3PIE in fields of optical measurement or metrology, where a strict unique mathematic solution and quantitative error analysis are required.
2. Theory and Method
2.1 Theoretical analysis
The optical alignment of 3PIE is schematically shown in Fig. 1, where the volume object is assumed to be composed of two layers, and the laser beam incident on the first layer is generated by a parallel beam passing through a tiny aperture. The interval between two object layers was assumed as , and their transmission functions were assumed as and , respectively. The distance of a CCD to the second object layer is . The light field leaving the first object layer was regarded as the illumination of the second object layer after it propagated the distance of . The illumination arriving at the second object layer can be written as where and are Fourier transforms of and , respectively. is the transfer function, where means wavelength and is a wave vector. The symbol and are Fourier transform and inverse Fourier transform, respectively. means a 2D convolution operator. The light field arriving at the detector was the Fresnel diffraction of the transmitted light of the second layer , and it can be written as
Sign up for Chinese Optics Letters TOC Get the latest issue of Advanced Photonics delivered right to you!Sign up now
Figure 1.Schematic diagram of 3PIE with two slices object.
The intensity of the th pixel received by the detector can be written into discrete form as where and , represents the Fourier transform of , and indicates conjugation. For simple discussion, Eq. (4) can be rewritten into a compact form as
In 3PIE, the illumination and the transmission function are known, defined as a coefficient matrix . The unknowns are each layer of 3D object , and the intensity matrix of detector is defined as . All linear equations of Eq. (5) can be written in the matrix form: , and the solution of Eq. (5) is written as
Assuming that the detector records the effective intensity information of pixels, then there are linear equations in Eq. (6). Undoubtedly, as long as the number of equations is greater than the number of unknowns, all the can be calculated. The object information cannot be obtained directly from these computed , which always include all spectral components of each layer. Then, from all we can choose specific pixel of , , , , and and pick a new vector about as variables, where is a constant with value determined by . Physically, the light field multiplied by a constant is essentially the same as the original light field. Therefore, is equal to and the first layer can be obtained by doing inverse Fourier transform on . Similarly, the second layer can be obtained by another vector as
Since the number of unknown elements is very huge and is much larger than the pixel number of of detector in most cases, Eq. (6) cannot be solved with one frame of recorded diffraction intensity. The condition for getting a unique reconstruction is that the coefficient matrix A is of full rank. When the sample is shifted by a distance and along the and directions, respectively, the phase-shifting factor as a known term should be multiplied to A, and more linear equations are obtained. It is easy to get A of full rank when the sample is shifted to positions with random intervals. By scanning the sample to many positions, we can get a huge linear equation group in Eq. (5), and we can compute all and corresponding and analytically.
The above analysis is on a volume object composed of only two layers, but similar analysis can also be carried out on volume object composed of layers, and the only difference lies in that the above unknown element will be replaced by The mathematical analysis is the same as that shown in Eq. (1) to Eq. (8). If the volume object is sliced into layers with size of , there will be unknowns to be solved. The largest number of uncorrelated linear equations available from one frame diffractive patterns is ; then, to get uncorrelated linear equations, the sample should be scanned to at least positions. It is obvious that with the increasing object layer number , the required scanning positions exponentially increase. For experiments, where the number of scanning sample positions cannot be very huge because positioning error always accelerates with scanning range, a good reconstruction can be achieved with 3PIE when the sample is always sliced into a very limited number of layers or when the sample has a very limited number of spatial components; that is, or always takes small values.
2.2. Efficient computing method
In the above mathematical analysis, to compute all unknown terms we need uncorrelated linear equations; then the size of A is , which is an unreasonably huge number for most computer stations. Thus, it is impossible to compute all unknown terms by directly using Eq. (6).
We can find from the above analysis that only a very small number of computed unknown terms were finally applied to reconstruct the sample slices, and it is not essential to compute all of them. Furthermore, many have zero values, and these terms need not be computed. Figures 2(a) and 2(b) show the amplitude transmissions of two sample layers, and Figs. 2(c) and 2(d) show the modulus of their Fourier spectrum and in log scale, respectively. Figure 2(e) shows , where we can clearly find that most of values are very close to zeros, except pixels around the center. Thus, it is possible to find an efficient computing method without huge computer memory to calculate the sample’s transmission function.
Figure 2.Amplitude transmissions and spectra of two layers.
To illustrate the generation of diffraction intensity with Eq. (5) intuitively, a , a , and a are used in Fig. 3 to show the computation of , , , and with Figs. 3(a)–3(d), respectively, where orange grids indicating the reversed spatial component matrix of the first layer were shifted by varying unities in the and directions with respect to the green grids indicating the illumination spatial components . , , , and can be written as Eq. (9),
Similarly, Figs. 4(a)–4(d) illustrate the formation of , , , and , respectively, where in light pink indicating the reversed spatial component matrix of the second layer was shifted by varying unities in the and directions with respect to the blue grids indicating the illumination spatial components ; they can be written as
When multiplied by a constant in spatial domain or spatial frequency domain, the sample’s transmission function does not change essentially, and for simplicity we can assume has a value of 1.0 or other given values without losing generality. Then can be computed as with the first equation of Eq. (12), and then can be determined as with the first equation of Eq. (9). According to Figs. 3(b) and 4(b), and can be computed using the computed and . For clarity, we defined , , and the intensity at the th pixel when the sample was scanned to the th position can be written as Eq. (12), where , , and are additional phase factors of , , , and caused by the th shifting of the sample, respectively. Since do not change with the sample’s positions, these three terms can be eliminated by subtracting from to yield six linear equations. Then, with these six linear equations, we can compute six unknowns: of the Eq. (12); then the values of and are computed. With the same strategy, the values of and can be computed in the next step, and all other spatial components of two sample layers can be computed in the same way. The transmission functions of the two sample layers can be computed by doing inverse Fourier transform on all computed and . To be suitable for large matrix objects, the point-by-point calculation takes a certain amount of time. However, when calculating objects with large layers, compared with the traditional 3PIE algorithm, which needs to wait for convergence, this method can compute each layer at the same time without extra time cost.
With the above computing method, two spatial components of sample slices can be computed in each step using seven linear equations, and the computer memory required was very small. Then solving Eq. (7) becomes quite easier than directly using Eq. (6). A two-layer sample was used as an example in the above analysis, and the transmission function of the volume object composed of many layers can also be computed in a similar way. When the sample was composed of three layers, four layers, and layers, the number of diffraction patterns required will be 13, 21, and , respectively.
3. Numerical Simulations
To check the feasibility of the above theoretical analysis and proposed computing method, a series of numerical simulations were carried out. Two biological images of pixels shown in Figs. 5(a) and 5(b) were used as amplitude transmissions of two layers of a volume sample. Two images shown in Figs. 5(c) and 5(d) were used as phase retardations of two layers, respectively. The interval between two layers was assumed as 1 mm. The probe light illuminating on the sample was a parallel laser beam of 632.8 nm passing through an aperture 0.7 mm in radius, and the distance of this aperture from the sample was 30 mm, equal to the distance from the sample to the detector. The amplitude and phase of illumination are shown in Figs. 5(e) and 5(f), respectively. The strength of the Fourier components of two sample layers and the illumination are shown in Figs. 5(g), 5(h) and 5(i) in log scale, respectively. When the sample was shifted by distances of (450 µm, 450 µm), (446 µm, 900 µm), (450 µm, 1320 µm), (890 µm, 450 µm), (905 µm, 920 µm), (900 m, 1361 µm), and (1330 µm, 450 µm), seven frames of diffraction patterns shown in Fig. 6 can be obtained. The pixel size of the detector was assumed to be 9 µm.
Figure 5.Object and illumination. (a) and (b) are the amplitude of two layers of object; (c) and (d) are the corresponding phase of two layers of object; (e) and (f) are the amplitude and phase of illumination; the spectra in log scale of two layers object and illumination are shown in (g)–(i), respectively. The scale bar of (a) is suitable for (b)–(d) and (g)–(i).
With the computing method discussed above, the strengths of the Fourier components of two sample layers computed with Eqs. (9)–(12) are shown in Figs. 7(a) and 7(b) in log scale, respectively. For quantitative comparison, the differences between the reconstructed images and the original images were calculated based on the formula: , shown in Figs. 7(c) and 7(d). We can find that the difference is around 0.05%, which is the computing accuracy of a common desktop of 32 bits.
Figure 7.Reconstructed spectra of two layers. (a) and (b) represent the recovered spectra of two layers, and (c) and (d) depict the difference between reconstructed spectrum and original spectrum.
By doing inverse Fourier transform on computed Fourier components in Fig. 7, we get the modulus and phase of two sample layers, shown in Figs. 8(a)–8(d). The differences of computed modulus and phase to their original values are shown in Figs. 8(e)–8(h), which are all at the scale of about , and are the computing accuracy of a common workstation of 32 bits. Results in Figs. 7 and 8 perfectly match our theoretical expectations of Eqs. (10)–(12), proving the correctness of the above theoretical analysis and suggested computing methods.
Figure 8.Reconstruction of two layers object. (a)–(d) are the amplitudes and phases of two layers object; (e)–(h) are the differences of modulus and phases to their original values.
In the above studies, we did not touch experimental factors, including detector noises, which are a kind of inevitable error source of PIE experiments. If there is random noise in the diffraction intensity, the linear equation set will become . Then, spatial components of sample slices cannot be accurately computed by directly using noisy diffraction intensities. If the sample was scanned at many positions to record a large enough number of diffraction intensities, we can add a large number of linear equations corresponding to the same detector pixel together as
When is large enough, will become close to zero, and Eq. (13) will become . Then, , without the influence of detector noise, can also be computed as . That means that, by shifting sample to more positions and recording more diffraction patterns, we can remarkably suppress the influence of detector noise and get accurate reconstruction for 3PIE with the above illustrated analytical method. As the approximation holds only when takes a small value, this method cannot be available when external noises are too large.
To verify the robustness of this anti-noise computing method, another set of simulations was carried out by adding Poisson noise to diffractive intensities, shown in Fig. 6. The sample was shifted by positions, yielding 49 frames of diffraction patterns, and Poisson noise [Fig. 9(a)] with strength between and 2 was added to each pattern, resulting in the 20 dB signal-to-noise ratio (SNR). These noisy intensities are divided into seven groups; after diffraction patterns in each group are summed up, seven frames of new hybrid diffractive intensities, shown in Figs. 9(b)–9(h), are obtained.
Figure 9.Noise and seven new diffraction patterns. (a) is the Poisson noise. (b)–(h) are seven new hybrid diffraction patterns.
With our suggested computing method, the spatial components of two sample layers were computed from hybrid diffraction patterns in Fig. (10), where Figs. 10(a) and 10(b) are the modulus of computed spectral components of the two sample layers in log scale. By doing inverse Fourier transform, the complex transmission function of the two sample layers can be obtained. Figures 10(c) and 10(d) are the modulus of the two layers, and Figs. 10(e) and 10(f) are corresponding phases. Figures 10(g) and 10(h) show the differences between the recovered modulus and their corresponding original values, respectively. The maximum difference is about 0.5%, which matches our expectation well that the influence of noise can be effectively eliminated by shifting the sample to more positions and recording more diffraction intensities.
Figure 10.Results reconstructed from noisy data. (a) and (b) are the recovered spectra in log scale; (c) and (d) are the modulus of the reconstructed two-layer object; the corresponding phase is shown in (e) and (f); amplitude differences to original image are shown in (g) and (h).
The underlying physical mechanism and mathematics of the 3PIE imaging method was revealed by writing its diffraction intensities as a linear equation set. The spatial components of all sample slices can be analytically determined using an efficient computing method to solve this linear equation set. The robustness of this suggested computing method in dealing with noisy data was also studied, and it was demonstrated that the influence of detector noise can be effectively eliminated by simply dividing many recorded intensities into groups and summing each group up. While theoretical analysis was illustrated, numerical simulations were also carried out to verify the feasibility of the proposed model and computing method by taking a two-slice thick sample as an example. This study clarified the mathematical uniqueness of the 3PIE algorithm for the first time and suggested a speedy computing method to get analytical reconstruction from diffraction intensity, breaking the theoretical bottleneck that hinders the application of 3PIE in fields of optical measurement or metrology, where mathematical uniqueness and error analysis are very crucial.
[11] R. Hegerl, W. Hoppe. Phase evaluation in generalized diffraction (ptychography). Proceedings of the Fifth European Congress Electron Microscopy(1972).