1National and Local Joint Engineering Research Center of Space Optoelectronics Technology, Changchun University of Science and Technology, Changchun 130022, China
2Beijing Institute of Space Mechanics & Electricity, Beijing 100094, China
3School of Computer Science and Technology, Changchun University of Science and Technology, Changchun 130022, China
The degradation of image quality in space-based remote sensing is a critical challenge due to atmospheric disturbances. In this paper, we propose a new model to simulate image blur effects caused by turbulence and aerosol scattering. It also analyzes a distortion vector field to simulate the distortion effects from atmospheric turbulence. Using this time-varying physical model, we present a generative adversarial network called MSFFA-GAN. It uses a multi-scale feature fusion and attention mechanism to analyze and apply optimal constraints on deep neural networks for atmospheric impact parameters. This helps our network handle complex atmospheric conditions that cause image degradation. Experimental results show that MSFFA-GAN improves the peak signal-to-noise ratio (PSNR) by 5.05 dB and the structural similarity index (SSIM) by 4.43%. It effectively restores degraded images and enhances the image quality of remote sensing systems.
【AIGC One Sentence Reading】:MSFFA-GAN restores remote sensing images by simulating atmospheric disturbances, improving PSNR and SSIM.
【AIGC Short Abstract】:This paper addresses image degradation in space-based remote sensing due to atmospheric disturbances. A new model simulates blur and distortion effects from turbulence and aerosol scattering. A generative adversarial network, MSFFA-GAN, utilizes multi-scale feature fusion and attention mechanisms to optimize constraints for atmospheric parameters, enhancing image restoration. Results show significant improvements in PSNR and SSIM, demonstrating effective image quality enhancement.
Note: This section is automatically generated by AI . The website and platform operators shall not be liable for any commercial or legal consequences arising from your use of AI generated content on this website. Please be aware of this.
During the space-based remote sensing long-distance imaging process, the impact of the atmospheric environment leads to a decline in the image quality captured by the remote imaging system, specifically manifested as increased image blur and geometric distortion. This degradation has a particularly negative impact on computer vision tasks like object recognition and detection, significantly reducing processing accuracy and efficiency.
In recent years, there has been a lot of focus on the impact of atmospheric turbulence and aerosols on image quality. Many methods have been proposed to address the blurring caused by turbulence. Carhart et al. improved the frame selection method using short-exposure images[1]. Potvin et al. simulated realistic turbulent imagery using a point spread function (PSF)[2]. Valley et al. proposed a modulation transfer function (MTF) with tilt correction to predict Strehl ratio and resolution degradation[3]. Burckel et al. enhanced the simulation accuracy of phase screens using logarithmic sampling of filtered white noise, reducing memory requirements[4]. Reinhardt et al. proposed an algorithm to extract atmospheric turbulence and extinction parameters from passive images[5]. Hunt et al. used statistical methods to describe resolution variations caused by turbulence[6]. Schwartzman et al. developed a two-dimensional (2D) image deformation method to render turbulence distortion[7]. Wu et al. developed an image degradation model incorporating turbulence and aerosols, optimizing the atmospheric MTF curve[8]. Sadot et al. studied the influence of aerosols on MTF, finding that aerosols dominate at high spatial frequencies[9]. Hu et al. built a PSF model that incorporates aerosol properties to analyze the blurring effect on remote sensing images[10].
Deep learning has also made significant progress in image processing, especially through the use of convolutional neural networks for feature extraction. Researchers have proposed various deep learning-based methods to mitigate atmospheric turbulence[11–14]. Barbastathis et al. used deep learning and computational imaging to address ill-posed problems[15]. Chimitt et al. simulated turbulence quickly using Zernike coefficients[16]. Hoffmire et al. proposed a restoration method for long-distance turbulence-induced deformation[17]. Cheng et al. used a generative adversarial network (GAN) to simulate and restore turbulent images[18]. Jin et al. developed TSR-WGAN, which generated high-quality image sequences without assumptions and performed exceptionally well[19].
Sign up for Chinese Optics Letters TOC Get the latest issue of Advanced Photonics delivered right to you!Sign up now
First, in this paper, we propose a new time-variant physical degradation model for remote sensing image atmospheric impact. This model accounts for both the turbulence short-exposure function and the single scattering aerosol function to simulate image blur effects. Additionally, we introduce the Schwartzman distortion vector field to model the distortion effects caused by atmospheric turbulence. Furthermore, we develop a multi-scale feature fusion and attention-based generative adversarial network (MSFFA-GAN) to optimize constraints on deep neural networks for atmospheric parameters in the physical degradation model. This network effectively restores images degraded by atmospheric temporal physical processes through multi-scale fusion. The encoder of the generator is built upon the Inception-ResNet-v2 architecture, incorporating multi-scale feature extraction units that work in tandem to enhance the encoder’s functionality. The decoder, in turn, aligns features from multiple scales to a common resolution through up-sampling while applying spatial weights to different levels of the feature maps during the fusion process. The discriminator employs a Markov discriminator (PatchGAN) structure to further enhance the network’s performance in distinguishing between real and restored images. Finally, by utilizing the time-varying physical degradation model of atmospheric impact, we generate synthetic image datasets degraded by turbulence and aerosols. These datasets are then input into the network for adversarial training, allowing the generator and discriminator to collaboratively enhance the image restoration performance. To validate the model, experimental imaging of real atmospheric turbulence and aerosols was conducted in the Jingyue region of Changchun City. During this process, real-time monitoring of vital atmospheric parameters, such as the atmospheric structure constant and aerosol extinction coefficient, was performed under actual atmospheric turbulence conditions. The resulting dataset was used to evaluate the restoration capability of the proposed network on real-world images degraded by atmospheric turbulence. The overall structure block diagram is shown in Fig. 1.
2.1. Atmospheric impact time-varying physical degradation model
Atmospheric turbulence and aerosols were the main contributors to the degradation of imaging quality in remote sensing systems. The effects of turbulence and aerosols on imaging quality can be modeled as where represents the image blur effect, represents the image jitter effect, represents the degraded image, and represents the original image. The image blur can be modeled as where and represent the optical transfer functions of atmospheric turbulence and the optical system, respectively, and are expressed as where represents the spatial frequency, represents the wavelength, represents the telescope focal length, and represents the telescope aperture diameter. When , the above equation represents the long-exposure atmospheric optical transfer function. When , it represents the atmospheric optical transfer function under short-exposure conditions. represents the optical cutoff frequency, represents the F-number, and represents the atmospheric coherence length, given by the following equation: where represents the zenith angle and denotes the atmospheric refractive index structure constant.
represents the point spread function of aerosol particles, which is expressed as where represents the single scattering albedo, , , is the zenith angle, and is the angle between the scattering light propagation direction and the vertical direction. is the scattering phase function, which describes the scattering intensity distribution of particles at different angles. According to the Mie scattering theory, the scattering phase function of aerosol particles can be expressed as where represents the scattering cross-sectional area, and denote the complex conjugates. represents the aerosol optical thickness, which describes the degree of attenuation of light radiation during transmission through the aerosol medium. It can be calculated by integrating the aerosol extinction coefficient over direction, where represents the aerosol extinction coefficient and represents the distance in the vertical direction.
Based on physical principles, Schwartzman’s method directly generates a 2D random deformation vector field for image distortion. where and are the distortion motion vectors of two pixels, and represents the distortion motion vector function.
2.2. Short-exposure atmospheric turbulence and aerosol image degradation simulation
Using this physical model, the generation of short-exposure atmospheric turbulence and aerosol-degraded images was simulated, with the simulation parameters provided in Table 1.
Atmospheric turbulence leads to independent effects of image blur and jitter, which can be calculated separately. With a fixed wavelength and a focal length, various values for the atmospheric refractive index structure constant and aerosol optical thickness were set, and the target image was input into the simulation model. It was observed that under strong turbulence, the resulting degraded image became nearly indistinguishable, while under weaker turbulence, the image remained discernible but exhibited reduced contrast and blurriness, as shown in Fig. 2. Furthermore, turbulence-induced distortion was observed in the target image. A physical model based on a 2D random deformation vector field was used to simulate these distortions under varying turbulence intensities, as shown in Fig. 3.
To investigate the impact of atmospheric aerosols on long-range imaging, simulations were conducted to observe the degradation effects of different aerosol optical thickness values on the image, as shown in Fig. 4.
Figure 2.Image distortion caused by different turbulence intensities.
Variations in aerosol optical thickness strongly affect the range of influence of the atmospheric aerosol PSF, as we can observe from the calculated 2D spatial distribution map. As the aerosol optical thickness increases, the number of pixels affected by the PSF increases, leading to a greater degradation of image quality in the spatial domain.
2.3. Comparison of field trial degraded images and simulated degraded images
To validate the effectiveness of the atmospheric turbulence and aerosol image degradation simulation model, we conducted field trials on July 30 and August 1, 2024, at a 60-m-high rooftop platform in the National Science Park of Jingyue District, Changchun. In support of the field trials, Henghui Optoelectronics provided extensive instrumentation, range sponsorship, and coordination of the measurement program. The trials employed a NexStar 8SE telescope paired with a MER2-230-168U3C camera for image acquisition of rooftop targets at 3.3 km distance and streetlights at 8.1 km distance. Simultaneous atmospheric parameter monitoring was implemented using the HMTM-2 temperature fluctuation meter and the MSLR-200 dual-band aerosol lidar. Figure 5 shows the specific setup of the field trials.
Figure 5.Field trials in the Jingyue region of Changchun City.
During data acquisition, the temperature fluctuation meter continuously recorded the atmospheric refractive index structure constant at 5-s intervals. The lidar system executed sector scans in 10° increments across zenith angles from 10° to 90°, while conducting detection every minute at 15-m vertical intervals. To mitigate solar interference, each observation angle maintained 24-h continuous monitoring, ensuring continuous acquisition of aerosol extinction coefficients.
As illustrated in Fig. 6, turbulence intensity demonstrates distinct diurnal patterns, with periodic peak values occurring during daytime hours from 11:00 to 15:00 that correlate with substantial deterioration of imaging quality. Conversely, nocturnal observations from 20:00 to 07:00 exhibit markedly reduced turbulence intensity, resulting in minimal degradation of the imaging system performance. Figure 7(a) reveals that under clear-sky conditions on July 30, the atmosphere displays homogeneous distribution characteristics accompanied by a stable vertical profile of near-surface aerosol extinction coefficients. However, the observed distribution of thin clouds within the altitude layer from 7.5 to 13.5 km resulted in an aerosol extinction coefficient peak reaching approximately . Figure 7(c) shows that under cloudy conditions on August 1, there are larger fluctuations in the extinction coefficient within the 0.4–0.8 km altitude layer, with the enhancement of cloud coverage being the primary contributing factor. Comparative analysis of aerosol extinction coefficient temporal variations at 0.06, 0.3, and 0.6 km altitude levels in Figs. 7(b) and 7(d) quantitatively confirms altitude-dependent aerosol distribution characteristics.
Figure 6.Atmospheric refractive index structure constants at different times of the day. (a) July 30, 2024. (b) August 1, 2024.
Figure 7.Aerosol extinction coefficient data. (a) Aerosol extinction coefficients at different heights on July 30, 2024, at 08:00, 11:00, 12:00, 13:00, 14:00, 15:00, and 18:00. (b) Aerosol extinction coefficients were measured at 0.06, 0.3, and 0.6 km on July 30, 2024, at various times. (c) Aerosol extinction coefficients at different heights on August 1, 2024, at 08:00, 11:00, 12:00, 13:00, 14:00, 15:00, and 18:00. (d) Aerosol extinction coefficients were measured at 0.06, 0.3, and 0.6 km on August 1, 2024, at various times.
By comparing the actual captured images with the simulated images, the performance of the model can be qualitatively analyzed. When performing the simulation, the simulation parameters must correspond to the atmospheric parameters of the actual captured images. Based on the capture time of the actual images, along with the corresponding atmospheric refractive index structure constant and aerosol extinction coefficient, the simulation parameters are set as outlined in Table 2.
When the reference target is a streetlight situated 8.1 km from the observation point, simulations are performed based on the atmospheric refractive index structure parameters and the aerosol optical thickness corresponding to real degraded images at various times, as shown in Fig. 8.
Figure 8(b) shows an image captured on July 30, 2024, at 13:19:18. The relevant atmospheric parameters are as follows: the atmospheric refractive index structure constant is , and the aerosol optical thickness is 0.248. It can be seen that in the areas marked by the red, orange, and yellow boxes, strong turbulence causes significant distortion of traffic signs, windows, and prohibition signs in the image. In addition to distortion, turbulence and aerosols also cause blurring of the image.
Figure 8.Comparison of measured and simulated degraded images of streetlights in the Jingyue region. (a) Simulation of atmospheric parameters at different times. (b) Comparison of measured and simulated degraded images with high turbulence intensity and aerosol thickness.
Figure 9(b) shows an image captured on August 1, 2024, at 13:44:46. The atmospheric refractive index structure constant is , and the aerosol optical thickness is 0.188. As illustrated in the figure, the white line within the orange box is visibly distorted, and the image shows noticeable blurring.
To perform a more quantitative analysis of the effects of atmospheric turbulence and aerosols on image quality, this work compares the quality of clear images, actual degraded images obtained from tests, and simulated degraded images. The comparison focuses on parameters closely related to image quality, such as resolution and contrast, with particular emphasis on resolution as an indicator to evaluate the degradation of images before and after atmospheric effects.
Figure 10 compares the clarity and contrast of clear, actual degraded, and simulated degraded images under four sets of different atmospheric parameters. The results show that, compared to the clear images, both the clarity and contrast of the actual and simulated degraded images decrease, with a consistent trend observed. When the atmospheric parameters are and , the difference in clarity between real and simulated images is minimal. When the atmospheric parameters are and , the difference in contrast between them is the smallest. It is concluded that this physical model can effectively simulate the degradation process of images under actual atmospheric conditions.
Figure 11 illustrates the proposed generator network structure, which utilizes the pre-trained Inception-ResNet-V2[20] model as the backbone of the encoder, combined with a multi-scale feature extraction (MSFE) unit. These two components collaboratively form the functional framework of the encoder. Following down-sampling, the encoder produces feature maps at different scales. High-level feature maps exhibit a larger receptive field and stronger semantic representation but with lower resolution. In contrast, low-level feature maps have a smaller receptive field and higher resolution, though they provide weaker semantic representation. To reconcile this, a convolutional up-sampling layer is applied to rescale the feature maps to a uniform size. The fused feature maps undergo further up-sampling, and the generated output image is then passed to the discriminator for adversarial training against the initial image.
Figure 12 illustrates how the MSFE unit incorporates multi-scale feature fusion and attention mechanisms. By employing a multi-branch convolutional layer structure with varying receptive field sizes, the multi-scale feature fusion stage effectively handles the diverse feature information contained within images exhibiting different levels of degradation. An attention mechanism module is added to highlight important image parts and improve learning ability. The attention mechanisms refine feature maps by emphasizing relevant channels and spatial regions, leading to better data understanding and representation. To preserve crucial information and prevent vanishing gradients, a residual connection merges the input feature map with the processed feature map from the MSFE unit.
The attention mechanism module comprises channel and spatial attention mechanisms. The channel attention mechanism module compresses global spatial information into channel descriptors by generating channel statistics through global average pooling[21]. This descriptor embeds the global distribution of the channel feature responses, allowing information from the network’s global receptive field to be utilized by all layers, where represents the value of the input feature at channel and position , while the output represents the one-dimensional feature corresponding to channel c.
To utilize the aggregated information, an excitation operation is applied to fully capture the channel dependencies. A simple gating mechanism with a sigmoid activation function is employed, where the embedding is used as input to generate a set of modulation weights for each channel, where denotes the ReLU activation function, and represents the Sigmoid activation function. To constrain model complexity and improve generalization, we parameter the gating mechanism by introducing a bottleneck with two fully connected layers around the non-linearity, specifically a dimensionality reduction layer with a reduction ratio denoted as and . Finally, an element-wise multiplication is performed between the output of the gating mechanism and the input feature , producing the output of the channel attention,
Unlike the channel attention mechanism, the spatial attention mechanism focuses on the “spatial locations” of the relevant information in the feature map[22]. First, global max pooling and global average pooling are applied along the channel axis to the input feature , producing the 2D features and , respectively,
After stacking the obtained 2D features, they are passed through a convolutional layer to capture contextual information in the spatial domain. The output is then passed through a Sigmoid activation function to generate the spatial attention weight matrix . Finally, an element-wise multiplication is performed between the attention weights and the input feature , assigning different levels of importance to different spatial locations, thereby enhancing the model’s ability to understand and represent the input data.
The discriminator adopts a Markov discriminator structure[23], as shown in Fig. 13. PatchGAN maps the input image to an matrix and uses a fixed-size local receptive field (referred to as a “patch”) to assess different regions of the image. Since the size of the local receptive field is fixed, the discriminator does not need to know the dimensions of the entire image, making PatchGAN highly flexible and capable of handling images of varying sizes. represents the discriminator’s output for a small patch of the input image, indicating the probability that the patch is a true sample, with values between 0 and 1. The final output of the discriminator is obtained by averaging all the values.
Selecting an appropriate loss function is critical for optimizing neural network performance. In standard GAN frameworks, the discriminator typically employs a classifier with Sigmoid cross-entropy loss, which may lead to vanishing gradient issues during training. To address this, the least-squares loss variant modifies the discriminator’s objective to penalize the squared difference between its predictions and target values (1 for real data and 0 for generated samples),
The generator aims to minimize the discrepancy between the discriminator’s prediction for its output and the ideal value of 1,
Content loss combines mean squared error loss () and perceptual loss (). The mean square error measures the average squared difference, pixel by pixel, between the original image and the image corrected by the network. A pre-trained VGG19 network calculates the perceptual loss. Both the generated and clear images are processed by the VGG19 network, and their corresponding feature maps are used to compute the Euclidean distance, where represents the number of input channels of the image, and and denote the width and height of the image, respectively. is the number of feature layers, represents the clear image, represents the image reconstructed by the generator, and denotes the feature representation at the th layer of the pre-trained neural network. Therefore, the total loss function is composed of the adversarial loss and the content loss, where , , and are weight coefficients, set to 0.01, 0.5, and 0.01, respectively.
4. Experiment and Discussion
The dataset used for training and testing our network architecture is described at the beginning of this section. Following that, the parameters needed for training are outlined. After outlining our experimental design, we benchmark our approach against various leading methods.
4.1. Dataset preprocessing
A large dataset is crucial for training GANs to maintain stability and achieve optimal performance. To overcome the challenge of obtaining clear images in real-world conditions, this study utilizes simulations to create the necessary dataset. As shown in Fig. 14, 137 clear images were selected from the RS_C11 remote sensing dataset[24], and used to synthesize degraded images using the time-varying physical degradation model for atmospheric effects. Blurry images with varying levels of degradation were produced by altering parameters, including the atmospheric refractive index structure constant, the atmospheric coherence length, and the aerosol optical thickness. A total of 2466 training images were generated, with 2088 used for training and 378 reserved for testing the network’s performance.
4.2. Training parameters
The training was conducted using the PyTorch 1.13.0 deep learning framework and an NVIDIA GeForce RTX 4090 GPU. The training process employed the Adam optimizer, a method that incorporates momentum and RMSProp. It adjusts the learning rate for each parameter by estimating both the first and second moments of the gradients, enabling more efficient network training. The two hyperparameters of the Adam optimizer were set as follows: and , with a batch size of 8. The initial learning rates for the generator and discriminator were set to 0.001 and 0.005, respectively, and were reduced by a factor of 10 every 25 epochs. In total, the learning rate was decreased 4 times, and training was terminated after 100 epochs. The training process took 3.59 h.
4.3. Experimental results and analysis
After the model training was completed, this study applied the generator network to degraded image restoration tasks. To validate the performance advantages of the proposed algorithm, we compared it with various advanced image restoration models including VDSR[25], MemNet[26], MSICF[27], and EDSR[28]. The experiments employed the peak signal-to-noise ratio (PSNR) and the structural similarity index (SSIM) as objective evaluation metrics to quantify image reconstruction quality. Additionally, the computational efficiency of each algorithm was evaluated based on the number of model parameters, floating-point operations (FLOPs), and single inference time. The experimental results shown in Table 3 comprehensively demonstrate the performance differences and comparative characteristics of various methods.
Table 3. Performance Comparison of Algorithms for Degraded Image Recovery
Table 3. Performance Comparison of Algorithms for Degraded Image Recovery
Method
Param/MB
FLOPs
Time/s
PSNR/dB
SSIM/%
VDSR
2.66
174.25G
154.91
18.91
69.12
MemNet
11.64
759.37G
9658.94
18.75
67.20
MSICF
14.96
1.12T
2943.38
18.82
66.40
EDSR
172.28
13.17T
1463.44
18.83
61.73
MSFFA-GAN
323.75
706.76G
122.66
22.65
81.42
Degraded image
—
—
103.86
17.60
76.99
As shown in the quantitative comparison in Table 3, the MSFFA-GAN achieves the best multi-dimensional balance in image restoration tasks. The proposed method demonstrates superior restoration capabilities, with a PSNR of 22.65 dB and an SSIM of 81.42%, both significantly outperforming other comparative methods. In terms of computational efficiency, the MSFFA-GAN requires only 122.66 s to complete processing, which is much faster than the EDSR, the MSICF, and the MemNet, while maintaining excellent restoration quality. The model achieves an optimal balance between performance and efficiency, with 323.75M parameters and 706.76G FLOPs. It avoids the high computational cost of the EDSR while outperforming lightweight models like the VDSR and the MemNet. These advantages are attributed to the multi-scale feature fusion strategy, which allows the model to analyze features across various scales and is particularly beneficial for handling objects of diverse sizes and complexities. Additionally, the integrated channel attention mechanism captures the interactions between the feature channels while the spatial attention mechanism determines the prominence of different parts of the image. The synergy between these attention mechanisms enables a more thorough understanding and utilization of image features, allowing the model to better present these features.
Figure 15 compares the restored images generated by different algorithms. Figure 15(a) displays the original clear image as the reference for restoration. Figures 15(c) to 15(f) present the results processed by the VDSR, the MemNet, the MSICF, and the EDSR, respectively, while Fig. 15(g) shows the output of the proposed MSFFA-GAN. The proposed algorithm significantly improves the restoration of the blurred images, producing outputs with enhanced clarity and richer edge details that closely resemble the original image. The objective metrics listed in Table 3 confirm the subjectively perceived visual improvements.
Figure 15.Comparison of experimental results of different algorithms.
To systematically evaluate model stability and robustness, we performed 25, 50, 75, and 100 independent trials on the test set. The means and standard deviations of both the PSNR and the SSIM were computed for each trial configuration, with comprehensive results summarized in Table 4.
Table 4. Performance Statistics Across Different Trials
Table 4. Performance Statistics Across Different Trials
Trial
Time/h
PSNR/dB
Std/dB
SSIM
Std
25
0.68
22.620067
0.030980
0.814127
0.000341
50
1.45
22.616801
0.027515
0.814101
0.000352
75
2.17
22.614375
0.026441
0.814075
0.000324
100
2.90
22.614152
0.026673
0.814090
0.000336
Experimental analysis reveals that the computation time increases linearly from 0.68 to 2.90 h as the number of experiments increases. Both quality metrics exhibit remarkable stability: the PSNR maintains 22.614–22.620 dB () and the SSIM remains at 0.814 (), demonstrating exceptional model robustness and reproducibility.
To test the algorithm’s ability to restore images impacted by real atmospheric effects, we gathered a dataset of degraded images captured under actual atmospheric conditions using a NexStar 8SE telescope and a MER2-230-168U3C camera. The images, influenced by atmospheric distortion, served as input for the network’s evaluation. The results are shown in Fig. 16.
Figure 16.Recovery effects of real atmospheric turbulence degradation images.
Based on the comparison between the restored and degraded images, there is a significant improvement in overall clarity, with better preservation and presentation of local details. The experimental results further validate the restoration performance of the proposed model, demonstrating its feasibility and effectiveness in image restoration.
5. Conclusion
In this paper, we propose a model that analyzes how the physical degradation of atmospheric effects evolves over time. This time-variant model is not limited to processes caused by atmospheric turbulence and aerosols in short-exposure conditions. It can also model distortion effects under a range of atmospheric turbulence scenarios. Meanwhile, we conducted field trials to get the dataset and briefly analyzed important parameters, such as the atmospheric refractive index structure constant, the aerosol extinction coefficient, and the vector field. Comparing simulated and real degraded images in an experimental analysis confirmed the model’s effectiveness.
Moreover, in this work, we present a generative adversarial network called MSFFA-GAN that can be coupled with the time-varying physical model mentioned above. This framework leverages receptive fields of different sizes for multi-scale spatial feature extraction, while a channel attention mechanism dynamically adjusts channel weights in feature maps. Through a spatial attention mechanism, the model can prioritize important spatial features, allowing for more effective restoration of degraded images. Compared to images degraded by turbulence, the proposed network enhances the PSNR by 5.05 dB and the SSIM by 4.43%. Notably, the proposed model applies to moderate and weak atmospheric turbulence conditions. When the atmospheric refractive index structure constant falls within the range of and the aerosol optical thickness is configured with specific parametric combinations of 0.048, 0.148, and 0.300, it enables the generation of image datasets exhibiting varying degrees of degradation. Additionally, the model training relies on the residential area subset of the RS-C11 dataset, which demonstrates robust adaptability to remote sensing images in this specific scenario.
In future research, we plan to incorporate more diversified datasets that encompass various scenarios and environmental conditions of remote sensing imagery. This strategic enhancement aims to improve the model’s generalization capacity and strengthen its applicability across broader operational scenarios.
[7] A. Schwartzman, M. Alterman, R. Zamir et al. Turbulence-induced 2d correlated image distortion. 2017 IEEE International Conference on Computational Photography (ICCP), 1(2017).
[8] P.-F. Wu, S. Fang, Q.-S. Xu et al. Restoration of blurred image based on atmospheric MTF. J. Atmosph. Environ. Opt., 6, 196(2011).
[10] X. L. Hu, X. F. Gu, T. Yu et al. Effects of aerosol optical thickness on the optical remote sensing imaging quality. Spectrosc. Spect. Anal., 34, 735(2014).
[13] R. Yasarla, V. M. Patel. Learning to restore images degraded by atmospheric turbulence using uncertainty. 2021 IEEE International Conference on Image Processing (ICIP), 1694(2021).
[14] N. G. Nair, V. M. Patel. Confidence guided network for atmospheric turbulence mitigation. 2021 IEEE International Conference on Image Processing (ICIP), 1359(2021).
[18] J. Cheng, J. Li, C. Dai et al. Research on atmospheric turbulence-degraded image restoration based on generative adversarial networks. First International Conference on Spatial Atmospheric Marine Environmental Optics (SAME 2023), 37(2023).
[20] C. Szegedy, S. Ioffe, V. Vanhoucke et al. Inception-v4, inception-resnet and the impact of residual connections on learning. Proceedings of the AAAI Conference on Artificial Intelligence, 31(2017).
[21] J. Hu, L. Shen, G. Sun. Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 7132(2018).
[22] S. Woo, J. Park, J. Y. Lee et al. Cbam: Convolutional block attention module. Proceedings of the European Conference on Computer Vision (ECCV), 3(2018).
[23] P. Isola, J. Y. Zhu, T. Zhou et al. Image-to-image translation with conditional adversarial networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1125(2017).
[25] J. Kim, J. K. Lee, K. M. Lee. Accurate image super-resolution using very deep convolutional networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1646(2016).
[26] Y. Tai, J. Yang, X. Liu et al. Memnet: A persistent memory network for image restoration. Proceedings of the IEEE International Conference on Computer Vision, 4539(2017).
[28] B. Lim, S. Son, H. Kim et al. Enhanced deep residual networks for single image super-resolution. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 136(2017).
Tools
Get Citation
Copy Citation Text
Xinyi Qin, Hui Li, Yan Lou, Yongli Hu, Yunbiao Liu, Wenxuan Lü, "Remote sensing image restoration via atmospheric impact time-varying degraded physical models using neural networks," Chin. Opt. Lett. 23, 080101 (2025)