Boosting Restoration of Turbulence-Degraded Images With State Space Conditional Diffusion

被引:0
作者
Wu, Yubo [1 ]
Cheng, Kuanhong [1 ]
Chai, Tingting [2 ]
Lyu, Gengyu [3 ]
Zhao, Shuping [4 ]
Jia, Wei [5 ]
机构
[1] Xian Univ Technol, Sch Comp Sci & Engn, Xian 710048, Peoples R China
[2] Harbin Inst Technol, Fac Comp, Weihai 264209, Peoples R China
[3] Beijing Univ Technol, Coll Comp Sci, Beijing 100124, Peoples R China
[4] Guangdong Univ Technol, Sch Comp Sci & Technol, Guangzhou 510006, Peoples R China
[5] Hefei Univ Technol, Sch Comp & Informat, Hefei 230009, Peoples R China
基金
中国国家自然科学基金;
关键词
Image restoration; Frequency-domain analysis; Mathematical models; Degradation; Stochastic processes; Transformers; Noise reduction; Computer architecture; Computational modeling; Training; turbulence distortion; deep learning; DDPM; SSM; ATMOSPHERIC-TURBULENCE;
D O I
10.1109/TSP.2025.3580723
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recovering fine details from turbulence-distorted images is highly challenging due to the complex, spatially varying, and stochastic nature of the distortion process. Conventional multi-frame methods rely on extracting and averaging clear regions from pre-aligned frames, but their effectiveness is limited due to the rarity of "lucky regions". In contrast, learning based methods have shown superior performance across various vision tasks. However, existing deep learning approaches still face key limitations: (1) they struggle to efficiently model the global context required for correcting pixel dispersion caused by spatially varying Point Spread Functions (PSFs); (2) they often overlook the physical formation of turbulence, particularly the spatial-frequency relationship between phase distortions and PSFs; and (3) they rely on deterministic architectures that fail to capture the inherent uncertainty in turbulence, leading to visually implausible outputs. To address these issues, we propose the Two-Stage Turbulence Removal Network (TSTRNet). The first stage uses a UNet-based generator built on the State Space Model to perform efficient, coarse global restoration. The second stage refines the output through a Denoising Diffusion Probabilistic Model, introducing stochasticity and edge-guided conditioning for detail enhancement and realism. Both stages incorporate frequency-domain processing to align with the physical characteristics of turbulence. Experimental results on multiple benchmark datasets demonstrate that TSTRNet achieves superior restoration performance compared to state-of-the-art methods, with strong generalization from synthetic to real-world scenarios.
引用
收藏
页码:2631 / 2645
页数:15
相关论文
共 62 条
[41]  
Sohl-Dickstein J, 2015, PR MACH LEARN RES, V37, P2256
[42]  
Song J, 2021, P INT C LEARN REPR I
[43]   UU-Mamba: Uncertainty-aware U-Mamba for Cardiac Image Segmentation [J].
Tsai, Ting Yu ;
Lin, Li ;
Hu, Shu ;
Chang, Ming-Ching ;
Zhu, Hongtu ;
Wang, Xin .
2024 IEEE 7TH INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL, MIPR 2024, 2024, :267-273
[44]  
Vaswani A, 2017, ADV NEUR IN, V30
[45]   Monte Carlo-Based Restoration of Images Degraded by Atmospheric Turbulence [J].
Wang, Cong ;
Sun, Guodong ;
Wang, Cailing ;
Gao, Zixuan ;
Wang, Hongwei .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (11) :6610-6620
[46]   Non-local Neural Networks [J].
Wang, Xiaolong ;
Girshick, Ross ;
Gupta, Abhinav ;
He, Kaiming .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7794-7803
[47]   Dense Hybrid Attention Network for Palmprint Image Super-Resolution [J].
Wang, Yao ;
Fei, Lunke ;
Zhao, Shuping ;
Zhu, Qi ;
Wen, Jie ;
Jia, Wei ;
Rida, Imad .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (04) :2590-2602
[48]   Semi-supervised correction model for turbulence-distorted images [J].
Wu, Yubo ;
Cheng, Kuanhong ;
Cao, Ting ;
Zhao, Dong ;
Li, Junhuai .
OPTICS EXPRESS, 2024, 32 (12) :21160-21174
[49]  
Wu YX, 2020, INT J COMPUT VISION, V128, P742, DOI [10.1007/s11263-019-01198-w, 10.1109/CSTIC.2018.8369274]
[50]  
Yang J, 2025, IEEE T GEOSCI REMOTE, V63, DOI [10.1109/tgrs.2025.3556000, 10.1109/TGRS.2025.3556000]