DIP-based anisotropic total variation regularization for image denoising

Image denoising is a fundamental problem in image processing, with the primary objective of eliminating noise while effectively preserving image details and structural features. Recent advances in convolutional neural networks (CNNs) have led to the widespread adoption of deep learning-based denoisi...

Full description

Saved in:
Bibliographic Details
Published inAdvances in continuous and discrete models Vol. 2025; no. 1; p. 126
Main Authors Pang, Zhi-Feng, ShangGuan, Yi, Jiang, Fuyang, Tai, Xue-Cheng
Format Journal Article
LanguageEnglish
Published Cham Springer International Publishing 01.12.2025
Springer Nature B.V
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Image denoising is a fundamental problem in image processing, with the primary objective of eliminating noise while effectively preserving image details and structural features. Recent advances in convolutional neural networks (CNNs) have led to the widespread adoption of deep learning-based denoising methods. However, the representative deep image prior (DIP) exhibits limitations in capturing structural features, resulting in insufficient detail preservation during denoising. To address this issue, we propose a novel denoising model that integrates DIP with anisotropic total variation (ATV) regularization, termed DIP-ATV. In the ATV term, a weighted matrix is introduced to couple the gradient operator, enhancing the regularization term’s ability to characterize local image structures. To decouple the weighted matrix and gradient operator from the ℓ 1 -norm, we employ the alternating direction method of multipliers (ADMM), decomposing the proposed model into several efficiently solvable subproblems. Experimental results demonstrate that DIP-ATV achieves superior visual quality in qualitative evaluations and outperforms state-of-the-art unsupervised denoising methods in quantitative metrics, including PSNR, SSIM, and LPIPS.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:2731-4235
1687-1839
2731-4235
1687-1847
DOI:10.1186/s13662-025-03984-y