High-quality computational ghost imaging using ADMM fused with low-rank regularization
Computational ghost imaging (CGI) has emerged as a promising technique for diverse imaging applications, particularly in challenging environments. However, achieving high-quality image reconstruction under low sampling rates and noisy conditions remains a significant challenge hindering practical de...
Saved in:
Published in | Optics and lasers in engineering Vol. 193; p. 109029 |
---|---|
Main Authors | , , , , , , |
Format | Journal Article |
Language | English |
Published |
Elsevier Ltd
01.10.2025
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Computational ghost imaging (CGI) has emerged as a promising technique for diverse imaging applications, particularly in challenging environments. However, achieving high-quality image reconstruction under low sampling rates and noisy conditions remains a significant challenge hindering practical deployment. To overcome these limitations and achieve superior reconstruction quality, we present a novel CGI method based on the Alternating direction method of multipliers (ADMM) fused with Low-rank regularization (GIAL). We also develop a fiber-based ghost imaging setup for experimental validation. Numerical simulations and experimental results validate the exceptional and general reconstruction performance of the proposed GIAL algorithm. Our findings demonstrate the algorithm's remarkable capacity to reconstruct high-quality images at extremely low sampling rates (e.g., 1.56%) and highlight its inherent robustness to noise. These superior characteristics underscore the significant potential of the GIAL method for widespread applications in biomedical imaging and remote sensing scenarios.
•Proposed ADMM with low-rank regularization for ghost imaging at 1.56% sampling rate.•Validated universal applicability across 210+ diverse images from the USC-SIPI database.•Established theoretical framework for fiber-based ghost imaging with practical experimental validation. |
---|---|
ISSN: | 0143-8166 |
DOI: | 10.1016/j.optlaseng.2025.109029 |