Vision transformer empowered physics-driven deep learning for omnidirectional three-dimensional holography

The inter-plane crosstalk and limited axial resolution are two key points that hinder the performance of three-dimensional (3D) holograms. The state-of-the-art methods rely on increasing the orthogonality of the cross-sections of a 3D object at different depths to lower the impact of inter-plane cro...

Full description

Saved in:
Bibliographic Details
Published inOptics express Vol. 32; no. 8; pp. 14394 - 14404
Main Authors Jin, Zhongwei, Ren, Qiuyu, Chen, Tao, Dai, Zhigang, Shu, Fangzhou, Fang, Bin, Hong, Zhi, Shen, Changyu, Mei, Shengtao
Format Journal Article
LanguageEnglish
Published 08.04.2024
Online AccessGet full text

Cover

Loading…
More Information
Summary:The inter-plane crosstalk and limited axial resolution are two key points that hinder the performance of three-dimensional (3D) holograms. The state-of-the-art methods rely on increasing the orthogonality of the cross-sections of a 3D object at different depths to lower the impact of inter-plane crosstalk. Such strategy either produces unidirectional 3D hologram or induces speckle noise. Recently, learning-based methods provide a new way to solve this problem. However, most related works rely on convolution neural networks and the reconstructed 3D holograms have limited axial resolution and display quality. In this work, we propose a vision transformer (ViT) empowered physics-driven deep neural network which can realize the generation of omnidirectional 3D holograms. Owing to the global attention mechanism of ViT, our 3D CGH has small inter-plane crosstalk and high axial resolution. We believe our work not only promotes high-quality 3D holographic display, but also opens a new avenue for complex inverse design in photonics.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:1094-4087
1094-4087
DOI:10.1364/OE.519400