Accurate Single Image to 3D Using View-Specific Neural Renderer

Synthesizing a 3D model from a single 2D image is a significant challenge in computer vision and 3D modeling. Previous single image to-3D methods generate multi-view images from a single image first and then feed these images to Neural Radiance Fields (NeRF) for 3D reconstruction. Therefore, visual...

Full description

Saved in:
Bibliographic Details
Published inJournal of Multimedia Information System Vol. 11; no. 4; pp. 241 - 248
Main Authors Jun, U-Chae, Ko, Jaeeun, Hong, Kibeom
Format Journal Article
LanguageEnglish
Published 한국멀티미디어학회 31.12.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Synthesizing a 3D model from a single 2D image is a significant challenge in computer vision and 3D modeling. Previous single image to-3D methods generate multi-view images from a single image first and then feed these images to Neural Radiance Fields (NeRF) for 3D reconstruction. Therefore, visual consistency across viewpoints of these generated multi-view images directly affects the accuracy of 3D reconstruction. However, the previous methods tend to generate view-inconsistent images due to the projective ambiguity of a single image. To address the view inconsistency, we propose a viewpoint-specific learning method for single image-to-3D reconstruction using variants of NeRF. By introducing viewpoint-specific self-attention to NeRF, our method specializes the learning for viewpoints, enabling accurate 3D reconstruction even with visually discontinuous multi-view images. Experimental results demonstrate that the proposed method outperforms state-of-the-art single image-to-3D techniques by generating more accurate and coherent 3D models. KCI Citation Count: 0
ISSN:2383-7632
2383-7632
DOI:10.33851/JMIS.2024.11.4.241