Eyeing3D: perceiving 3D from 2D images

The recent vision foundation models, e.g. Segment Anything Model (SAM), have shown great potential in various downstream 2D tasks. However, their adaptability to 3D vision remains an unexplored area. In this paper, we propose a novel generative framework, namely Eyeing3D, by integrating generative v...

Full description

Saved in:
Bibliographic Details
Published inIET International Conference on Engineering Technologies and Applications (ICETA 2023) Vol. 2023; pp. 120 - 121
Main Authors Peng, T., Huang, K. -Y., Lu, S. -Y., Chen, R. -A., Fu, J., Shuai, H. -H., Cheng, W. -H.
Format Conference Proceeding
LanguageEnglish
Published The Institution of Engineering and Technology 2023
Online AccessGet full text

Cover

Loading…
More Information
Summary:The recent vision foundation models, e.g. Segment Anything Model (SAM), have shown great potential in various downstream 2D tasks. However, their adaptability to 3D vision remains an unexplored area. In this paper, we propose a novel generative framework, namely Eyeing3D, by integrating generative vision models of multiple purposes (including SAM and Neural Radiance Fields) to achieve human's uncanny capability to perceive and interpret the 3D structure of a visual object, even when it is represented in a single 2D image. Particularly, a user is granted the ability to select any visual object of interest in the input 2D image with a simple click or bounding box, facilitating the reconstruction of its 3D model, with the added ability to manipulate the visual style and viewing angle. In the experiments, the effectiveness of our proposed Eyeing3D is demonstrated, showcasing improved performance in image-based 3D reconstruction tasks.
DOI:10.1049/icp.2023.3227