Back to Search Start Over

Novel View Synthesis from a Single Image with Pretrained Diffusion Guidance

Authors :
Kang, Taewon
Kothandaraman, Divya
Manocha, Dinesh
Lin, Ming C.
Publication Year :
2024

Abstract

Recent 3D novel view synthesis (NVS) methods are limited to single-object-centric scenes generated from new viewpoints and struggle with complex environments. They often require extensive 3D data for training, lacking generalization beyond training distribution. Conversely, 3D-free methods can generate text-controlled views of complex, in-the-wild scenes using a pretrained stable diffusion model without tedious fine-tuning, but lack camera control. In this paper, we introduce HawkI++, a method capable of generating camera-controlled viewpoints from a single input image. HawkI++ excels in handling complex and diverse scenes without additional 3D data or extensive training. It leverages widely available pretrained NVS models for weak guidance, integrating this knowledge into a 3D-free view synthesis approach to achieve the desired results efficiently. Our experimental results demonstrate that HawkI++ outperforms existing models in both qualitative and quantitative evaluations, providing high-fidelity and consistent novel view synthesis at desired camera angles across a wide variety of scenes.<br />Comment: 6 pages, 7 figures

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2408.06157
Document Type :
Working Paper