Generative Novel View Synthesis with 3D-Aware Diffusion Models

Type: Article

Publication Date: 2023-10-01

Citations: 72

DOI: https://doi.org/10.1109/iccv51070.2023.00389

Abstract

We present a diffusion-based model for 3D-aware generative novel view synthesis from as few as a single input image. Our model samples from the distribution of possible renderings consistent with the input and, even in the presence of ambiguity, is capable of rendering diverse and plausible novel views. To achieve this, our method makes use of existing 2D diffusion backbones but, crucially, incorporates geometry priors in the form of a 3D feature volume. This latent feature field captures the distribution over possible scene representations and improves our method's ability to generate view-consistent novel renderings. In addition to generating novel views, our method has the ability to autoregressively synthesize 3D-consistent sequences. We demonstrate state-of-the-art results on synthetic renderings and room-scale scenes; we also show compelling results for challenging, real-world objects.

Locations

  • arXiv (Cornell University) - View - PDF
  • 2021 IEEE/CVF International Conference on Computer Vision (ICCV) - View

Similar Works

Action Title Year Authors
+ Generative Novel View Synthesis with 3D-Aware Diffusion Models 2023 Eric R. Chan
Koki Nagano
Matthew A. Chan
Alexander W. Bergman
Jeong Joon Park
Axel Levy
Miika Aittala
Shalini De Mello
Tero Karras
Gordon Wetzstein
+ SparseFusion: Distilling View-conditioned Diffusion for 3D Reconstruction 2022 Zhizhuo Zhou
Shubham Tulsiani
+ PDF Chat SparseFusion: Distilling View-Conditioned Diffusion for 3D Reconstruction 2023 Zhizhuo Zhou
Shubham Tulsiani
+ PDF Chat 3DGS-Enhancer: Enhancing Unbounded 3D Gaussian Splatting with View-consistent 2D Diffusion Priors 2024 Xi Liu
Chaoyi Zhou
Siyu Huang
+ DreamSparse: Escaping from Plato's Cave with 2D Frozen Diffusion Model Given Sparse Views 2023 Paul D. Yoo
Jiaxian Guo
Yutaka Matsuo
Shixiang Gu
+ Consistent-1-to-3: Consistent Image to 3D View Synthesis via Geometry-aware Diffusion Models 2023 Jianglong Ye
Peng Wang
Kejie Li
Yichun Shi
Heng Wang
+ Novel View Synthesis with Diffusion Models 2022 Daniel Watson
William Chan
Ricardo Martin-Brualla
Jonathan Ho
Andrea Tagliasacchi
Mohammad Norouzi
+ PDF Chat Consistent-1-to-3: Consistent Image to 3D View Synthesis via Geometry-aware Diffusion Models 2024 Jianglong Ye
Peng Wang
Kejie Li
Yichun Shi
Heng Wang
+ ViVid-1-to-3: Novel View Synthesis with Video Diffusion Models 2023 Jeong-gi Kwak
Erqun Dong
Yuhe Jin
Hanseok Ko
Shweta Mahajan
Kwang Moo Yi
+ NerfDiff: Single-image View Synthesis with NeRF-guided Distillation from 3D-aware Diffusion 2023 Jiatao Gu
Alex Trevithick
Kai-En Lin
Josh Susskind
Christian Theobalt
Lingjie Liu
Ravi Ramamoorthi
+ PDF Chat Skel3D: Skeleton Guided Novel View Synthesis 2024 Áron Fóthi
Bence Fazekas
Natabara Måté Gyöngyössy
Kristian Fenech
+ PDF Chat ViewCrafter: Taming Video Diffusion Models for High-fidelity Novel View Synthesis 2024 Wangbo Yu
Jinbo Xing
Yuan Li
Wenbo Hu
Xiaoyu Li
Zhipeng Huang
Xiangjun Gao
Tien‐Tsin Wong
Ying Shan
Yonghong Tian
+ PDF Chat LT3SD: Latent Trees for 3D Scene Diffusion 2024 Quan Meng
Lei Li
Matthias Nießner
Angela Dai
+ Light Field Diffusion for Single-View Novel View Synthesis 2023 Yifeng Xiong
Haoyu Ma
Shanlin Sun
Kun Han
Xiaohui Xie
+ NeuralField-LDM: Scene Generation with Hierarchical Latent Diffusion Models 2023 Seung Wook Kim
Bradley Brown
Kangxue Yin
Karsten Kreis
K Schwarz
Daiqing Li
Robin Rombach
Antonio Torralba
Sanja Fidler
+ PDF Chat NeuralField-LDM: Scene Generation with Hierarchical Latent Diffusion Models 2023 Seung Wook Kim
Bradley Brown
Kangxue Yin
Karsten Kreis
K Schwarz
Daiqing Li
Robin Rombach
Antonio Torralba
Sanja Fidler
+ PDF Chat Novel View Synthesis from a Single Image with Pretrained Diffusion Guidance 2024 Taewon Kang
Divya Kothandaraman
Dinesh Manocha
Ming–Chieh Lin
+ PDF Chat Pointmap-Conditioned Diffusion for Consistent Novel View Synthesis 2025 Thang-Anh-Quan Nguyen
Nathan Piasco
Luis RoldĂŁo
MoussĂąb Bennehar
Dzmitry Tsishkou
Laurent Caraffa
Jean‐Philippe Tarel
Roland Brémond
+ PDF Chat CAT4D: Create Anything in 4D with Multi-View Video Diffusion Models 2024 Rundi Wu
Ruiqi Gao
Ben Poole
Alex Trevithick
Changxi Zheng
Jonathan T. Barron
Aleksander Holynski
+ PDF Chat Denoising Diffusion via Image-Based Rendering 2024 Titas Anciukevicius
Fabian Manhardt
Federico Tombari
Paul Henderson

Works That Cite This (9)

Action Title Year Authors
+ PDF Chat Viewset Diffusion: (0-)Image-Conditioned 3D Generative Models from 2D Data 2023 StanisƂaw Szymanowicz
Christian Rupprecht
Andrea Vedaldi
+ iNVS: Repurposing Diffusion Inpainters for Novel View Synthesis 2023 Yash Kant
Aliaksandr Siarohin
Michael Vasilkovsky
Rıza Alp GĂŒler
Jian Feng Ren
Sergey Tulyakov
Igor Gilitschenski
+ PDF Chat Streetscapes: Large-scale Consistent Street View Generation Using Autoregressive Video Diffusion 2024 B. H. Deng
Richard Tucker
Zhengqi Li
Leonidas Guibas
Noah Snavely
Gordon Wetzstein
+ PDF Chat Recent Trends in 3D Reconstruction of General Non‐Rigid Scenes 2024 Raza Yunus
Jan Eric Lenssen
Michael Niemeyer
Yiyi Liao
Christian Rupprecht
Christian Theobalt
Gerard Pons‐Moll
Jia‐Bin Huang
Vladislav Golyanik
Eddy Ilg
+ PDF Chat State of the Art on Diffusion Models for Visual Computing 2024 Riccardo PĂČ
Yifan Wang
Vladislav Golyanik
Kfir Aberman
Jonathan T. Barron
Amit H. Bermano
Edwin P. Chan
Tali Dekel
Aleksander Holynski
Angjoo Kanazawa
+ PDF Chat TeCH: Text-Guided Reconstruction of Lifelike Clothed Humans 2024 Yangyi Huang
Hongwei Yi
Yuliang Xiu
Tingting Liao
Jiaxiang Tang
Deng Cai
Justus Thies
+ PDF Chat TADA! Text to Animatable Digital Avatars 2024 Tingting Liao
Hongwei Yi
Yuliang Xiu
Jiaxiang Tang
Yangyi Huang
Justus Thies
Michael J. Black
+ PDF Chat Consistent-1-to-3: Consistent Image to 3D View Synthesis via Geometry-aware Diffusion Models 2024 Jianglong Ye
Peng Wang
Kejie Li
Yichun Shi
Heng Wang
+ PDF Chat CombiNeRF: A Combination of Regularization Techniques for Few-Shot Neural Radiance Field View Synthesis 2024 Matteo Bonotto
Luigi Sarrocco
Daniele Evangelista
Marco Imperoli
Alberto Pretto

Works Cited by This (90)

Action Title Year Authors
+ PDF Chat Deep Stereo: Learning to Predict New Views from the World's Imagery 2016 John P. Flynn
Ivan Neulander
James Philbin
Noah Snavely
+ GANs Trained by a Two Time-Scale Update Rule Converge to a Nash Equilibrium 2017 Martin Heusel
Hubert Ramsauer
Thomas Unterthiner
Bernhard Nessler
GĂŒnter Klambauer
Sepp Hochreiter
+ Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation 2018 Liang-Chieh Chen
Yukun Zhu
George Papandreou
Florian Schroff
Hartwig Adam
+ PDF Chat HoloGAN: Unsupervised Learning of 3D Representations From Natural Images 2019 Thu Nguyen-Phuoc
Chuan Li
Lucas Theis
Christian Richardt
Yong‐Liang Yang
+ PDF Chat DeepView: View Synthesis With Learned Gradient Descent 2019 John P. Flynn
Michael Broxton
Paul Debevec
Matthew DuVall
Graham Fyffe
Ryan Overbeck
Noah Snavely
Richard Tucker
+ Generative Modeling by Estimating Gradients of the Data Distribution 2019 Yang Song
Stefano Ermon
+ PDF Chat The Unreasonable Effectiveness of Deep Features as a Perceptual Metric 2018 Richard Zhang
Phillip Isola
Alexei A. Efros
Eli Shechtman
Oliver Wang
+ PDF Chat Matterport3D: Learning from RGB-D Data in Indoor Environments 2017 Anne Lynn S. Chang
Angela Dai
Thomas Funkhouser
Maciej Halber
Matthias NieBner
Manolis Savva
Shuran Song
Andy Zeng
Yinda Zhang
+ PDF Chat DeepVoxels: Learning Persistent 3D Feature Embeddings 2019 Vincent Sitzmann
Justus Thies
Felix Heide
Matthias NieBner
Gordon Wetzstein
Michael Zollhöfer
+ Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations 2019 Vincent Sitzmann
Michael Zollhöfer
Gordon Wetzstein