Wonder3d: Single image to 3d using cross-domain diffusion

X Long, YC Guo, C Lin, Y Liu, Z Dou… - Proceedings of the …, 2024 - openaccess.thecvf.com
In this work we introduce Wonder3D a novel method for generating high-fidelity textured
meshes from single-view images with remarkable efficiency. Recent methods based on the …

Generative novel view synthesis with 3d-aware diffusion models

ER Chan, K Nagano, MA Chan… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present a diffusion-based model for 3D-aware generative novel view synthesis from as
few as a single input image. Our model samples from the distribution of possible renderings …

Syncdreamer: Generating multiview-consistent images from a single-view image

Y Liu, C Lin, Z Zeng, X Long, L Liu, T Komura… - arXiv preprint arXiv …, 2023 - arxiv.org
In this paper, we present a novel diffusion model called that generates multiview-consistent
images from a single-view image. Using pretrained large-scale 2D diffusion models, recent …

Diffusion with forward models: Solving stochastic inverse problems without direct supervision

A Tewari, T Yin, G Cazenavette… - Advances in …, 2023 - proceedings.neurips.cc
Denoising diffusion models are a powerful type of generative models used to capture
complex distributions of real-world signals. However, their applicability is limited to …

Ego-exo4d: Understanding skilled human activity from first-and third-person perspectives

K Grauman, A Westbury, L Torresani… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract We present Ego-Exo4D a diverse large-scale multimodal multiview video dataset
and benchmark challenge. Ego-Exo4D centers around simultaneously-captured egocentric …

Scenescape: Text-driven consistent scene generation

R Fridman, A Abecasis, Y Kasten… - Advances in Neural …, 2024 - proceedings.neurips.cc
We present a method for text-driven perpetual view generation--synthesizing long-term
videos of various scenes solely, given an input text prompt describing the scene and camera …

Viewdiff: 3d-consistent image generation with text-to-image models

L Höllein, A Božič, N Müller… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract 3D asset generation is getting massive amounts of attention inspired by the recent
success on text-guided 2D content creation. Existing text-to-3D methods use pretrained text …

Reconfusion: 3d reconstruction with diffusion priors

R Wu, B Mildenhall, P Henzler, K Park… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract 3D reconstruction methods such as Neural Radiance Fields (NeRFs) excel at
rendering photorealistic novel views of complex scenes. However recovering a high-quality …

Expressive text-to-image generation with rich text

S Ge, T Park, JY Zhu, JB Huang - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Plain text has become a prevalent interface for text-to-image synthesis. However, its limited
customization options hinder users from accurately describing desired outputs. For example …

Direct2. 5: Diverse text-to-3d generation via multi-view 2.5 d diffusion

Y Lu, J Zhang, S Li, T Fang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Recent advances in generative AI have unveiled significant potential for the creation of 3D
content. However current methods either apply a pre-trained 2D diffusion model with the …