Repurposing diffusion-based image generators for monocular depth estimation

B Ke, A Obukhov, S Huang, N Metzger… - Proceedings of the …, 2024 - openaccess.thecvf.com
Monocular depth estimation is a fundamental computer vision task. Recovering 3D depth
from a single image is geometrically ill-posed and requires scene understanding so it is not …

Repaint123: Fast and High-Quality One Image to 3D Generation with Progressive Controllable Repainting

J Zhang, Z Tang, Y Pang, X Cheng, P Jin, Y Wei… - … on Computer Vision, 2025 - Springer
Recent image-to-3D methods achieve impressive results with plausible 3D geometry due to
the development of diffusion models and optimization techniques. However, existing image …

Generative rendering: Controllable 4d-guided video generation with 2d diffusion models

S Cai, D Ceylan, M Gadelha… - Proceedings of the …, 2024 - openaccess.thecvf.com
Traditional 3D content creation tools empower users to bring their imagination to life by
giving them direct control over a scene's geometry appearance motion and camera path …

Dginstyle: Domain-generalizable semantic segmentation with image diffusion models and stylized semantic control

Y Jia, L Hoyer, S Huang, T Wang, L Van Gool… - … on Computer Vision, 2025 - Springer
Large, pretrained latent diffusion models (LDMs) have demonstrated an extraordinary ability
to generate creative content, specialize to user data through few-shot fine-tuning, and …

VCD-Texture: Variance Alignment based 3D-2D Co-Denoising for Text-Guided Texturing

S Liu, C Yu, C Cao, W Qian, F Wang - European Conference on Computer …, 2025 - Springer
Recent research on texture synthesis for 3D shapes benefits a lot from dramatically
developed 2D text-to-image diffusion models, including inpainting-based and optimization …

I-design: Personalized llm interior designer

A Çelen, G Han, K Schindler, L Van Gool… - arXiv preprint arXiv …, 2024 - arxiv.org
Interior design allows us to be who we are and live how we want-each design is as unique
as our distinct personality. However, it is not trivial for non-professionals to express and …

MatAtlas: Text-driven Consistent Geometry Texturing and Material Assignment

D Ceylan, V Deschaintre, T Groueix, R Martin… - arXiv preprint arXiv …, 2024 - arxiv.org
We present MatAtlas, a method for consistent text-guided 3D model texturing. Following
recent progress we leverage a large scale text-to-image generation model (eg, Stable …

RoomTex: Texturing Compositional Indoor Scenes via Iterative Inpainting

Q Wang, R Lu, X Xu, J Wang, MY Wang, B Dai… - arXiv preprint arXiv …, 2024 - arxiv.org
The advancement of diffusion models has pushed the boundary of text-to-3D object
generation. While it is straightforward to composite objects into a scene with reasonable …

Consistency^ 2: Consistent and Fast 3D Painting with Latent Consistency Models

T Wang, A Obukhov, K Schindler - arXiv preprint arXiv:2406.11202, 2024 - arxiv.org
Generative 3D Painting is among the top productivity boosters in high-resolution 3D asset
management and recycling. Ever since text-to-image models became accessible for …

ConTEXTure: Consistent Multiview Images to Texture

J Ahn, S Cho, H Jung, K Hong, S Ban… - arXiv preprint arXiv …, 2024 - arxiv.org
We introduce ConTEXTure, a generative network designed to create a texture map/atlas for
a given 3D mesh using images from multiple viewpoints. The process begins with …