Text-guided texturing by synchronized multi-view diffusion

Y Liu, M **e, H Liu, TT Wong - SIGGRAPH Asia 2024 Conference Papers, 2024 - dl.acm.org
This paper introduces a novel approach to synthesize texture to dress up a 3D object, given
a text prompt. Based on the pre-trained text-to-image (T2I) diffusion model, existing methods …

Dreammat: High-quality pbr material generation with geometry-and light-aware diffusion models

Y Zhang, Y Liu, Z **e, L Yang, Z Liu, M Yang… - ACM Transactions on …, 2024 - dl.acm.org
Recent advancements in 2D diffusion models allow appearance generation on untextured
raw meshes. These methods create RGB textures by distilling a 2D diffusion model, which …

Blenderalchemy: Editing 3d graphics with vision-language models

I Huang, G Yang, L Guibas - European Conference on Computer Vision, 2024 - Springer
Graphics design is important for various applications, including movie production and game
design. To create a high-quality scene, designers usually need to spend hours in software …

Shapegpt: 3d shape generation with a unified multi-modal language model

F Yin, X Chen, C Zhang, B Jiang, Z Zhao… - IEEE Transactions …, 2025 - ieeexplore.ieee.org
The advent of large language models, which enable flexibility through instruction-driven
approaches, has revolutionized many traditional generative tasks, but large models for 3D …

Dilightnet: Fine-grained lighting control for diffusion-based image generation

C Zeng, Y Dong, P Peers, Y Kong, H Wu… - ACM SIGGRAPH 2024 …, 2024 - dl.acm.org
This paper presents a novel method for exerting fine-grained lighting control during text-
driven diffusion-based image generation. While existing diffusion models already have the …

Meta 3d texturegen: Fast and consistent texture generation for 3d objects

R Bensadoun, Y Kleiman, I Azuri, O Harosh… - arxiv preprint arxiv …, 2024 - arxiv.org
The recent availability and adaptability of text-to-image models has sparked a new era in
many related domains that benefit from the learned text priors as well as high-quality and …

Cascade-zero123: One image to highly consistent 3d with self-prompted nearby views

Y Chen, J Fang, Y Huang, T Yi, X Zhang, L **e… - … on Computer Vision, 2024 - Springer
Synthesizing multi-view 3D from one single image is a significant but challenging task. Zero-
1-to-3 methods have achieved great success by lifting a 2D latent diffusion model to the 3D …

Mapa: Text-driven photorealistic material painting for 3d shapes

S Zhang, S Peng, T Xu, Y Yang, T Chen… - ACM SIGGRAPH 2024 …, 2024 - dl.acm.org
This paper aims to generate materials for 3D meshes from text descriptions. Unlike existing
methods that synthesize texture maps, we propose to generate segment-wise procedural …

Meta 3d gen

R Bensadoun, T Monnier, Y Kleiman… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce Meta 3D Gen (3DGen), a new state-of-the-art, fast pipeline for text-to-3D asset
generation. 3DGen offers 3D asset creation with high prompt fidelity and high-quality 3D …

Roomtex: Texturing compositional indoor scenes via iterative inpainting

Q Wang, R Lu, X Xu, J Wang, MY Wang, B Dai… - … on Computer Vision, 2024 - Springer
The advancement of diffusion models has pushed the boundary of text-to-3D object
generation. While it is straightforward to composite objects into a scene with reasonable …