Grounded Compositional and Diverse Text-to-3D with Pretrained Multi-View
Diffusion Model
Grounded Compositional and Diverse Text-to-3D with Pretrained Multi-View
Diffusion Model
In this paper, we propose an effective two-stage approach named Grounded-Dreamer to generate 3D assets that can accurately follow complex, compositional text prompts while achieving high fidelity by using a pre-trained multi-view diffusion model. Multi-view diffusion models, such as MVDream, have shown to generate high-fidelity 3D assets using score distillation …