DreamView: Injecting View-specific Text Guidance into Text-to-3D Generation

Type: Preprint

Publication Date: 2024-04-09

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2404.06119

Abstract

Text-to-3D generation, which synthesizes 3D assets according to an overall text description, has significantly progressed. However, a challenge arises when the specific appearances need customizing at designated viewpoints but referring solely to the overall description for generating 3D objects. For instance, ambiguity easily occurs when producing a T-shirt with distinct patterns on its front and back using a single overall text guidance. In this work, we propose DreamView, a text-to-image approach enabling multi-view customization while maintaining overall consistency by adaptively injecting the view-specific and overall text guidance through a collaborative text guidance injection module, which can also be lifted to 3D generation via score distillation sampling. DreamView is trained with large-scale rendered multi-view images and their corresponding view-specific texts to learn to balance the separate content manipulation in each view and the global consistency of the overall object, resulting in a dual achievement of customization and consistency. Consequently, DreamView empowers artists to design 3D objects creatively, fostering the creation of more innovative and diverse 3D assets. Code and model will be released at https://github.com/iSEE-Laboratory/DreamView.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ PDF Chat SeMv-3D: Towards Semantic and Mutil-view Consistency simultaneously for General Text-to-3D Generation with Triplane Priors 2024 Xiao Cai
Pengpeng Zeng
Lianli Gao
Junchen Zhu
Jiaxin Zhang
Sitong Su
Heng Tao Shen
Jingkuan Song
+ PDF Chat COMOGen: A Controllable Text-to-3D Multi-object Generation Framework 2024 Shaorong Sun
Shuchao Pang
Yazhou Yao
Xiaoshui Huang
+ PDF Chat GradeADreamer: Enhanced Text-to-3D Generation Using Gaussian Splatting and Multi-View Diffusion 2024 Trapoom Ukarapol
Kevin Pruvost
+ PDF Chat PlacidDreamer: Advancing Harmony in Text-to-3D Generation 2024 Shuo Huang
S. S. Sun
Zixuan Wang
Xiaoyu Qin
Yanmin Xiong
Yuan Zhang
Pengfei Wan
Di Zhang
Jia Jia
+ PDF Chat MVLight: Relightable Text-to-3D Generation via Light-conditioned Multi-View Diffusion 2024 Dongseok Shim
Yichun Shi
Kejie Li
H. Jin Kim
Peng Wang
+ PDF Chat DreamPolisher: Towards High-Quality Text-to-3D Generation via Geometric Diffusion 2024 Yuanze Lin
Ronald A. Clark
Philip H. S. Torr
+ PDF Chat Focus on Neighbors and Know the Whole: Towards Consistent Dense Multiview Text-to-Image Generator for 3D Creation 2024 Bonan Li
Zicheng Zhang
Xingyi Yang
Xinchao Wang
+ EfficientDreamer: High-Fidelity and Robust 3D Creation via Orthogonal-view Diffusion Prior 2023 Minda Zhao
Chaoyi Zhao
Xinyue Liang
Lincheng Li
Zeng Zhao
Zhipeng Hu
Changjie Fan
Yu Xin
+ PDF Chat Phidias: A Generative Model for Creating 3D Content from Text, Image, and 3D Conditions with Reference-Augmented Diffusion 2024 Zhenwei Wang
Tengfei Wang
Zexin He
Gerhard P. Hancke
Ziwei Liu
Rynson W. H. Lau
+ PDF Chat CompGS: Unleashing 2D Compositionality for Compositional Text-to-3D via Dynamically Optimizing 3D Gaussians 2024 Chongjian Ge
Chenfeng Xu
Yuanfeng Ji
Chensheng Peng
Masayoshi Tomizuka
Ping Luo
Mingyu Ding
Varun Jampani
Wangcheng Zhan
+ PDF Chat Sculpt3D: Multi-View Consistent Text-to-3D Generation with Sparse 3D Prior 2024 Cheng Chen
Xiaofeng Yang
Yang Fan
Chengzeng Feng
Zhoujie Fu
Chuan-Sheng Foo
Guosheng Lin
Fayao Liu
+ PDF Chat DreamStone: Image as a Stepping Stone for Text-Guided 3D Shape Generation 2023 Zhengzhe Liu
Peng Dai
Ruihui Li
Xiaojuan Qi
Chi‐Wing Fu
+ IPDreamer: Appearance-Controllable 3D Object Generation with Image Prompts 2023 Bohan Zeng
Shanglin Li
Yutang Feng
Hong Li
Sicheng Gao
Jiaming Liu
Huaxia Li
Xu Tang
Jianzhuang Liu
Baochang Zhang
+ DreamStone: Image as Stepping Stone for Text-Guided 3D Shape Generation 2023 Zhengzhe Liu
Peng Dai
Ruihui Li
Xiaojuan Qi
Chi‐Wing Fu
+ A Unified Approach for Text- and Image-guided 4D Scene Generation 2023 Yufeng Zheng
Xueting Li
Koki Nagano
Sifei Liu
Karsten Kreis
Otmar Hilliges
Shalini De Mello
+ PDF Chat BoostDream: Efficient Refining for High-Quality Text-to-3D Generation from Multi-View Diffusion 2024 Yong-Hao Yu
Shunan Zhu
Huai Qin
Haorui Li
+ Direct2.5: Diverse Text-to-3D Generation via Multi-view 2.5D Diffusion 2023 Yuanxun Lu
Jingyang Zhang
Shiwei Li
Tian Fang
David McKinnon
Yanghai Tsin
Long Quan
Xun Cao
Yao Yao
+ MVDream: Multi-view Diffusion for 3D Generation 2023 Yichun Shi
Peng Wang
Jianglong Ye
Long Mai
Kejie Li
Xiao Yang
+ PDF Chat Multi-view Image Prompted Multi-view Diffusion for Improved 3D Generation 2024 Seungwook Kim
Yichun Shi
Kejie Li
Minsu Cho
Peng Wang
+ PDF Chat Grounded Compositional and Diverse Text-to-3D with Pretrained Multi-View Diffusion Model 2024 Xiaolong Li
Jiawei Mo
Ying Wang
Chethan M. Parameshwara
Xiaohan Fei
Ashwin Swaminathan
Chris Taylor
Zhuowen Tu
Paolo Favaro
Stefano Soatto

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors