Text to 3D Generation
Text-to-3D generation aims to create three-dimensional models from textual descriptions, bridging the gap between natural language and 3D content creation. Current research heavily utilizes diffusion models, often coupled with techniques like Score Distillation Sampling (SDS) and Gaussian splatting, to generate high-fidelity 3D objects represented as neural radiance fields or meshes. These advancements are improving the realism, detail, and efficiency of 3D model generation, impacting fields such as computer graphics, animation, and virtual/augmented reality by offering faster and more intuitive content creation pipelines. Ongoing efforts focus on addressing challenges like geometric consistency, view consistency, and efficient generation of complex scenes.
Papers
Chasing Consistency in Text-to-3D Generation from a Single Image
Yichen Ouyang, Wenhao Chai, Jiayi Ye, Dapeng Tao, Yibing Zhan, Gaoang Wang
Text2Control3D: Controllable 3D Avatar Generation in Neural Radiance Fields using Geometry-Guided Text-to-Image Diffusion Model
Sungwon Hwang, Junha Hyung, Jaegul Choo
ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation
Zhengyi Wang, Cheng Lu, Yikai Wang, Fan Bao, Chongxuan Li, Hang Su, Jun Zhu
T2TD: Text-3D Generation Model based on Prior Knowledge Guidance
Weizhi Nie, Ruidong Chen, Weijie Wang, Bruno Lepri, Nicu Sebe