Text to 3D Generation
Text-to-3D generation aims to create three-dimensional models from textual descriptions, bridging the gap between natural language and 3D content creation. Current research heavily utilizes diffusion models, often coupled with techniques like Score Distillation Sampling (SDS) and Gaussian splatting, to generate high-fidelity 3D objects represented as neural radiance fields or meshes. These advancements are improving the realism, detail, and efficiency of 3D model generation, impacting fields such as computer graphics, animation, and virtual/augmented reality by offering faster and more intuitive content creation pipelines. Ongoing efforts focus on addressing challenges like geometric consistency, view consistency, and efficient generation of complex scenes.
Papers
InterFusion: Text-Driven Generation of 3D Human-Object Interaction
Sisi Dai, Wenhao Li, Haowen Sun, Haibin Huang, Chongyang Ma, Hui Huang, Kai Xu, Ruizhen Hu
LATTE3D: Large-scale Amortized Text-To-Enhanced3D Synthesis
Kevin Xie, Jonathan Lorraine, Tianshi Cao, Jun Gao, James Lucas, Antonio Torralba, Sanja Fidler, Xiaohui Zeng
DreamFlow: High-Quality Text-to-3D Generation by Approximating Probability Flow
Kyungmin Lee, Kihyuk Sohn, Jinwoo Shin
Hyper-3DG: Text-to-3D Gaussian Generation via Hypergraph
Donglin Di, Jiahui Yang, Chaofan Luo, Zhou Xue, Wei Chen, Xun Yang, Yue Gao
Sculpt3D: Multi-View Consistent Text-to-3D Generation with Sparse 3D Prior
Cheng Chen, Xiaofeng Yang, Fan Yang, Chengzeng Feng, Zhoujie Fu, Chuan-Sheng Foo, Guosheng Lin, Fayao Liu