Multi View Diffusion
Multi-view diffusion models are revolutionizing 3D content generation by synthesizing multiple consistent views of an object from limited input (e.g., a single image or text prompt), enabling subsequent high-fidelity 3D reconstruction. Current research emphasizes improving the quality, consistency, and efficiency of these models, often employing architectures like transformers and diffusion models, sometimes coupled with Gaussian splatting for efficient 3D representation. This work has significant implications for various fields, including computer graphics, virtual and augmented reality, and digital asset creation, by offering faster and more realistic 3D content generation from diverse input sources.
Papers
Sampling 3D Gaussian Scenes in Seconds with Latent Diffusion Models
Paul Henderson, Melonie de Almeida, Daniela Ivanova, Titas Anciukevičius
HumanSplat: Generalizable Single-Image Human Gaussian Splatting with Structure Priors
Panwang Pan, Zhuo Su, Chenguo Lin, Zhen Fan, Yongjie Zhang, Zeming Li, Tingting Shen, Yadong Mu, Yebin Liu