Depth to Image Diffusion
Depth-to-image diffusion models leverage the power of diffusion processes to generate realistic images conditioned on depth information, aiming to improve 3D scene generation, depth estimation, and related tasks. Current research focuses on enhancing model robustness in challenging scenarios (e.g., adverse weather, specular surfaces), improving efficiency through optimized inference pipelines and fine-tuning strategies, and exploring applications in diverse areas such as robotic manipulation and video editing. These advancements are significant because they enable more accurate and efficient 3D scene understanding and manipulation, with implications for fields ranging from computer vision and robotics to digital content creation.
Papers
SceneTex: High-Quality Texture Synthesis for Indoor Scenes via Diffusion Priors
Dave Zhenyu Chen, Haoxuan Li, Hsin-Ying Lee, Sergey Tulyakov, Matthias Nießner
RichDreamer: A Generalizable Normal-Depth Diffusion Model for Detail Richness in Text-to-3D
Lingteng Qiu, Guanying Chen, Xiaodong Gu, Qi Zuo, Mutian Xu, Yushuang Wu, Weihao Yuan, Zilong Dong, Liefeng Bo, Xiaoguang Han