3D Reconstruction
3D reconstruction aims to create three-dimensional models from various two-dimensional data sources, such as images or videos, with applications spanning diverse fields. Current research emphasizes improving accuracy and efficiency, particularly in challenging scenarios like sparse viewpoints, dynamic scenes, and occluded objects. Popular approaches utilize neural radiance fields (NeRFs), Gaussian splatting, and other deep learning architectures, often incorporating techniques like active view selection and multi-view stereo to enhance reconstruction quality. These advancements are driving progress in areas such as robotics, medical imaging, and remote sensing, enabling more accurate and detailed 3D models for various applications.
Papers
A Comparative Neural Radiance Field (NeRF) 3D Analysis of Camera Poses from HoloLens Trajectories and Structure from Motion
Miriam Jäger, Patrick Hübner, Dennis Haitz, Boris Jutzi
Farm3D: Learning Articulated 3D Animals by Distilling 2D Diffusion
Tomas Jakab, Ruining Li, Shangzhe Wu, Christian Rupprecht, Andrea Vedaldi
Deep learning-based image exposure enhancement as a pre-processing for an accurate 3D colon surface reconstruction
Ricardo Espinosa, Carlos Axel Garcia-Vega, Gilberto Ochoa-Ruiz, Dominique Lamarque, Christian Daul
DITTO-NeRF: Diffusion-based Iterative Text To Omni-directional 3D Model
Hoigi Seo, Hayeon Kim, Gwanghyun Kim, Se Young Chun
DeLiRa: Self-Supervised Depth, Light, and Radiance Fields
Vitor Guizilini, Igor Vasiljevic, Jiading Fang, Rares Ambrus, Sergey Zakharov, Vincent Sitzmann, Adrien Gaidon
End-to-End Latency Optimization of Multi-view 3D Reconstruction for Disaster Response
Xiaojie Zhang, Mingjun Li, Andrew Hilton, Amitangshu Pal, Soumyabrata Dey, Saptarshi Debroy
FineRecon: Depth-aware Feed-forward Network for Detailed 3D Reconstruction
Noah Stier, Anurag Ranjan, Alex Colburn, Yajie Yan, Liang Yang, Fangchang Ma, Baptiste Angles