Monocular Depth Estimation
Monocular depth estimation aims to reconstruct three-dimensional scene depth from a single image, a challenging inverse problem due to the inherent loss of depth information during image formation. Current research focuses on improving accuracy and robustness, particularly in challenging scenarios like low-texture regions, viewpoint changes, and non-Lambertian surfaces, often employing deep learning models such as transformers and diffusion networks, along with techniques like multi-view rendering and radar fusion. These advancements have significant implications for various applications, including autonomous driving, robotics, and augmented reality, by enabling more accurate and reliable 3D scene understanding from readily available monocular vision data.
Papers
ECoDepth: Effective Conditioning of Diffusion Models for Monocular Depth Estimation
Suraj Patni, Aradhye Agarwal, Chetan Arora
$\mathrm{F^2Depth}$: Self-supervised Indoor Monocular Depth Estimation via Optical Flow Consistency and Feature Map Synthesis
Xiaotong Guo, Huijie Zhao, Shuwei Shao, Xudong Li, Baochang Zhang
Leveraging Near-Field Lighting for Monocular Depth Estimation from Endoscopy Videos
Akshay Paruchuri, Samuel Ehrenstein, Shuxian Wang, Inbar Fried, Stephen M. Pizer, Marc Niethammer, Roni Sengupta
Physical 3D Adversarial Attacks against Monocular Depth Estimation in Autonomous Driving
Junhao Zheng, Chenhao Lin, Jiahao Sun, Zhengyu Zhao, Qian Li, Chao Shen