Cross View
Cross-view research focuses on bridging the significant visual discrepancies between images captured from different viewpoints, primarily aiming to improve the accuracy and robustness of tasks like geolocalization, scene understanding, and 3D reconstruction. Current research heavily utilizes deep learning models, including transformers, autoencoders, and diffusion models, often incorporating techniques like contrastive learning, bird's-eye-view transformations, and geometric constraints to align and fuse information across views. This field is crucial for advancing autonomous navigation, remote sensing, and human-computer interaction applications by enabling more reliable and efficient processing of multi-perspective data.
Papers
Geometry-guided Cross-view Diffusion for One-to-many Cross-view Image Synthesis
Tao Jun Lin, Wenqing Wang, Yujiao Shi, Akhil Perincherry, Ankit Vora, Hongdong Li
Appearance Matching Adapter for Exemplar-based Semantic Image Synthesis
Siyoon Jin, Jisu Nam, Jiyoung Kim, Dahyun Chung, Yeong-Seok Kim, Joonhyung Park, Heonjeong Chu, Seungryong Kim