Semantic Correspondence
Semantic correspondence aims to identify corresponding elements (pixels, points, regions) in different images or 3D shapes that share the same semantic meaning, despite variations in appearance, viewpoint, or modality. Current research heavily utilizes deep learning, particularly transformer networks and diffusion models, to learn robust feature representations and establish correspondences, often incorporating techniques like contrastive learning and self-supervised training to address data scarcity and annotation challenges. This field is crucial for advancing numerous computer vision and robotics applications, including image translation, object pose estimation, 3D shape reconstruction, and video editing, by enabling more accurate and robust understanding of visual scenes and object interactions.
Papers
VideoSwap: Customized Video Subject Swapping with Interactive Semantic Point Correspondence
Yuchao Gu, Yipin Zhou, Bichen Wu, Licheng Yu, Jia-Wei Liu, Rui Zhao, Jay Zhangjie Wu, David Junhao Zhang, Mike Zheng Shou, Kevin Tang
StableVITON: Learning Semantic Correspondence with Latent Diffusion Model for Virtual Try-On
Jeongho Kim, Gyojung Gu, Minho Park, Sunghyun Park, Jaegul Choo