Cross Modality Matching
Cross-modality matching focuses on aligning and comparing data from different sensory modalities (e.g., images and text, visible and infrared light, images and point clouds). Current research emphasizes developing robust algorithms, often leveraging contrastive learning, optimal transport, and pre-trained models like CLIP, to bridge the "modality gap" and improve cross-modal matching accuracy. This work is crucial for applications ranging from person re-identification and medical image analysis to zero-shot learning and image retrieval, enabling more powerful and versatile AI systems. Significant advancements are being made through techniques like generating homogeneous modalities and incorporating multi-granularity feature extraction.
Papers
November 7, 2024
July 17, 2024
April 27, 2024
January 11, 2024
December 16, 2023
November 25, 2023
October 5, 2023
May 17, 2023