Cross Modal Retrieval
Cross-modal retrieval aims to find relevant items across different data types (e.g., images and text, audio and video) by learning shared representations that capture semantic similarities. Current research focuses on improving retrieval accuracy in the face of noisy data, mismatched pairs, and the "modality gap" using techniques like contrastive learning, masked autoencoders, and optimal transport. These advancements are crucial for applications ranging from medical image analysis and robotics to multimedia search and music recommendation, enabling more effective information access and integration across diverse data sources.
Papers
April 4, 2023
March 10, 2023
February 28, 2023
February 13, 2023
January 12, 2023
January 10, 2023
December 29, 2022
December 15, 2022
December 11, 2022
November 30, 2022
November 23, 2022
November 21, 2022
November 7, 2022
October 26, 2022
October 24, 2022
October 19, 2022
October 9, 2022
September 30, 2022