Modal Adapter
Modal adapters are lightweight modules designed to enhance pre-trained multimodal models, such as CLIP, without requiring extensive retraining. Current research focuses on developing efficient adapter architectures, often employing attention mechanisms or bottleneck structures, to improve performance on various downstream tasks like image classification, video retrieval, and machine translation by effectively fusing visual and textual information. This approach offers significant advantages in terms of parameter efficiency, reduced training costs, and improved generalization across different datasets and modalities, making it a valuable tool for advancing multimodal learning and its applications.
Papers
November 9, 2024
October 18, 2024
October 15, 2024
September 3, 2024
August 13, 2024
July 3, 2024
May 26, 2024
April 19, 2024
April 13, 2024
February 20, 2024
January 22, 2024
October 19, 2023
September 11, 2023
August 3, 2023
May 18, 2023
May 12, 2023
May 9, 2023
January 19, 2023