Multi Modal Learning
Multi-modal learning aims to improve machine learning performance by integrating information from diverse data sources like images, text, and audio. Current research emphasizes developing robust methods for aligning and fusing these modalities, often employing techniques like contrastive learning, latent variable models, and attention mechanisms within various architectures including transformers and generative models. This field is significant because it enables more accurate and comprehensive analyses across numerous domains, from medical diagnosis (e.g., using images and genomic data) to action recognition (e.g., combining RGB and skeletal data), improving both scientific understanding and practical applications.
Papers
July 26, 2023
July 19, 2023
June 2, 2023
May 24, 2023
May 2, 2023
March 28, 2023
December 6, 2022
September 25, 2022
September 1, 2022
July 22, 2022
May 16, 2022
March 25, 2022
March 23, 2022
March 22, 2022
March 20, 2022
March 18, 2022
January 25, 2022
January 15, 2022
December 23, 2021
December 8, 2021