Modal Representation
Modal representation focuses on effectively combining information from multiple data sources (modalities) like text, images, and audio to improve the performance of machine learning models. Current research emphasizes developing robust methods for handling missing modalities, efficiently scaling up multimodal analysis for large datasets, and improving the fusion of information from different modalities using architectures like transformers and memory networks. This work has significant implications for various applications, including emotion recognition, content moderation, and scientific discovery, by enabling more accurate and comprehensive analysis of complex data.
Papers
April 18, 2024
February 7, 2024
October 31, 2023
May 31, 2023
April 4, 2022
February 7, 2022