Multimodal Data
Multimodal data analysis focuses on integrating information from diverse sources like text, images, audio, and sensor data to achieve a more comprehensive understanding than any single modality allows. Current research emphasizes developing effective fusion techniques, often employing transformer-based architectures, variational autoencoders, or large language models to combine and interpret these heterogeneous data types for tasks ranging from sentiment analysis and medical image interpretation to financial forecasting and summarization. This field is significant because it enables more robust and accurate models across numerous applications, improving decision-making in areas like healthcare, finance, and environmental monitoring.
238papers
Papers - Page 11
October 3, 2023
October 1, 2023
September 25, 2023
September 18, 2023
September 15, 2023
September 10, 2023
August 23, 2023
Multimodal Latent Emotion Recognition from Micro-expression and Physiological Signals
Liangfei Zhang, Yifei Qian, Ognjen Arandjelovic, Anthony ZhuLarge Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages
Jinyi Hu, Yuan Yao, Chongyi Wang, Shan Wang, Yinxu Pan, Qianyu Chen, Tianyu Yu, Hanghao Wu, Yue Zhao, Haoye Zhang, Xu Han, Yankai Lin+4SUMMIT: Source-Free Adaptation of Uni-Modal Models to Multi-Modal Targets
Cody Simons, Dripta S. Raychaudhuri, Sk Miraj Ahmed, Suya You, Konstantinos Karydis, Amit K. Roy-Chowdhury
July 19, 2023