Multi Modal Data
Multi-modal data analysis focuses on integrating information from diverse sources, such as images, text, audio, and sensor data, to achieve more comprehensive and accurate insights than using any single modality alone. Current research emphasizes developing robust models, often based on transformer architectures and contrastive learning, that can effectively fuse these disparate data types, handle missing data, and address issues like noisy labels and modality mismatches. This field is crucial for advancing numerous applications, including medical diagnosis, urban planning, materials science, and traffic prediction, by enabling more sophisticated and reliable analyses of complex systems.
Papers
Hierarchical Semi-Supervised Learning Framework for Surgical Gesture Segmentation and Recognition Based on Multi-Modality Data
Zhili Yuan, Jialin Lin, Dandan Zhang
Bridging the Gap: Exploring the Capabilities of Bridge-Architectures for Complex Visual Reasoning Tasks
Kousik Rajesh, Mrigank Raman, Mohammed Asad Karim, Pranit Chawla
Multi-modal Graph Neural Network for Early Diagnosis of Alzheimer's Disease from sMRI and PET Scans
Yanteng Zhanga, Xiaohai He, Yi Hao Chan, Qizhi Teng, Jagath C. Rajapakse
MM-DAG: Multi-task DAG Learning for Multi-modal Data -- with Application for Traffic Congestion Analysis
Tian Lan, Ziyue Li, Zhishuai Li, Lei Bai, Man Li, Fugee Tsung, Wolfgang Ketter, Rui Zhao, Chen Zhang
Cross-Modal Vertical Federated Learning for MRI Reconstruction
Yunlu Yan, Hong Wang, Yawen Huang, Nanjun He, Lei Zhu, Yuexiang Li, Yong Xu, Yefeng Zheng