Representation Fusion
Representation fusion aims to combine information from multiple sources (e.g., images, text, sensor data) to create more comprehensive and robust representations for various machine learning tasks. Current research focuses on developing efficient fusion techniques, including attention mechanisms, transformer networks, and autoencoders, often tailored to specific data modalities and addressing challenges like data heterogeneity and missing views. These advancements are improving performance in diverse applications such as image classification, natural language processing, and healthcare monitoring, leading to more accurate and reliable models.
Papers
Layer-wise Representation Fusion for Compositional Generalization
Yafang Zheng, Lei Lin, Shuangtao Li, Yuxuan Yuan, Zhaohong Lai, Shan Liu, Biao Fu, Yidong Chen, Xiaodong Shi
SMURF: Spatial Multi-Representation Fusion for 3D Object Detection with 4D Imaging Radar
Jianan Liu, Qiuchi Zhao, Weiyi Xiong, Tao Huang, Qing-Long Han, Bing Zhu