Multi Level Fusion
Multi-level fusion integrates data from multiple sources or levels of representation to improve the accuracy and robustness of various tasks. Current research focuses on developing sophisticated fusion architectures, including transformers and neural networks, to effectively combine diverse data modalities (e.g., images, text, sensor data) and feature levels (e.g., low-level pixel data, high-level semantic information). This approach is proving valuable across numerous fields, enhancing performance in applications such as medical diagnosis, autonomous driving, and time series forecasting by leveraging the complementary strengths of different data types. The resulting improvements in accuracy and efficiency have significant implications for both scientific understanding and practical applications.
Papers
HiDAnet: RGB-D Salient Object Detection via Hierarchical Depth Awareness
Zongwei Wu, Guillaume Allibert, Fabrice Meriaudeau, Chao Ma, Cédric Demonceaux
HSTFormer: Hierarchical Spatial-Temporal Transformers for 3D Human Pose Estimation
Xiaoye Qian, Youbao Tang, Ning Zhang, Mei Han, Jing Xiao, Ming-Chun Huang, Ruei-Sung Lin