Attention Mechanism
Attention mechanisms are computational processes that selectively focus on relevant information within data, improving efficiency and performance in various machine learning models. Current research emphasizes optimizing attention's computational cost (e.g., reducing quadratic complexity to linear), enhancing its expressiveness (e.g., through convolutional operations on attention scores), and improving its robustness (e.g., mitigating hallucination in vision-language models and addressing overfitting). These advancements are significantly impacting fields like natural language processing, computer vision, and time series analysis, leading to more efficient and accurate models for diverse applications.
Papers
Modeling Teams Performance Using Deep Representational Learning on Graphs
Francesco Carli, Pietro Foini, Nicolò Gozzi, Nicola Perra, Rossano Schifanella
SALO: An Efficient Spatial Accelerator Enabling Hybrid Sparse Attention Mechanisms for Long Sequences
Guan Shen, Jieru Zhao, Quan Chen, Jingwen Leng, Chao Li, Minyi Guo
Balancing Exploration and Exploitation for Solving Large-scale Multiobjective Optimization via Attention Mechanism
Haokai Hong, Min Jiang, Liang Feng, Qiuzhen Lin, Kay Chen Tan
Deep Learning-based Inertial Odometry for Pedestrian Tracking using Attention Mechanism and Res2Net Module
Boxuan Chen, Ruifeng Zhang, Shaochu Wang, Liqiang Zhang, Yu Liu