Attention Mechanism
Attention mechanisms are computational processes that selectively focus on relevant information within data, improving efficiency and performance in various machine learning models. Current research emphasizes optimizing attention's computational cost (e.g., reducing quadratic complexity to linear), enhancing its expressiveness (e.g., through convolutional operations on attention scores), and improving its robustness (e.g., mitigating hallucination in vision-language models and addressing overfitting). These advancements are significantly impacting fields like natural language processing, computer vision, and time series analysis, leading to more efficient and accurate models for diverse applications.
Papers
Outlier-Efficient Hopfield Layers for Large Transformer-Based Models
Jerry Yao-Chieh Hu, Pei-Hsuan Chang, Robin Luo, Hong-Yu Chen, Weijian Li, Wei-Po Wang, Han Liu
Dissecting Query-Key Interaction in Vision Transformers
Xu Pan, Aaron Philip, Ziqian Xie, Odelia Schwartz
Part-Attention Based Model Make Occluded Person Re-Identification Stronger
Zhihao Chen, Yiyuan Ge
Enhancing Efficiency in Vision Transformer Networks: Design Techniques and Insights
Moein Heidari, Reza Azad, Sina Ghorbani Kolahi, René Arimond, Leon Niggemeier, Alaa Sulaiman, Afshin Bozorgpour, Ehsan Khodapanah Aghdam, Amirhossein Kazerouni, Ilker Hacihaliloglu, Dorit Merhof
Topological Cycle Graph Attention Network for Brain Functional Connectivity
Jinghan Huang, Nanguang Chen, Anqi Qiu
TCNet: Continuous Sign Language Recognition from Trajectories and Correlated Regions
Hui Lu, Albert Ali Salah, Ronald Poppe
COVID-19 detection from pulmonary CT scans using a novel EfficientNet with attention mechanism
Ramy Farag, Parth Upadhyay, Yixiang Gao, Jacket Demby, Katherin Garces Montoya, Seyed Mohamad Ali Tousi, Gbenga Omotara, Guilherme DeSouza