Attention Mechanism
Attention mechanisms are computational processes that selectively focus on relevant information within data, improving efficiency and performance in various machine learning models. Current research emphasizes optimizing attention's computational cost (e.g., reducing quadratic complexity to linear), enhancing its expressiveness (e.g., through convolutional operations on attention scores), and improving its robustness (e.g., mitigating hallucination in vision-language models and addressing overfitting). These advancements are significantly impacting fields like natural language processing, computer vision, and time series analysis, leading to more efficient and accurate models for diverse applications.
Papers
TUNeS: A Temporal U-Net with Self-Attention for Video-based Surgical Phase Recognition
Isabel Funke, Dominik Rivoir, Stefanie Krell, Stefanie Speidel
AesPA-Net: Aesthetic Pattern-Aware Style Transfer Networks
Kibeom Hong, Seogkyu Jeon, Junsoo Lee, Namhyuk Ahn, Kunhee Kim, Pilhyeon Lee, Daesik Kim, Youngjung Uh, Hyeran Byun
Supervised Attention Using Homophily in Graph Neural Networks
Michail Chatzianastasis, Giannis Nikolentzos, Michalis Vazirgiannis
ResMatch: Residual Attention Learning for Local Feature Matching
Yuxin Deng, Jiayi Ma
A Modular Multimodal Architecture for Gaze Target Prediction: Application to Privacy-Sensitive Settings
Anshul Gupta, Samy Tafasca, Jean-Marc Odobez