Kernel Attention
Kernel attention mechanisms enhance deep learning models by selectively weighting features based on their spatial relationships, addressing limitations of traditional attention methods that treat images as 1D sequences or suffer from high computational costs. Current research focuses on developing efficient large kernel attention modules within various architectures, including convolutional neural networks (CNNs) and transformers, for tasks such as image super-resolution, medical image segmentation, and natural language processing. These advancements improve model accuracy and efficiency across diverse applications, particularly in resource-constrained settings or those involving high-resolution data, leading to significant improvements in performance across various fields.
Papers
Self-supervised Semantic Segmentation: Consistency over Transformation
Sanaz Karimijafarbigloo, Reza Azad, Amirhossein Kazerouni, Yury Velichko, Ulas Bagci, Dorit Merhof
Beyond Self-Attention: Deformable Large Kernel Attention for Medical Image Segmentation
Reza Azad, Leon Niggemeier, Michael Huttemann, Amirhossein Kazerouni, Ehsan Khodapanah Aghdam, Yury Velichko, Ulas Bagci, Dorit Merhof