Transformer Based Architecture
Transformer-based architectures, initially developed for natural language processing, are rapidly expanding into diverse fields like computer vision and robotics, aiming to improve efficiency and accuracy in various tasks. Current research focuses on optimizing transformer models for specific applications, including developing more efficient attention mechanisms (e.g., FlashAttention), exploring alternative architectures like state-space models, and adapting transformers for resource-constrained environments. This surge in transformer applications is significantly impacting various scientific domains and practical applications, leading to advancements in areas such as image captioning, anomaly detection, and medical image analysis.
Papers
November 11, 2024
October 31, 2024
October 30, 2024
October 28, 2024
September 18, 2024
August 27, 2024
August 25, 2024
August 19, 2024
July 11, 2024
June 28, 2024
June 25, 2024
June 24, 2024
June 8, 2024
June 7, 2024
June 6, 2024
May 31, 2024
May 23, 2024
April 28, 2024
April 22, 2024