Decoder Only Transformer
Decoder-only transformers, a type of neural network architecture, are being extensively studied for their potential in various applications, primarily focusing on autoregressive sequence generation. Current research emphasizes improving their efficiency and capabilities, particularly addressing limitations in context length and computational complexity through techniques like optimized attention mechanisms (e.g., FlashAttention, LeanAttention) and key-value cache compression. This research is significant because it pushes the boundaries of large language models and other sequence-based tasks, impacting fields ranging from natural language processing and speech recognition to computer vision and even materials science.
Papers
January 31, 2024
January 25, 2024
January 11, 2024
December 4, 2023
November 27, 2023
November 8, 2023
October 11, 2023
October 5, 2023
September 28, 2023
August 30, 2023
April 20, 2023
January 12, 2023