Transformer Based Model
Transformer-based models are a class of neural networks achieving state-of-the-art results across diverse fields by leveraging self-attention mechanisms to capture long-range dependencies in sequential data. Current research focuses on addressing limitations such as quadratic computational complexity for long sequences, leading to the development of alternative architectures like Mamba and modifications such as LoRA for efficient adaptation and inference. These advancements are significantly impacting various applications, from speech recognition and natural language processing to computer vision and time-series forecasting, by improving both accuracy and efficiency on resource-constrained devices.
Papers
March 19, 2022
February 24, 2022
February 21, 2022
February 17, 2022
February 13, 2022
February 4, 2022
February 1, 2022
January 25, 2022
January 24, 2022
January 23, 2022
January 19, 2022
January 11, 2022
December 30, 2021
December 19, 2021
December 14, 2021
December 10, 2021
December 7, 2021
November 27, 2021