Mamba Based Model
Mamba models are a novel class of deep learning architectures based on selective state-space models, offering a computationally efficient alternative to transformers for sequential data processing. Current research focuses on refining Mamba's architecture, exploring its application across diverse domains (including natural language processing, computer vision, and biomedical data analysis), and comparing its performance against established methods like transformers and convolutional neural networks. This work is significant due to Mamba's potential to improve the efficiency and scalability of large language models and other deep learning applications, particularly in resource-constrained environments.
Papers
November 6, 2024
November 5, 2024
October 21, 2024
October 9, 2024
October 7, 2024
October 4, 2024
September 22, 2024
September 15, 2024
September 11, 2024
August 23, 2024
August 11, 2024
August 2, 2024
July 29, 2024
July 4, 2024
June 24, 2024
June 12, 2024
May 11, 2024
April 30, 2024
April 9, 2024
March 21, 2024