Limited Memory
Limited memory in machine learning focuses on developing efficient algorithms and architectures that can operate effectively with constrained memory resources, primarily addressing the challenges posed by increasingly large models like LLMs and deep neural networks. Current research emphasizes techniques such as memory-aware attention mechanisms, adaptive memory management strategies (e.g., dynamic caching, swapping), and model compression methods to reduce memory footprint without significant performance loss. This research is crucial for deploying advanced AI models on resource-constrained devices (e.g., edge devices, mobile phones) and for making large-scale model training more accessible.
Papers
November 3, 2024
October 22, 2024
October 19, 2024
October 14, 2024
September 8, 2024
August 23, 2024
July 16, 2024
June 12, 2024
March 19, 2024
January 30, 2024
December 23, 2023
December 19, 2023
December 12, 2023
October 25, 2023
October 10, 2023
August 29, 2023
June 17, 2023
June 16, 2023
June 13, 2023