Knowledge Storage
Knowledge storage in large language models (LLMs) is a burgeoning research area focused on understanding how these models encode and retrieve factual information. Current investigations explore the localization of knowledge within model architectures, examining the roles of attention mechanisms and MLP weights, and developing efficient caching strategies like RAGCache and FedCache 2.0 to improve knowledge access and reduce computational costs. These efforts aim to improve LLMs' accuracy, efficiency, and robustness, with implications for various applications including personalized federated learning and enhanced retrieval-augmented generation.
Papers
September 17, 2024
September 1, 2024
May 23, 2024
May 22, 2024
April 18, 2024
April 8, 2024
September 25, 2023
August 25, 2023