Memorization Effect
Memorization, the tendency of large language models (LLMs) and other deep learning models to reproduce training data verbatim, is a significant research area focusing on quantifying its extent, understanding its mechanisms (including the roles of attention and cross-attention), and developing mitigation strategies. Current research investigates memorization across various architectures, including transformers and recurrent neural networks, employing techniques like nucleus sampling and soft prompting to either measure or reduce this effect. Addressing memorization is crucial for ensuring data privacy, mitigating copyright infringement, and improving the reliability and generalizability of AI models across diverse applications.
Papers
October 31, 2024
October 7, 2024
October 3, 2024
September 20, 2024
August 29, 2024
August 28, 2024
June 23, 2024
June 20, 2024
June 14, 2024
June 6, 2024
May 20, 2024
March 20, 2024
March 17, 2024
March 1, 2024
February 4, 2024
December 6, 2023
November 28, 2023
October 10, 2023
October 4, 2023