Memorization Effect
Memorization, the tendency of large language models (LLMs) and other deep learning models to reproduce training data verbatim, is a significant research area focusing on quantifying its extent, understanding its mechanisms (including the roles of attention and cross-attention), and developing mitigation strategies. Current research investigates memorization across various architectures, including transformers and recurrent neural networks, employing techniques like nucleus sampling and soft prompting to either measure or reduce this effect. Addressing memorization is crucial for ensuring data privacy, mitigating copyright infringement, and improving the reliability and generalizability of AI models across diverse applications.
Papers
November 28, 2023
October 10, 2023
October 4, 2023
April 21, 2023
November 22, 2022
October 24, 2022
October 17, 2022
May 22, 2022
May 20, 2022
March 23, 2022
December 16, 2021