D\'ej\`a Vu Memorization
Déjà vu memorization refers to the unintended storage and retrieval of specific training data by large machine learning models, impacting generalization and raising privacy concerns. Current research focuses on quantifying this memorization in various architectures, including vision-language models and large language models, and developing methods to mitigate it, such as employing reinforcement learning to discourage verbatim or approximate reproduction of training examples. Understanding and controlling déjà vu memorization is crucial for improving model robustness, ensuring data privacy, and advancing the trustworthiness of AI systems.
Papers
November 15, 2024
March 19, 2024
February 3, 2024
May 2, 2023
April 26, 2023
February 3, 2023
May 30, 2022
May 22, 2022