Task Specific Memorization Measure

Task-specific memorization measures aim to quantify how much training data a machine learning model, particularly large language models (LLMs) and generative models like GANs and diffusion models, directly memorizes rather than generalizes from. Current research focuses on developing and applying these measures across various architectures, including Vision Transformers (ViTs) and autoencoders, to assess memorization in different tasks like code completion, quotation attribution, and speech recognition. Understanding and mitigating memorization is crucial for addressing privacy concerns, improving model reliability, and ensuring the integrity of generated content, particularly in high-stakes applications.

Papers