Information Leakage
Information leakage in machine learning models, particularly large language models (LLMs) and diffusion models, refers to the unintended exposure of sensitive training data through model outputs or gradients. Current research focuses on quantifying leakage risks across various architectures, including retrieval-augmented generation, mixture-of-experts models, and diffusion models, and developing mitigation strategies like knowledge sanitization and differential privacy. Understanding and addressing information leakage is crucial for ensuring the responsible development and deployment of AI systems, protecting user privacy, and maintaining public trust.
Papers
October 19, 2024
July 9, 2024
June 19, 2024
June 13, 2024
May 15, 2024
April 24, 2024
March 24, 2024
January 25, 2024
September 21, 2023
September 20, 2023
June 12, 2023
May 30, 2023
March 31, 2023