Training Data Leakage
Training data leakage refers to the vulnerability of machine learning models, particularly large language models (LLMs) and deep neural networks, to revealing sensitive information from their training datasets. Current research focuses on understanding how various attack vectors, including gradient inversion and exploiting specific character patterns, can extract this data, even from seemingly secure training methods like differential privacy. This is a significant concern for privacy and intellectual property, impacting the development and deployment of machine learning systems across diverse applications, driving efforts to develop more robust training and defense mechanisms.
Papers
November 17, 2024
October 9, 2024
June 3, 2024
May 9, 2024
October 12, 2023
August 8, 2023
October 20, 2022