Extraction Attack
Extraction attacks exploit the tendency of large language models (LLMs) and other deep learning models to memorize training data, enabling adversaries to illicitly retrieve sensitive information. Current research focuses on developing and evaluating these attacks against various model architectures, including LLMs like GPT and specialized models for tasks such as image generation, analyzing the effectiveness of different attack strategies and exploring mitigation techniques like model editing and data deduplication. Understanding and mitigating extraction attacks is crucial for ensuring the privacy and security of AI systems and their applications, particularly in sensitive domains like healthcare and finance.
Papers
November 18, 2024
October 20, 2024
October 14, 2024
September 12, 2024
August 5, 2024
July 2, 2024
June 20, 2024
June 13, 2024
June 10, 2024
May 30, 2024
January 24, 2024
December 4, 2023
October 26, 2023
September 29, 2023
September 21, 2023
September 19, 2023
June 20, 2023
May 25, 2023
May 15, 2023