Prompt Extraction
Prompt extraction research focuses on how attackers can recover the secret instructions (prompts) used to customize large language models (LLMs), thereby compromising intellectual property and potentially enabling malicious attacks. Current research investigates the vulnerability of various LLMs, including models from OpenAI and others, to different attack methods, and develops and evaluates defensive strategies against prompt leakage. This area is crucial because the widespread use of prompt-based LLM services necessitates robust security measures to protect proprietary information and maintain the integrity of these systems.
Papers
August 5, 2024
June 10, 2024