Adversarial Question
Adversarial question research focuses on crafting questions designed to expose vulnerabilities in question-answering systems, particularly large language models (LLMs) and retrieval-augmented generation (RAG) systems. Current research emphasizes developing metrics to evaluate the effectiveness of adversarial questions, creating robust benchmarks, and exploring various attack strategies, including prompt leaking and entity substitution, to assess model robustness. This work is crucial for improving the reliability and safety of LLMs in real-world applications by identifying and mitigating their weaknesses, ultimately leading to more trustworthy and robust AI systems.
Papers
June 24, 2024
June 3, 2024
May 10, 2024
March 15, 2024
March 12, 2024
February 25, 2024
February 16, 2024
January 20, 2024
December 26, 2023
October 5, 2023
August 8, 2022
June 29, 2022
May 7, 2022