LLM Detector

LLM detectors aim to distinguish text generated by large language models (LLMs) from human-written text, addressing concerns about misinformation and academic integrity. Current research focuses on improving the robustness of these detectors against adversarial attacks, exploring various approaches including zero-shot methods that leverage pre-trained language models and those employing instruction tuning or contrasting model outputs. The development of reliable and accurate LLM detectors is crucial for mitigating the risks associated with AI-generated content in diverse applications, from combating online scams to ensuring authenticity in academic settings. However, challenges remain in achieving consistent performance across different LLMs and domains, and in developing detectors that are resistant to sophisticated evasion techniques.

Papers