Generated Text
Generated text research focuses on understanding and mitigating the challenges posed by increasingly sophisticated large language models (LLMs) producing human-quality text. Current efforts concentrate on detecting machine-generated text, often employing techniques like latent-space analysis and fine-tuned transformer models (e.g., RoBERTa, DeBERTa) to identify subtle differences in writing style and structure between human and AI-generated content. This field is crucial for addressing concerns about misinformation, plagiarism, and authenticity, impacting various domains from education and journalism to legal and scientific publishing.
Papers
SMLT-MUGC: Small, Medium, and Large Texts -- Machine versus User-Generated Content Detection and Comparison
Anjali Rawal, Hui Wang, Youjia Zheng, Yu-Hsuan Lin, Shanu Sushmita
Detection and Measurement of Syntactic Templates in Generated Text
Chantal Shaib, Yanai Elazar, Junyi Jessy Li, Byron C. Wallace