Long Form Generation
Long-form generation focuses on creating extended and coherent text outputs from large language models (LLMs), aiming to improve both the factual accuracy and overall quality of generated content. Current research emphasizes mitigating issues like hallucinations (factual inaccuracies) and improving control over length and structure, often employing techniques like retrieval-augmented generation, hierarchical clustering, and reinforcement learning to guide the generation process. These advancements are crucial for enhancing the reliability and usability of LLMs in diverse applications, ranging from scientific writing and report generation to personalized content creation and human-computer interaction.
Papers
VERISCORE: Evaluating the factuality of verifiable claims in long-form text generation
Yixiao Song, Yekyung Kim, Mohit Iyyer
LongLaMP: A Benchmark for Personalized Long-form Text Generation
Ishita Kumar, Snigdha Viswanathan, Sushrita Yerra, Alireza Salemi, Ryan A. Rossi, Franck Dernoncourt, Hanieh Deilamsalehy, Xiang Chen, Ruiyi Zhang, Shubham Agarwal, Nedim Lipka, Chien Van Nguyen, Thien Huu Nguyen, Hamed Zamani