Long Text
Research on long text focuses on enabling large language models (LLMs) to effectively process and generate extended textual content, overcoming limitations of traditional transformer architectures. Current efforts concentrate on improving efficiency through optimized tokenization, novel attention mechanisms (like sparse attention and multi-kernel transformers), and techniques for semantic compression to handle longer sequences. This work is crucial for advancing numerous NLP applications, including improved machine translation, relation extraction from lengthy documents, and more accurate and efficient factual text generation.
Papers
Length-Induced Embedding Collapse in Transformer-based Models
Yuqi Zhou, Sunhao Dai, Zhanshuo Cao, Xiao Zhang, Jun Xu
Language Models can Self-Lengthen to Generate Long Texts
Shanghaoran Quan, Tianyi Tang, Bowen Yu, An Yang, Dayiheng Liu, Bofei Gao, Jianhong Tu, Yichang Zhang, Jingren Zhou, Junyang Lin
Suri: Multi-constraint Instruction Following for Long-form Text Generation
Chau Minh Pham, Simeng Sun, Mohit Iyyer
VERISCORE: Evaluating the factuality of verifiable claims in long-form text generation
Yixiao Song, Yekyung Kim, Mohit Iyyer
LongLaMP: A Benchmark for Personalized Long-form Text Generation
Ishita Kumar, Snigdha Viswanathan, Sushrita Yerra, Alireza Salemi, Ryan A. Rossi, Franck Dernoncourt, Hanieh Deilamsalehy, Xiang Chen, Ruiyi Zhang, Shubham Agarwal, Nedim Lipka, Chien Van Nguyen, Thien Huu Nguyen, Hamed Zamani