Long Text
Research on long text focuses on enabling large language models (LLMs) to effectively process and generate extended textual content, overcoming limitations of traditional transformer architectures. Current efforts concentrate on improving efficiency through optimized tokenization, novel attention mechanisms (like sparse attention and multi-kernel transformers), and techniques for semantic compression to handle longer sequences. This work is crucial for advancing numerous NLP applications, including improved machine translation, relation extraction from lengthy documents, and more accurate and efficient factual text generation.
58papers
Papers - Page 2
October 30, 2024
October 23, 2024
October 6, 2024
September 15, 2024
September 10, 2024
August 5, 2024
June 28, 2024
June 27, 2024
June 17, 2024