Summarization Task
Text summarization research aims to automatically generate concise and informative summaries of text, focusing on improving the alignment of machine-generated summaries with human preferences and addressing challenges in evaluation. Current research emphasizes the use of large language models (LLMs) and explores various architectures like BART and Mixture-of-Experts models, along with novel prompting techniques and fine-tuning strategies to enhance summarization quality across diverse domains and input types (e.g., medical reports, financial documents, code). This field is crucial for managing information overload and has significant implications for various applications, including information retrieval, content recommendation, and knowledge synthesis.
Papers
DSGPT: Domain-Specific Generative Pre-Training of Transformers for Text Generation in E-commerce Title and Review Summarization
Xueying Zhang, Yunjiang Jiang, Yue Shang, Zhaomeng Cheng, Chi Zhang, Xiaochuan Fan, Yun Xiao, Bo Long
LongT5: Efficient Text-To-Text Transformer for Long Sequences
Mandy Guo, Joshua Ainslie, David Uthus, Santiago Ontanon, Jianmo Ni, Yun-Hsuan Sung, Yinfei Yang