Summarization Task
Text summarization research aims to automatically generate concise and informative summaries of text, focusing on improving the alignment of machine-generated summaries with human preferences and addressing challenges in evaluation. Current research emphasizes the use of large language models (LLMs) and explores various architectures like BART and Mixture-of-Experts models, along with novel prompting techniques and fine-tuning strategies to enhance summarization quality across diverse domains and input types (e.g., medical reports, financial documents, code). This field is crucial for managing information overload and has significant implications for various applications, including information retrieval, content recommendation, and knowledge synthesis.
Papers
Open Domain Multi-document Summarization: A Comprehensive Study of Model Brittleness under Retrieval
John Giorgi, Luca Soldaini, Bo Wang, Gary Bader, Kyle Lo, Lucy Lu Wang, Arman Cohan
Socratic Pretraining: Question-Driven Pretraining for Controllable Summarization
Artidoro Pagnoni, Alexander R. Fabbri, Wojciech Kryściński, Chien-Sheng Wu