Abstractive Summarization Model
Abstractive summarization models aim to generate concise, coherent summaries that paraphrase the source text, unlike extractive methods which simply select sentences. Current research focuses on improving the factual consistency and reducing hallucinations in these models, often employing techniques like contrastive learning, reward shaping, and auxiliary information integration within architectures such as BART, PEGASUS, and large language models (LLMs). These advancements are crucial for enhancing the reliability and applicability of abstractive summarization across diverse domains, including legal text processing and disaster response, where accurate and trustworthy summaries are paramount.
Papers
Correcting Diverse Factual Errors in Abstractive Summarization via Post-Editing and Language Model Infilling
Vidhisha Balachandran, Hannaneh Hajishirzi, William W. Cohen, Yulia Tsvetkov
Salience Allocation as Guidance for Abstractive Summarization
Fei Wang, Kaiqiang Song, Hongming Zhang, Lifeng Jin, Sangwoo Cho, Wenlin Yao, Xiaoyang Wang, Muhao Chen, Dong Yu