Retrieval Augmentation
Retrieval augmentation enhances large language models (LLMs) by incorporating external knowledge sources to improve accuracy, address hallucinations, and handle long contexts. Current research focuses on optimizing retrieval methods (e.g., k-NN, dense retrieval), integrating retrieved information effectively into LLMs (e.g., through modality fusion), and developing frameworks for managing and utilizing this external knowledge (e.g., dynamic retrieval based on model confidence). This approach is proving valuable across diverse applications, including question answering, text summarization, code generation, and even medical diagnosis, by improving factual accuracy and mitigating the limitations of LLMs trained solely on parametric knowledge.
Papers
Making Retrieval-Augmented Language Models Robust to Irrelevant Context
Ori Yoran, Tomer Wolfson, Ori Ram, Jonathan Berant
RA-DIT: Retrieval-Augmented Dual Instruction Tuning
Xi Victoria Lin, Xilun Chen, Mingda Chen, Weijia Shi, Maria Lomeli, Rich James, Pedro Rodriguez, Jacob Kahn, Gergely Szilvasy, Mike Lewis, Luke Zettlemoyer, Scott Yih
BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models
Qingqing Cao, Sewon Min, Yizhong Wang, Hannaneh Hajishirzi