Retrieval Augmented
Retrieval-Augmented Generation (RAG) enhances large language models (LLMs) and other machine learning models by incorporating external knowledge sources during inference, improving accuracy and addressing limitations like hallucinations and factual errors. Current research focuses on optimizing retrieval methods (e.g., using graph structures, determinantal point processes, or hierarchical representations), improving the integration of retrieved information with LLMs (e.g., through various reasoning modules and adaptive retrieval strategies), and applying RAG across diverse domains, including autonomous vehicles, robotics, and biomedical applications. This approach significantly impacts various fields by improving the reliability and efficiency of AI systems, particularly in knowledge-intensive tasks where access to and effective use of external information is crucial.
Papers
RAG-Star: Enhancing Deliberative Reasoning with Retrieval Augmented Verification and Refinement
Jinhao Jiang, Jiayi Chen, Junyi Li, Ruiyang Ren, Shijie Wang, Wayne Xin Zhao, Yang Song, Tao Zhang
PERC: Plan-As-Query Example Retrieval for Underrepresented Code Generation
Jaeseok Yoo, Hojae Han, Youngwon Lee, Jaejin Kim, Seung-won Hwang
BioRAGent: A Retrieval-Augmented Generation System for Showcasing Generative Query Expansion and Domain-Specific Search for Scientific Q&A
Samy Ateia, Udo Kruschwitz
Leveraging Retrieval-Augmented Tags for Large Vision-Language Understanding in Complex Scenes
Antonio Carlos Rivera, Anthony Moore, Steven Robinson