Retrieval Augmented Generation
Retrieval-Augmented Generation (RAG) enhances large language models (LLMs) by incorporating external knowledge sources to improve accuracy and address limitations like hallucinations. Current research focuses on optimizing retrieval strategies (e.g., using hierarchical graphs, attention mechanisms, or determinantal point processes for diverse and relevant information selection), improving the integration of retrieved information with LLM generation (e.g., through various prompting techniques and model adaptation methods), and mitigating biases and ensuring fairness in RAG systems. The impact of RAG is significant, improving performance on various tasks like question answering and enabling more reliable and contextually aware applications across diverse domains, including healthcare and scientific research.
Papers
Eliciting In-context Retrieval and Reasoning for Long-context Large Language Models
Yifu Qiu, Varun Embar, Yizhe Zhang, Navdeep Jaitly, Shay B. Cohen, Benjamin Han
ASTRID -- An Automated and Scalable TRIaD for the Evaluation of RAG-based Clinical Question Answering Systems
Mohita Chowdhury, Yajie Vera He, Aisling Higham, Ernest Lim
ReARTeR: Retrieval-Augmented Reasoning with Trustworthy Process Rewarding
Zhongxiang Sun, Qipeng Wang, Weijie Yu, Xiaoxue Zang, Kai Zheng, Jun Xu, Xiao Zhang, Song Yang, Han Li
WebWalker: Benchmarking LLMs in Web Traversal
Jialong Wu, Wenbiao Yin, Yong Jiang, Zhenglin Wang, Zekun Xi, Runnan Fang, Deyu Zhou, Pengjun Xie, Fei Huang
Parallel Key-Value Cache Fusion for Position Invariant RAG
Philhoon Oh, Jinwoo Shin, James Thorne
Enhancing Retrieval-Augmented Generation: A Study of Best Practices
Siran Li, Linus Stenzel, Carsten Eickhoff, Seyed Ali Bahrainian
Research on the Online Update Method for Retrieval-Augmented Generation (RAG) Model with Incremental Learning
Yuxin Fan, Yuxiang Wang, Lipeng Liu, Xirui Tang, Na Sun, Zidong Yu
A Proposed Large Language Model-Based Smart Search for Archive System
Ha Dung Nguyen, Thi-Hoang Anh Nguyen, Thanh Binh Nguyen
LLMQuoter: Enhancing RAG Capabilities Through Efficient Quote Extraction From Large Contexts
Yuri Facanha Bezerra, Li Weigang
RAG-WM: An Efficient Black-Box Watermarking Approach for Retrieval-Augmented Generation of Large Language Models
Peizhuo Lv, Mengjie Sun, Hao Wang, Xiaofeng Wang, Shengzhi Zhang, Yuxuan Chen, Kai Chen, Limin Sun
A General Retrieval-Augmented Generation Framework for Multimodal Case-Based Reasoning Applications
Ofir Marom
Advancing Retrieval-Augmented Generation for Persian: Development of Language Models, Comprehensive Benchmarks, and Best Practices for Optimization
Sara Bourbour Hosseinbeigi, Sina Asghari, Mohammad Ali Seif Kashani, Mohammad Hossein Shalchian, Mohammad Amin Abbasi
Re-ranking the Context for Multimodal Retrieval Augmented Generation
Matin Mortaheb, Mohammad A. Amir Khojastepour, Srimat T. Chakradhar, Sennur Ulukus
Multi-task retriever fine-tuning for domain-specific and efficient RAG
Patrice Béchard, Orlando Marquez Ayala
End-to-End Bangla AI for Solving Math Olympiad Problem Benchmark: Leveraging Large Language Model Using Integrated Approach
H.M. Shadman Tabib, Jaber Ahmed Deedar
RAG-Check: Evaluating Multimodal Retrieval Augmented Generation Performance
Matin Mortaheb, Mohammad A. Amir Khojastepour, Srimat T. Chakradhar, Sennur Ulukus
Retrieval-Augmented Generation by Evidence Retroactivity in LLMs
Liang Xiao, Wen Dai, Shuai Chen, Bin Qin, Chongyang Shi, Haopeng Jing, Tianyu Guo