Multi FAct
Multi-FAct research centers on evaluating and improving the factuality of large language models (LLMs), particularly in the context of retrieval-augmented generation (RAG) and multilingual applications. Current efforts focus on developing robust evaluation datasets and frameworks, such as those incorporating knowledge graphs and multi-hop reasoning, to assess LLMs' ability to accurately retrieve and synthesize information. This work is crucial for building trustworthy AI systems, improving the reliability of information access, and advancing the development of more accurate and reliable LLMs across diverse languages and domains.
Papers
Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback
Baolin Peng, Michel Galley, Pengcheng He, Hao Cheng, Yujia Xie, Yu Hu, Qiuyuan Huang, Lars Liden, Zhou Yu, Weizhu Chen, Jianfeng Gao
Spanish Built Factual Freectianary (Spanish-BFF): the first AI-generated free dictionary
Miguel Ortega-Martín, Óscar García-Sierra, Alfonso Ardoiz, Juan Carlos Armenteros, Jorge Álvarez, Adrián Alonso