Great Truth
Research on "truth" in the context of large language models (LLMs) focuses on developing methods to assess and improve the factual accuracy and reliability of LLM outputs. Current efforts involve analyzing the underlying causes of LLM inaccuracies (e.g., multi-step reasoning failures, biases in training data), designing game-theoretic approaches to enhance consistency and reliability during decoding, and developing robust lie detection methods using techniques like spectral analysis of model activations. This research is crucial for mitigating the risks of misinformation spread by LLMs and building more trustworthy AI systems across various applications, from healthcare to social media.
Papers
October 30, 2023
October 10, 2023
October 2, 2023
August 30, 2023
July 18, 2023
May 22, 2023
May 6, 2023
February 13, 2023
January 28, 2023
January 23, 2023
December 29, 2022
November 22, 2022
September 26, 2022
September 22, 2022
September 16, 2022
July 14, 2022
May 4, 2022
February 24, 2022