Critique Ability
Critique ability in large language models (LLMs) focuses on evaluating their capacity to identify and correct errors in their own reasoning and generated outputs. Current research emphasizes benchmarking this ability across diverse tasks, using metrics beyond simple accuracy to assess aspects like reasoning steps, constraint satisfaction, and handling of complex instructions, often employing techniques like chain-of-thought prompting and self-critique mechanisms. This research is crucial for improving LLM reliability and trustworthiness, impacting fields ranging from automated reasoning and code generation to more nuanced applications requiring robust and explainable AI.
Papers
October 6, 2023
October 5, 2023
October 3, 2023
September 29, 2023
September 19, 2023
September 4, 2023
September 2, 2023
August 16, 2023
July 28, 2023
July 21, 2023
July 15, 2023
July 12, 2023
June 12, 2023
June 11, 2023
June 2, 2023
May 24, 2023
May 4, 2023
April 25, 2023
April 12, 2023