Critique Ability
Critique ability in large language models (LLMs) focuses on evaluating their capacity to identify and correct errors in their own reasoning and generated outputs. Current research emphasizes benchmarking this ability across diverse tasks, using metrics beyond simple accuracy to assess aspects like reasoning steps, constraint satisfaction, and handling of complex instructions, often employing techniques like chain-of-thought prompting and self-critique mechanisms. This research is crucial for improving LLM reliability and trustworthiness, impacting fields ranging from automated reasoning and code generation to more nuanced applications requiring robust and explainable AI.
Papers
May 4, 2024
April 24, 2024
April 17, 2024
April 3, 2024
March 29, 2024
March 26, 2024
March 1, 2024
February 29, 2024
February 28, 2024
February 21, 2024
February 19, 2024
February 16, 2024
January 26, 2024
January 24, 2024
January 7, 2024
December 13, 2023
December 12, 2023
December 11, 2023
December 4, 2023