Human Explanation
Human explanation in AI focuses on improving model interpretability and performance by integrating human-provided explanations into the training and inference processes. Current research heavily utilizes large language models (LLMs) like GPT-3 and others, exploring methods to leverage these models for generating, interpreting, and incorporating explanations into various tasks, including robot control, toxicity moderation, and mental health applications. This research aims to bridge the gap between complex model outputs and human understanding, ultimately leading to more trustworthy, robust, and effective AI systems across diverse domains.
Papers
September 25, 2024
July 30, 2024
June 17, 2024
May 8, 2024
April 18, 2024
April 3, 2024
March 11, 2024
February 16, 2024
May 23, 2023
May 8, 2023
May 4, 2023
March 11, 2023
January 23, 2023
November 14, 2022
September 13, 2022