Model Trustworthiness
Model trustworthiness, encompassing reliability, accuracy, and fairness, is a critical area of research aiming to ensure that AI systems make dependable predictions and decisions. Current efforts focus on improving model calibration (aligning confidence with accuracy), mitigating biases and vulnerabilities, and enhancing explainability through techniques like knowledge transfer between large language models and concept bottleneck models. These advancements are crucial for building trust in AI across diverse applications, from healthcare and finance to environmental modeling and security, ultimately promoting responsible and beneficial AI deployment.
Papers
November 13, 2024
October 30, 2024
July 5, 2024
May 27, 2024
May 25, 2024
April 29, 2024
January 4, 2024
November 16, 2023
October 25, 2023
August 6, 2023
April 21, 2023
April 7, 2023
November 7, 2022
October 31, 2022
August 18, 2022