Model Understanding
Model understanding focuses on interpreting the internal workings and decision-making processes of complex machine learning models, particularly large language and vision-language models, to improve their reliability and trustworthiness. Current research emphasizes developing rigorous benchmarks and evaluation frameworks to assess model comprehension, including tests of reasoning abilities, handling of ambiguous queries, and sensitivity to data perturbations. This work is crucial for identifying and mitigating biases, improving model robustness, and fostering greater transparency and explainability in AI systems, ultimately leading to safer and more effective applications across various domains.
Papers
October 9, 2024
September 9, 2024
June 15, 2024
May 6, 2024
January 25, 2024
January 24, 2024
January 2, 2024
December 13, 2023
November 6, 2023
October 30, 2023
October 19, 2023
November 3, 2022
October 24, 2022
August 26, 2022
June 6, 2022
April 30, 2022