Model Benchmark
Model benchmarking aims to objectively evaluate the performance of machine learning models, focusing on accuracy, bias, and efficiency across diverse tasks and datasets. Current research emphasizes developing more precise and nuanced evaluation metrics, particularly for addressing model biases and ensuring local validity, and explores the use of active learning and empirical Bayes methods to improve efficiency. These advancements are crucial for building trustworthy and reliable models, impacting various fields from manufacturing and official statistics to computer vision and natural language processing.
Papers
November 7, 2024
October 7, 2024
October 2, 2024
July 15, 2024
June 11, 2024
May 24, 2024
May 10, 2024
January 10, 2024
July 28, 2023
May 22, 2023
April 25, 2023
April 19, 2023
July 28, 2022
July 14, 2022
December 15, 2021