Calibration Performance
Calibration performance, the alignment of predicted probabilities with observed frequencies, is crucial for reliable machine learning models across diverse applications. Current research focuses on improving calibration in various model types, including deep neural networks, large language models, and Gaussian processes, employing techniques like temperature scaling, isotonic regression, and novel loss functions designed to directly optimize calibration metrics such as Expected Calibration Error (ECE). These advancements are vital for ensuring trustworthy predictions in high-stakes domains like medical diagnosis, autonomous driving, and climate forecasting, where accurate uncertainty quantification is paramount. Furthermore, research is exploring the relationship between calibration and other desirable model properties, such as robustness and generalization.
Papers
Using Platt's scaling for calibration after undersampling -- limitations and how to address them
Nathan Phelps, Daniel J. Lizotte, Douglas G. Woolford
Reinforcement Learning for Data-Driven Workflows in Radio Interferometry. I. Principal Demonstration in Calibration
Brian M. Kirk, Urvashi Rau, Ramyaa Ramyaa