Trustworthy Prediction

Trustworthy prediction focuses on developing machine learning models that not only produce accurate predictions but also provide reliable uncertainty estimates and are robust against biases and out-of-distribution data. Current research emphasizes methods for improving calibration, such as novel calibration measures and k-Nearest Neighbor uncertainty estimation, alongside techniques for enhancing fairness and robustness, including causal debiasing and stratified invariance. These advancements are crucial for deploying machine learning models in high-stakes applications like healthcare, finance, and safety-critical systems, where confidence in predictions is paramount. The ultimate goal is to build models that are not only accurate but also transparent and explainable, fostering trust and responsible use of AI.

Papers