High Confidence
High confidence in machine learning models, particularly large language models (LLMs), is a crucial area of research focusing on aligning a model's expressed certainty with the actual accuracy of its predictions. Current efforts involve developing methods to estimate confidence, often using techniques like multi-modal similarity analysis, activation-based calibration, and conformal prediction, and applying these methods to various model architectures including diffusion models and Bayesian neural networks. Achieving high confidence is vital for improving the reliability and trustworthiness of AI systems across diverse applications, from code generation and medical diagnosis to autonomous driving and financial forecasting, ultimately fostering greater user trust and safer deployment.
Papers
Beyond Confidence: Reliable Models Should Also Consider Atypicality
Mert Yuksekgonul, Linjun Zhang, James Zou, Carlos Guestrin
Analysis of Perceived Stress Test using Machine Learning
Toygar Tanyel
Improving Confidence in Evolutionary Mine Scheduling via Uncertainty Discounting
Michael Stimson, William Reid, Aneta Neumann, Simon Ratcliffe, Frank Neumann