Confidence Score
Confidence scores, representing a model's certainty in its predictions, are crucial for building trustworthy AI systems, particularly in high-stakes applications like healthcare and autonomous driving. Current research focuses on improving the calibration and reliability of these scores across diverse model architectures (including LLMs, transformers, and conformers) and tasks, often employing techniques like self-consistency, multicalibration, and novel scoring functions tailored to specific data characteristics (e.g., ordinal data, long-form text). The accurate estimation of confidence is vital for enhancing model performance, enabling selective classification (rejecting low-confidence predictions), and facilitating human-in-the-loop systems where trust and transparency are paramount.
Papers
Calibrating Large Language Models with Sample Consistency
Qing Lyu, Kumar Shridhar, Chaitanya Malaviya, Li Zhang, Yanai Elazar, Niket Tandon, Marianna Apidianaki, Mrinmaya Sachan, Chris Callison-Burch
A Comprehensive Study of Multilingual Confidence Estimation on Large Language Models
Boyang Xue, Hongru Wang, Rui Wang, Sheng Wang, Zezhong Wang, Yiming Du, Bin Liang, Kam-Fai Wong