High Confidence
High confidence in machine learning models, particularly large language models (LLMs), is a crucial area of research focusing on aligning a model's expressed certainty with the actual accuracy of its predictions. Current efforts involve developing methods to estimate confidence, often using techniques like multi-modal similarity analysis, activation-based calibration, and conformal prediction, and applying these methods to various model architectures including diffusion models and Bayesian neural networks. Achieving high confidence is vital for improving the reliability and trustworthiness of AI systems across diverse applications, from code generation and medical diagnosis to autonomous driving and financial forecasting, ultimately fostering greater user trust and safer deployment.
Papers
Confidence v.s. Critique: A Decomposition of Self-Correction Capability for LLMs
Zhe Yang, Yichang Zhang, Yudong Wang, Ziyao Xu, Junyang Lin, Zhifang Sui
A Self-Efficacy Theory-based Study on the Teachers Readiness to Teach Artificial Intelligence in Public Schools in Sri Lanka
Chathura Rajapakse, Wathsala Ariyarathna, Shanmugalingam Selvakan