Back to news
AI Research
Apr 22, 2026

MIT researchers develop method to improve AI confidence estimates

Apr 22, 2026
AI Summary

Researchers at MIT's CSAIL have created a technique called RLCR that enhances AI models' ability to express uncertainty in their answers. This method significantly reduces overconfidence in AI systems while maintaining accuracy, which is crucial for applications in sensitive fields like medicine and finance.

MIT researchers develop method to improve AI confidence estimates
  • Current AI models often exhibit overconfidence, providing answers with unwavering certainty regardless of their actual accuracy.
  • MIT researchers have identified a flaw in traditional reinforcement learning training methods that leads to this overconfidence.
  • The new technique, RLCR (Reinforcement Learning with Calibration Rewards), introduces a reward structure that encourages models to assess their uncertainty and produce calibrated confidence scores alongside their answers.
  • In testing, RLCR reduced calibration error by up to 90% while maintaining or improving accuracy across various benchmarks, including new datasets.
  • The method penalizes models for confidently incorrect answers and encourages them to express uncertainty when appropriate.
  • Results indicate that models trained with RLCR not only perform better in terms of calibration but also enhance the utility of confidence estimates during inference.
  • The research highlights the importance of models being able to express uncertainty, especially in high-stakes decision-making environments.
  • The findings will be presented at the International Conference on Learning Representations later this month.
ai confidencehallucinationtraining methodsreasoning modelsreliability