
Ever asked an AI a question and got a super confident answer, only to find it was totally wrong? It's a bit like that loudmouth in the pub who knows everything, even when they don't. Well, scientists at MIT have just dropped a bombshell, teaching AI models to finally say 'I'm not sure' – making them far more reliable.
Today's most advanced AI reasoning models are often overconfident. They spout every answer with unshakable certainty, whether they're actually right or just guessing.
Researchers at MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL) have pinpointed this overconfidence.
They've traced it to a flaw in how these models are trained, and now they've fixed it without losing any accuracy.
Their new technique, called Reinforcement Learning with Calibration Rewards (RLCR), trains language models to give a confidence score alongside their answers. So, they don't just answer; they also tell you how certain they are.
In tests, RLCR reduced calibration error by a massive 90%.
Crucially, it maintained or even improved accuracy on both familiar and brand-new tasks.
It turns out the issue comes from a surprisingly simple source: current training methods for AI models reward correct answers but heavily penalise wrong ones. There’s nothing in between.
This means a model that guesses correctly gets the same praise as one that carefully reasoned its way to the right answer. Over time, this teaches AI to confidently answer everything, even when it’s effectively flipping a coin.
This can be dangerous. Imagine an AI used in medicine or finance that claims 'I'm 95 percent sure' but is only right half the time. Users have no clue they should seek a second opinion.
Mehul Damani, an MIT PhD student and co-lead author, said: "The standard training approach is simple and powerful, but it gives the model no incentive to express uncertainty or say I don’t know. So the model naturally learns to guess when it is unsure."
RLCR tackles this by adding a 'Brier score' to the training process. This score penalises the difference between what the model says it knows and its actual accuracy.
Confidently wrong answers are hit, but so are correct answers given with unnecessary uncertainty.
Isha Puri, another MIT PhD student and co-lead author, highlighted the issue: "What’s striking is that ordinary RL training doesn't just fail to help calibration. It actively hurts it. The models become more capable and more overconfident at the same time."
The team found that even the act of reasoning about uncertainty itself holds value, making the AI's self-reflection useful information, not just fluff.
MORE: AI Gets Brainier: Groundbreaking Tech Makes Models Leaner, Faster & Cheaper — https://trendwiremedia.com/2026/04/11/ai-gets-brainier-groundbreaking-tech-makes-models-leaner-faster-cheaper/
OFFICIAL SOURCE VERIFICATION: This report is based on official data from MIT. Document: Teaching AI models to say “I’m not sure” Source Link: [Read the official report from MIT](https://news.mit.edu/2026/teaching-ai-models-to-say-im-not-sure-0422)
Subscribe for $2 every four weeks for the first six months
Subscribe for $20 every four weeks for the first six months
Editorial Note: This report utilises automated data-sourcing and drafting technologies to ensure rapid coverage. Every article undergoes rigorous human fact-checking and editorial review by the Trend Wire Media Editorial Desk to ensure accuracy and adherence to our journalistic standards.