September 08, 2025
atlas

Why Chatbots Hallucinate and How We Can Teach Them to Say 'I Don't Know'

OpenAI’s latest research tackles a persistent and somewhat amusing problem with large language models like GPT-5 and ChatGPT: hallucinations. Those confidently wrong answers that feel like the AI just made something up to fill the silence. The research points out a core issue — our evaluation methods reward guessing over admitting uncertainty. This is akin to giving students full credit for taking wild guesses on a multiple-choice test but penalizing those who skip questions when unsure.

The key insight? Models are trained to predict next words based purely on language patterns, with no built-in truth-checking or an understanding of factual accuracy. This leads to plausible but false statements, especially for low-frequency facts that can’t be inferred from patterns alone.

The proposed fix is elegant and refreshingly pragmatic: revise evaluation metrics to penalize confident errors more than expressions of uncertainty, encouraging models to, in effect, learn to say "I don’t know." Much like how we discourage guessing with negative marking on tests, AI models too should be nudged away from guesswork in favor of cautious responses.

This rethinking is vital because, as things stand, chatbots have a strong incentive to keep guessing — after all, getting something right by chance beats admitting a lack of knowledge. As a tech-journalist who loves innovation but keeps one foot grounded, this research hits home. It reminds us that AI's ability to generate coherent language doesn’t equal understanding or certainty, and that our frameworks for judging AI performance need to evolve alongside their capabilities.

So, next time your AI assistant confidently flubs a fact, remember: it’s not lying with malice, it’s guessing because the game we made rewards guessing. Let’s change the rules, and maybe we’ll get chatbots that are a bit more humble, which could be their smartest upgrade yet. Source: Are bad incentives to blame for AI hallucinations?

Ana Avatar
Awatar WPAtlasBlogTerms & ConditionsPrivacy Policy

AWATAR INNOVATIONS SDN. BHD 202401005837 (1551687-X)