×
Uncertainty Training: How AI experts are fighting back against the AI hallucination problem
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Virtual assistants and AI language models have a significant challenge with acknowledging uncertainty and admitting when they don’t have accurate information. This problem of AI “hallucination” – where models generate false information rather than admitting ignorance – has become a critical focus for researchers working to improve AI reliability.

The core challenge: AI models demonstrate a concerning tendency to fabricate answers when faced with questions outside their training data, rather than acknowledging their limitations.

  • When asked about personal details that aren’t readily available online, AI models consistently generate false but confident responses
  • In a test by WSJ writer Ben Fritz, multiple AI models provided entirely fictional answers about his marital status
  • Google’s Gemini similarly generated a completely fabricated response about a reporter being married to a deceased Syrian artist

Current research and solutions: Scientists at Germany’s Hasso Plattner Institut are developing methods to teach AI models about uncertainty during their training process.

  • Researchers Roi Cohen and Konstantin Dobler have created an intervention that helps AI systems learn to respond with “I don’t know” when appropriate
  • Their approach has shown promise in improving both the accuracy of responses and the ability to acknowledge uncertainty
  • However, the modified models sometimes display overcautiousness, declining to answer questions even when they have correct information

Industry implementation: Major AI companies are beginning to incorporate uncertainty training into their systems.

  • Anthropic has integrated uncertainty awareness into their Claude chatbot, which now explicitly declines to answer questions when it lacks confidence
  • This approach represents a shift from the traditional AI training paradigm that prioritized always providing an answer
  • Early results suggest that acknowledging uncertainty may actually increase user trust in AI systems

Expert perspectives: Leading researchers emphasize the importance of AI systems that can admit their limitations.

  • Professor José Hernández-Orallo explains that hallucination stems from AI training that prioritizes making guesses over acknowledging uncertainty
  • The ability to admit uncertainty may ultimately build more trust between humans and AI systems
  • Researchers argue that having reliable but limited AI systems is preferable to those that appear more capable but provide false information

Future implications: The challenge of managing AI hallucination represents a crucial inflection point in the development of trustworthy AI systems that can safely integrate into various aspects of daily life and professional applications.

Even the Most Advanced AI Has a Problem: If It Doesn’t Know the Answer, It Makes One Up

Recent News

Niantic plans $3.5B ‘Pokemon Go’ sale as HP acquires AI Pin

As gaming companies cut AR assets loose, Niantic is looking to sell its most valuable property while HP absorbs a struggling hardware startup.

This AI-powered wireless tree network detects and autonomously suppresses wildfires

A network of solar-powered sensors installed beneath forest canopies detects smoke and alerts authorities within minutes of a fire's start.

DeepSeek goes beyond ‘open weights’ with plans to release source code

Open-source AI firm will release internal code and model training infrastructure used in its commercial products.