Google Gemini ranked worst among major AI systems for news summarization accuracy, according to a comprehensive study by the European Broadcasting Union and BBC. The research found that while 42% of UK adults trust AI for news accuracy, Gemini showed significantly more problematic results than competitors like ChatGPT, Microsoft Copilot, and Perplexity.
What you should know: The study evaluated AI news summaries across multiple platforms, with concerning results across the board but Gemini standing out as particularly problematic.
• Most AI responses contained some form of inaccuracy or issue, despite user confidence in these systems.
• Among users under 25, 15% now rely on AI for their news consumption.
• While 84% of respondents said factual errors would significantly damage their trust, most users aren’t detecting these problems.
Gemini’s specific problems: The research identified several key areas where Google’s AI consistently underperformed compared to its rivals.
• Failed to provide clear links to source materials.
• Couldn’t distinguish between reliable sources and satirical content.
• Over-relied on Wikipedia as a primary source.
• Struggled to establish relevant context for news stories.
• Frequently butchered direct quotations from original sources.
The big picture: While all AI news summary tools showed improvement over the six-month study period, Gemini continues to lag behind despite making some gains in accuracy.
• The study tracked evolution across two main data collection periods, showing general improvement industry-wide.
• Gemini saw “some of the biggest gains when it came to accuracy” but still performed significantly worse than competitors.
• The gap between Gemini and other platforms remained substantial even after improvements.
Why this matters: With growing reliance on AI for news consumption, especially among younger users, the accuracy gap raises serious concerns about information quality and media literacy.
• The research challenges assumptions about AI reliability in news processing.
• Results suggest users may be overconfident in AI accuracy while missing significant errors.
• The findings highlight the need for more scrutiny of AI-generated news content.