Google AI Overviews deliver millions of wrong answers every hour, analysis reveals

Reviewed byNidhi Govil

10 Sources

Share

A study by AI startup Oumi found that Google AI Overviews are accurate only 91% of the time, meaning tens of millions of inaccurate AI answers are served every hour across Google's 5 trillion annual searches. The analysis also revealed that over half of the AI-generated summaries cite sources that don't support their claims, raising concerns about user trust in AI and the unprecedented scale of misinformation.

News article

Google AI Overviews Face Scrutiny Over AI Accuracy

Google AI Overviews are delivering incorrect information at a scale that has alarmed researchers and raised questions about the reliability of AI-generated search overviews. According to a recent analysis conducted by AI startup Oumi and reported by The New York Times, Google's AI summaries are accurate approximately 91% of the time—a figure that sounds impressive until you consider the volume

1

2

. With Google processing roughly 5 trillion searches annually, this error rate translates to more than 57 million inaccurate AI answers every hour, or nearly 1 million wrong responses every minute

3

.

The Oumi report analyzed 4,326 searches using SimpleQA, a widely used benchmark developed by OpenAI for testing AI accuracy. In October 2025, Gemini 2 produced accurate responses around 85% of the time. By February 2026, when the same test was conducted using Gemini 3, AI accuracy improved to 91%

1

. While this demonstrates progress, it also reveals that Google was willing to deploy a model that hallucinated even more frequently, exposing hundreds of millions of users to inaccurate AI answers

5

.

Misleading Source Links Compound the Problem

Beyond the issue of AI hallucinations, the analysis uncovered a troubling trend in source attribution. Oumi's data suggests that Gemini 2 produced misleading source links 37% of the time, but this figure jumped to more than 56% with Gemini 3

1

. This means that even when Google's AI summaries provide correct information, the cited sources often don't support the claims being made. In some cases, an accurate overview links to a source with inaccurate information, while in others, the linked pages contain no relevant information at all

3

.

Researchers found that Facebook and Reddit were the second- and fourth-most-cited references for AI-generated search overviews. Accurate answers cited Facebook 5% of the time, while inaccurate responses cited the social media site 7% of the time

2

. This reliance on user-generated content platforms raises concerns about fact-checking standards and the unprecedented scale of misinformation that could spread through Google's dominant search platform.

User Trust in AI Creates Dangerous Blind Spots

The subtle nature of these errors makes them particularly dangerous. Unlike obvious AI hallucinations, most mistakes identified in Google AI Overviews weren't outrageous—they were subtle, leaving out important context, simplifying complex topics too aggressively, or presenting partially correct information as fully accurate

4

. Because AI-generated summaries appear at the top of search results and sound authoritative, most users won't question them.

Studies have shown that people tend to trust what an AI tells them without question, with one report finding that only 8% of users actually double-checked an AI's answer. Another experiment found that users still listened to AI when it gave them the wrong answer nearly 80% of the time—a grim trend researchers dubbed cognitive surrender

5

. This psychological effect means that the cleaner and more confident an answer feels, the more we trust it, creating a perfect storm for misinformation to spread unchecked.

Vulnerability to Manipulation and Gaming

The system's flaws extend beyond accidental errors. The New York Times report showed how a BBC journalist used a deliberately misleading article to poison the AI. Google's summary bot took the bait and, within 24 hours, repeated phony information from the source article

1

. This demonstrates that anyone who understands these bugs in the system can potentially game AI-generated search overviews into giving inaccurate statements by authoring misleading blog posts and artificially boosting traffic to their websites

2

.

Google Disputes Findings, But Concerns Remain

Google spokesperson Ned Adriance disputed the results, arguing that Oumi used the SimpleQA benchmark, which contains incorrect information in its own right, and that the test doesn't reflect actual search behavior. Google also noted that Oumi uses its own AI systems to analyze the AI summaries, which could introduce errors

1

. However, Google's own internal testing of Gemini 3 found that the AI model produced incorrect information 28% of the time when operating independently of Google Search

3

.

The implications for publisher traffic and the broader information ecosystem remain significant. Google's AI search summaries have been blamed for a downturn in publisher site traffic and job losses, which Google also disputes

1

. As AI-generated summaries become the default presentation for search results, users may never click through to verify information, meaning the AI answer becomes the final answer and context from original sources gets lost

4

. Experts recommend treating these tools with skepticism and using them only for quick summaries rather than as definitive sources, especially when accuracy matters.

Today's Top Stories

TheOutpost.ai

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Instagram logo
LinkedIn logo
Youtube logo
© 2026 TheOutpost.AI All rights reserved