Large Language Models

Google AI Overviews Lies: Millions Per Hour

You're googling a simple fact, and bam—top result confidently feeds you nonsense. New tests expose Google's AI Overviews lying to millions hourly, eroding trust in the search giant we all rely on.

Chart of Google's AI Overviews error rate from NYT-Oumi tests showing millions of daily lies

Key Takeaways

  • Google's AI Overviews errs on 9% of simple facts, equating to millions of lies hourly across billions of searches.
  • Accuracy improved from 85% to 91% with Gemini 3, but still a massive risk for users relying on quick answers.
  • This erodes Google's core trust advantage, potentially sparking lawsuits and user exodus.

Imagine you’re rushing to the ER with a kid’s fever spiking, typing ‘safe dose Tylenol 5 year old’ into Google. Boom, AI Overviews pops up at the top, spouting some half-baked dosage pulled from who-knows-where. Wrong? You bet—could be disastrous.

That’s the nightmare Google’s AI Overviews is flirting with, every single day. Not some abstract benchmark score, but real stakes for regular folks who trust search to deliver truth, not hallucinations dressed as answers.

How Many Lies Hit Your Screen Hourly?

The New York Times teamed up with startup Oumi to run the numbers, using SimpleQA—a no-frills benchmark from OpenAI with 4,000+ questions that demand straight facts. Google’s AI Overviews, juiced by the latest Gemini 3 model, nailed 91% correctly. Sounds decent, right? Wrong.

“If you extrapolate this miss rate out to all Google searches, AI Overviews is generating tens of millions of incorrect answers per day.”

Tens of millions daily. That’s your tax on using Google: roughly millions of lies per hour beamed to eyeballs worldwide. Oumi first tested last year on Gemini 2.5—85% accuracy. Bumped to 91% now. Progress? Sure. But from lousy to merely alarming.

Look, I’ve covered this Valley circus for two decades. Back in the ’90s, AltaVista and Lycos spewed garbage too, but they didn’t pretend to be omniscient overlords. Google won by being brutally accurate. Now? They’re betting the farm on probabilistic word salad, all to keep you from clicking ads lower down.

And here’s my unique take, one you won’t find in the Times piece: this mirrors the Theranos blood-test debacle. Remember Elizabeth Holmes hawking 99% accuracy demos while the real machines flopped? Google demos shiny AI Overviews wins, but live? 9% error rate on simple facts. Who’s auditing the auditors when billions ride on it?

Why Do These Errors Even Happen?

Take Bob Marley’s old house turning museum. AI Overviews cites three sources—two irrelevant, Wikipedia’s got conflicting years—and picks the dud: wrong date, served straight-faced.

Or Yo-Yo Ma’s hall of fame nod. It quotes the org’s own site, then flatly denies the hall exists. “There’s no such thing,” it claims. Hallucination 101.

These aren’t edge cases. SimpleQA is baby’s first fact-check: dates, names, basics. If Gemini chokes here, imagine nuanced stuff like tax advice or medical symptoms. Google’s PR spins ‘getting better’—yeah, from F to D-minus.

But. Short sentence for emphasis.

The money angle—who’s cashing in? Not you. Google. AI Overviews keeps eyes on the page longer, fewer outbound clicks to publishers starving for traffic. Publishers sue (hello, News Media Alliance cases), regulators sniff (EU’s AI Act looming), yet Alphabet stock? Mooning. Investors love the ‘efficiency,’ screw the fallout.

Is Google’s AI Overviews Actually Improving Enough?

91% from 85%. That’s a 6-point jump post-Gemini 3. Oumi’s tools scraped real Overviews responses, no cherry-picking. But extrapolate to billions of queries? Hundreds of thousands wrong per minute. Per. Minute.

Real people pay. Small biz owner googles ‘OSHA forklift rules’—botches compliance, fined thousands. Student crams history dates, fails exam. Grandma’s recipe swap turns toxic.

Google’s fix? More compute, fancier RAG (retrieval-augmented generation, for the acronym haters). But hallucinations lurk in LLMs’ DNA—probabilities, not certainties. My bold prediction: by 2026, class-action suits over AI-misinfo damages hit Google harder than adboycott fears. Valley’s arrogance will crack.

Skeptical vet mode: I’ve seen hype cycles crash. Dot-com 1.0 promised search utopia; delivered spam. AI 1.0? Same script, bigger servers.

Paragraph of one: Trust is fracturing.

Then sprawl: Google’s not alone—Perplexity, Bing Copilot stumble too—but they own 90% search. Monopoly amplifies the sin. Users bolt to DuckDuckGo or organic results (hold Shift+click, folks). Or ChatGPT direct. Google’s moat? Crumbling under its own bot weight.

Who Wins When AI Search Lies?

Advertisers? Kinda—they pay for impressions, not accuracy. Publishers? Hemorrhaging. Users? Suckers.

Google execs tout ‘helpful’ AI, but read the room: backlash since launch, memes galore. ‘Don’t eat rocks’ glue saga still haunts. Yet they double down. Why? Data flywheel—lies train better models? Ironic.

Historical parallel: 2000s SEO black-hat era. Google crushed it with Panda, Penguin updates. Now they’re the spammer. Poetic justice.

Wrap the rant: Ditch the buzz. Demand verifiable sources atop every Overview. Or kill the feature. Real people deserve better than probabilistic BS.


🧬 Related Insights

Frequently Asked Questions

Will Google’s AI Overviews replace traditional search results?

Not yet—it’s supplemental, but eats clicks. Shift-click bypasses it; expect more users wising up.

How accurate is Google’s AI Overviews really?

91% on SimpleQA benchmarks, but that’s simple facts. Real-world queries? Likely worse, with millions of errors daily.

What are examples of Google AI Overviews lies?

Wrong dates for Bob Marley’s house museum; denying Yo-Yo Ma’s hall of fame exists despite citing the source.

Aisha Patel
Written by

Former ML engineer turned writer. Covers computer vision and robotics with a practitioner perspective.

Frequently asked questions

Will Google's AI Overviews replace traditional search results?
Not yet—it's supplemental, but eats clicks. Shift-click bypasses it; expect more users wising up.
How accurate is Google's AI Overviews really?
91% on SimpleQA benchmarks, but that's simple facts. Real-world queries? Likely worse, with millions of errors daily.
What are examples of <a href="/tag/google-ai-overviews/">Google AI Overviews</a> lies?
Wrong dates for Bob Marley's house museum; denying Yo-Yo Ma's hall of fame exists despite citing the source.

Worth sharing?

Get the best AI stories of the week in your inbox — no noise, no spam.

Originally reported by Ars Technica - AI

Stay in the loop

The week's most important stories from theAIcatchup, delivered once a week.