Rates Of Hallucination In AI Models From Google, OpenAI On The Rise

Zinger Key Points
  • AI models, like Google’s AI Overviews and OpenAI’s o3 and o4-mini, are experiencing higher levels of hallucination than their predecessors
  • Features like AI Overviews are also leading to a decrease in clicks through to articles with accurate information
  • Experts warn that these compounding issues could lead to a decrease in efficiency and an increase in the spread of fake news
  • 9 Out of the Last 10 Summers this "Power Pattern" Delivered Winners - Get The Details Now.

Google's AI Overviews are "hallucinating" false information and drawing clicks away from accurate sources, experts warned The Times of London late last week.

Google introduced its AI Overviews, a feature that aims to provide quick answers to search queries, in May 2024. Summaries are written by Google's Gemini AI – a large language model similar to ChatGPT – scans through the results of the search to create the graphs and includes links to some of the sources.

Don't Miss:

Google Vice President of Search Elizabeth Reid said In a blog post that the overviews were designed to be a "jumping off point" that provided higher-quality clicks to webpages. "People are more likely to stay on [those pages], because we've done a better job of finding the right info and helpful webpages for them."

However, experts told The Times of London that these answers can be "confidently wrong" and direct searchers away from legitimate information.

When generative AI imagines facts or otherwise makes mistakes, computer scientists refer to it as hallucinating. These hallucinations can include references to non-existent scientific papers, like those NOTUS found were cited in Health and Human Services Secretary Robert F. Kennedy Jr.'s "Make America Healthy Again" report, and a host of other errors in judgment.

Trending: Invest where it hurts — and help millions heal: Invest in Cytonics and help disrupt a $390B Big Pharma stronghold.

Shortly after AI Overviews were launched last year, users began to point out the frequency with which these summaries included inaccurate information, the Times of London reports. One of its most notorious hallucinations was the suggestion that a user add non-toxic glue to pizza sauce to help the cheese stick better. 

  • Massive Demand & Disruptive Potential – Boxabl has received interest for over 190,000 homes, positioning itself as a major disruptor in the housing market.
  • Revolutionary Manufacturing Approach – Inspired by Henry Ford’s assembly line, Boxabl’s foldable tiny homes are designed for high-efficiency production, making homeownership more accessible.
  • Affordable Investment Opportunity – With homes priced at $60,000, Boxabl is raising $1 billion to scale production, offering investors a chance to own a stake in its growth.
Share Price: $0.80
Min. Investment: $1,000
Valuation: $3.5B

Google pushed back, claiming that many of the examples circulating were fake, but Reid acknowledged in her blog post that "some odd, inaccurate or unhelpful AI Overviews certainly did show up. And while these were generally for queries that people don't commonly do, it highlighted some specific areas that we needed to improve."

According to the experts who spoke to The Times of London, despite the technological advancements and improvements, hallucinations are getting worse rather than better. New reasoning systems are producing more incorrect responses than their predecessors, and designers aren't sure why.

See Also: Maximize saving for your retirement and cut down on taxes: Schedule your free call with a financial advisor to start your financial journey – no cost, no obligation

In a recent study, it was found that two recent OpenAI models, o3 and 04-mini, hallucinated in 33% and 48% of answers, respectively, according to The Times of London. These percentages are more than double those of previous models.

Features like Google's AI Overviews or ChatGPT summaries are also drawing clicks away from more accurate resources. Laurence O'Toole, founder of tech firm Authoritas, tracked the impact of AI Overviews and told The Times of London that when they exist click click-throughs to articles go down by 40% to 60%. 

The compounding issues of AI presenting inaccurate information and causing searchers to click through to more accurate articles have many worried about efficiency and the spread of fake news. 

"You spend a lot of time trying to figure out which responses are factual and which aren't," the chief executive of Okahu, Pratik Verma, told The New York Times last month. Okahu works with AI engineers to make the technology better, and helps companies troubleshoot issues, including hallucinations. "Not dealing with these errors properly basically eliminates the value of A.I. systems, which are supposed to automate tasks for you," he said. 

Read Next:

Image: Shutterstock

Market News and Data brought to you by Benzinga APIs

Comments
Loading...