Welcome to DU! The truly grassroots left-of-center political community where regular people, not algorithms, drive the discussions and set the standards. Join the community: Create a free account Support DU (and get rid of ads!): Become a Star Member Latest Breaking News Editorials & Other Articles General Discussion The DU Lounge All Forums Issue Forums Culture Forums Alliance Forums Region Forums Support Forums Help & Search

General Discussion

Showing Original Post only (View all)

marble falls

(71,404 posts)
Wed Jul 9, 2025, 09:19 PM Jul 2025

A.I. Is Getting More Powerful, but Its Hallucinations Are Getting Worse [View all]

A.I. Is Getting More Powerful, but Its Hallucinations Are Getting Worse

A new wave of “reasoning” systems from companies like OpenAI is producing incorrect information more often. Even the companies don’t know why.

https://www.nytimes.com/2025/05/05/technology/ai-hallucinations-chatgpt-google.html

By Cade Metz and Karen Weise
Published May 5, 2025Updated May 6, 2025


-snip-

Today’s A.I. bots are based on complex mathematical systems that learn their skills by analyzing enormous amounts of digital data. They do not — and cannot — decide what is true and what is false. Sometimes, they just make stuff up, a phenomenon some A.I. researchers call hallucinations. On one test, the hallucination rates of newer A.I. systems were as high as 79 percent.

-snip-

The company found that o3 — its most powerful system — hallucinated 33 percent of the time when running its PersonQA benchmark test, which involves answering questions about public figures. That is more than twice the hallucination rate of OpenAI’s previous reasoning system, called o1. The new o4-mini hallucinated at an even higher rate: 48 percent.

-snip-

Hannaneh Hajishirzi, a professor at the University of Washington and a researcher with the Allen Institute for Artificial Intelligence, is part of a team that recently devised a way of tracing a system’s behavior back to the individual pieces of data it was trained on. But because systems learn from so much data — and because they can generate almost anything — this new tool can’t explain everything. “We still don’t know how these models work exactly,” she said.

-snip-

Another issue is that reasoning models are designed to spend time “thinking” through complex problems before settling on an answer. As they try to tackle a problem step by step, they run the risk of hallucinating at each step. The errors can compound as they spend more time thinking.

-snip-

Audio produced by Adrienne Hurst.

Cade Metz is a Times reporter who writes about artificial intelligence, driverless cars, robotics, virtual reality and other emerging areas of technology.

Karen Weise writes about technology for The Times and is based in Seattle. Her coverage focuses on Amazon and Microsoft, two of the most powerful companies in America.



“You spend a lot of time trying to figure out which responses are factual and which aren’t,” said Pratik Verma, co-founder and chief executive of Okahu, a company that helps businesses navigate the hallucination problem. “Not dealing with these errors properly basically eliminates the value of A.I. systems, which are supposed to automate tasks for you.”




22 replies = new reply since forum marked as read
Highlight: NoneDon't highlight anything 5 newestHighlight 5 most recent replies
OK, I'm anything but a high tech guru, boonecreek Jul 2025 #1
Frightening? Beyond frightening! There are server banks for single companies using the energy of small cities ... marble falls Jul 2025 #3
I had a hunch it was worse than I thought. boonecreek Jul 2025 #7
A few links: highplainsdem Jul 2025 #5
And thank you for the links. boonecreek Jul 2025 #8
Garbage in garbage out. enough Jul 2025 #2
An unfiltered mishmash of garbage and fact with no editing protocol for truth or fact. marble falls Jul 2025 #4
Those newer AI models are also worse for the environment. See this: highplainsdem Jul 2025 #6
one word: B.See Jul 2025 #9
What the Ukraine war seems to have developed into a lab for. marble falls Jul 2025 #21
Forbidden Planet 1956............... Lovie777 Jul 2025 #10
Exactamundo. I had that thought myself. marble falls Jul 2025 #20
Anybody here Faux pas Jul 2025 #11
This is why all these companies purple_haze Jul 2025 #12
Companies that try to rely heavily on AI The Madcap Jul 2025 #13
I can probably give you the reason for the hallucinations. PurgedVoter Jul 2025 #14
A good explanation RainCaster Jul 2025 #17
Do incorrect arithmetical calculations count as hallucinations? Disaffected Jul 2025 #18
Grammar error -- "but" should be "so". nt eppur_se_muova Jul 2025 #15
Thanks! marble falls Jul 2025 #16
AI bots are COMPLETELY incapable of actual logic William Seger Jul 2025 #19
It's logic absent volunteered intent. It doesn't meant to be right or wrong, it looks to bridging a gap with ... marble falls Jul 2025 #22
Latest Discussions»General Discussion»A.I. Is Getting More Powe...