Chatbots Still Making Things Up? OpenAI Admits the Truth – KnowTechie
OpenAI says it comes down to how these models are trained.
by
OpenAI has a confession to make: even the smartest chatbots are still making stuff up.
In a new research paper, the company asks the question many of us have screamed at our screens: why do large language models like GPT-5 still hallucinate?
And the short answer is: because guessing is baked into their DNA.
Hallucinations, OpenAI explains, are those totally confident but totally wrong answers AI loves to serve up.
To prove the point, researchers ran a little experiment on one of the most popular chatbots (guess which one).
When asked for the title of co-author Adam Tauman Kalai’s PhD dissertation, the bot spat out three different titles, all fake. Same deal with his birthday: three dates, zero accuracy.
So how can a machine that crunches billions of data points still bomb on basic facts?
OpenAI says it comes down to how these models are trained. During pretraining, LLMs aren’t told what’s “true” or “false,” they’re just rewarded for predicting the next word. (Via: TechCrunch)
That works great for consistent patterns, like spelling or punctuation, but when it comes to obscure trivia, the model is basically winging it.
The paper doesn’t propose redoing pretraining from scratch. Instead, it points the finger at how models are evaluated.
Right now, evals are like multiple-choice tests graded only on correct answers. If you guess, you might get lucky. If you skip, you’re guaranteed a zero.
So, surprise surprise, the models learn to bluff instead of admitting, “I don’t know.”
OpenAI’s fix? Change the scoring system. Think of it like the SAT: wrong answers should hurt more than leaving it blank, and uncertainty should earn partial credit.
If models are rewarded for honesty, they’ll stop confidently fabricating nonsense just to climb the leaderboard.
The takeaway: hallucinations aren’t going away entirely, but maybe we can train AI to BS a little less often.
Until then, treat your chatbot like a charming friend who will absolutely lie to your face, but with style.
Should AI companies prioritize making chatbots admit uncertainty over giving confident but wrong answers? Do you think changing evaluation systems to penalize hallucinations will actually make AI more trustworthy, or will users just get frustrated with more “I don’t know” responses? Tell us below in the comments, or reach us via our Twitter or Facebook.
Ronil is a Computer Engineer by education and a consumer technology writer by choice. Over the course of his professional career, his work has appeared in reputable publications like MakeUseOf, TechJunkie, GreenBot, and many more. When not working, you’ll find him at the gym breaking a new PR.
zorse
September 10, 2025 at 12:06 pm
这文章讲得挺有意思的。OpenAI想让AI更诚实,不再瞎编,这想法不错。不过,改变评分系统就能行吗?我有点怀疑,用户会不会反而对满篇我不知道感到烦躁呢?这真是让人头疼的选择啊。
Your email address will not be published.
Shake it up! Mozilla’s new “Shake to Summarize” feature on iOS is here to…
The proposed settlement promised $3,000 per work, which the judge rejected.
Hinton say his girlfriend broke up with him using none other than OpenAI’s ChatGPT…
The irony: Sam Altman’s company develops the AI that is supposed to talk like…
Tal Melenboim, the visionary behind VFR.ai and StyleTech.ai, champions the cause of high-quality data…
Anthropic has to destroy all the pirated files it downloaded.
OAL Labs focuses on building new interfaces for how people collaborate with AI.
OpenAI wants AI play matchmaker between job seekers and employers.
Copyright © 2025 KnowTechie LLC / Powered by Kinsta