By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: Openai believes having discovered why the IAS hallucinates: they don’t know how to say “I don’t know”
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > Mobile > Openai believes having discovered why the IAS hallucinates: they don’t know how to say “I don’t know”
Mobile

Openai believes having discovered why the IAS hallucinates: they don’t know how to say “I don’t know”

News Room
Last updated: 2025/09/08 at 1:36 PM
News Room Published 8 September 2025
Share
SHARE

Alfa. Give answers that are not true. Invents, and also does it with a simply amazing ease. The answers seem coherent thanks to that apparent coherence and security, but the truth is that that can end up causing disturbing problems. For example, that you recommend putting glue on the pizza so that the cheese is well stuck.

Hallucinations are not a mystical error. In Openai they know the problem well, and they have just published a report in which they analyze the causes of hallucinations. According to the study, these arise from “statistical pressures” in the training and evaluation stages.

Good news, you don't have to choose model using GPT-5. Bad news, it is GPT-5 who chooses it without notifying you

In WorldOfSoftware

Good news, you don’t have to choose model using GPT-5. Bad news, it is GPT-5 who chooses it without notifying you

Award for guessing. The problem, they explain, is that in these procedures it is being rewarded that the “guess” instead of admitting that there may be uncertainty in the answers, “like when a student faces a difficult question in an exam” and answers some of the options available to see if he is lucky and hits. In Openai they point out how AI does something similar in those cases, and in training it is encouraged to answer the answer instead of answering with a simple “I don’t know.”

Damn probabilities. In the pre-training phase the models learn the distribution of language from a large text corpus. And that is where the authors emphasize that although the data of origin are completely free of errors, the statistical techniques that are used cause the model to make mistakes. The generation of a valid text is much more complex than answering a simple question with a yes or a not as “is this output valid?”

Predicting the word has trap. Language models learn to “speak” with preventive, in which they learn to predict the next word of a phrase thanks to the intake of huge amounts of text. There are no “true/false” labels in each sentence with which it is trained, only “positive (valid) examples of language. That makes it harder to avoid hallucinations, but in Openai they think they have a possible response that in fact they have already applied in GPT-5.

A new training. To mitigate the problem in Openai they propose to introduce a binary classification that they call “IS-ID-VALID” (IIV, “Is it valid?”), Which trains a model to distinguish between valid and erroneous responses.

Screen capture 2025 09 08 at 13 35 49

GPT-5 is somewhat more humble. When there is a correct answer, Openai models classify the answers that the model gives in three groups: correct, errors, and abstentions, which reflect some “humility.” According to its data, GPT-5 has improved in terms of hallucinations rate because in its tests much more abstain (52%) than O4-mini (1%), for example. Although O4-mini is slightly better in correct answers, it is much worse in error rate.

Benchmarks reward the successes. The study also indicates how benchmarks and technical cards of the current models (Model Cards) are fully focused on the successes. In this way, although the models of AI effectively improve and succeed more and more, they continue to hallucinate and there are hardly any data on these hallucinations rates that should be replaced by a simple “I do not know.”

{“videoid”: “x8jpy2b”, “Autoplay”: fals, “title”: “What is behind it like chatgpt, dall-e or midjourney? | artificial intelligence”, “tag”: “Webedia-prod”, “Duration”: “1173”}

Easy solution. But as in the testing tests, there is a way to prevent students from playing the pool: penalize errors rather than uncertainty. In these exams, answering well can use a point but answering badly can subtract 0.5 points and not answer zero points. If you don’t know the answer, guessing can be very expensive. Well, with AI models, the same.

In WorldOfSoftware | We have an AI problem: there is no reliable way to know if Chatgpt is better than Gemini, Copilot or Claude

(function() {
window._JS_MODULES = window._JS_MODULES || {};
var headElement = document.getElementsByTagName(‘head’)(0);
if (_JS_MODULES.instagram) {
var instagramScript = document.createElement(‘script’);
instagramScript.src=”https://platform.instagram.com/en_US/embeds.js”;
instagramScript.async = true;
instagramScript.defer = true;
headElement.appendChild(instagramScript);
}
})();

–
The news

Openai believes having discovered why the IAS hallucinates: they don’t know how to say “I don’t know”

It was originally posted in

WorldOfSoftware

By Javier Pastor.

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Huge flaw on many Bluetooth helmets, can we spy on your conversations easily?
Next Article Leaked iPhone 17 battery sizes reveal a surprising eSIM difference
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

AMD to roll out AI chip customized for China this July · TechNode
Computing
PSA: Plex suffers security breach, change your password right now!
News
The HackerNoon Newsletter: Can ChatGPT Outperform the Market? Week 4 (9/8/2025) | HackerNoon
Computing
Today's NYT Mini Crossword Answers for Sept. 9 – CNET
News

You Might also Like

Mobile

China is no longer made up of moving away from Nvidia. His next step is the heart of the AI ​​with a system that breaks molds

5 Min Read
Mobile

We ranked all the films of the saga, from the worst to the best

10 Min Read
Mobile

Ghost of Yotei PS5 And DualSense Controllers Could Soon Arrive In India: Here’s What We Know

3 Min Read
Mobile

Huge flaw on many Bluetooth helmets, can we spy on your conversations easily?

3 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?