By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: Are bad incentives to blame for AI hallucinations? | News
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > Are bad incentives to blame for AI hallucinations? | News
News

Are bad incentives to blame for AI hallucinations? | News

News Room
Last updated: 2025/09/07 at 5:23 PM
News Room Published 7 September 2025
Share
Are bad incentives to blame for AI hallucinations? |  News
SHARE

A new research paper from OpenAI asks why large language models like GPT-5 and chatbots like ChatGPT still hallucinate, and whether anything can be done to reduce those hallucinations.

In a blog post summarizing the paper, OpenAI defines hallucinations as “plausible but false statements generated by language models,” and it acknowledges that despite improvements, hallucinations “remain a fundamental challenge for all large language models” — one that will never be completely eliminated.

To illustrate the point, researchers say that when they asked “a widely used chatbot” about the title of Adam Tauman Kalai’s Ph.D. dissertation, they got three different answers, all of them wrong. (Kalai is one of the paper’s authors.) They then asked about his birthday and received three different dates. Once again, all of them were wrong.

How can a chatbot be so wrong — and sound so confident in its wrongness? The researchers suggest that hallucinations arise, in part, because of a pretraining process that focuses on getting models to correctly predict the next word, without true or false labels attached to the training statements: “The model sees only positive examples of fluent language and must approximate the overall distribution.”

“Spelling and parentheses follow consistent patterns, so errors there disappear with scale,” they write. “But arbitrary low-frequency facts, like a pet’s birthday, cannot be predicted from patterns alone and hence lead to hallucinations.”

The paper’s proposed solution, however, focuses less on the initial pretraining process and more on how large language models are evaluated. It argues that the current evaluation models don’t cause hallucinations themselves, but they “set the wrong incentives.”

The researchers compare these evaluations to the kind of multiple choice tests random guessing makes sense, because “you might get lucky and be right,” while leaving the answer blank “guarantees a zero.” 

Techcrunch event

San Francisco
|
October 27-29, 2025

“In the same way, when models are graded only on accuracy, the percentage of questions they get exactly right, they are encouraged to guess rather than say ‘I don’t know,’” they say.

The proposed solution, then, is similar to tests (like the SAT) that include “negative [scoring] for wrong answers or partial credit for leaving questions blank to discourage blind guessing.” Similarly, OpenAI says model evaluations need to “penalize confident errors more than you penalize uncertainty, and give partial credit for appropriate expressions of uncertainty.”

And the researchers argue that it’s not enough to introduce “a few new uncertainty-aware tests on the side.” Instead, “the widely used, accuracy-based evals need to be updated so that their scoring discourages guessing.”

“If the main scoreboards keep rewarding lucky guesses, models will keep learning to guess,” the researchers say.

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Transatlantic Twofer in Google’s Display-Ads Case: EU Levies €2.95B Fine, DOJ Proposes Breakup Transatlantic Twofer in Google’s Display-Ads Case: EU Levies €2.95B Fine, DOJ Proposes Breakup
Next Article Today's NYT Connections Hints, Answers for Sept. 8, #820 Today's NYT Connections Hints, Answers for Sept. 8, #820
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Kubernetes Community Retires Popular Ingress NGINX Controller
Kubernetes Community Retires Popular Ingress NGINX Controller
News
Best Residential Mortgage Rates Ottawa Your Guide for Home Financing
Best Residential Mortgage Rates Ottawa Your Guide for Home Financing
Gadget
Prediction: This Stock Bubble Will Burst in 2026 and Three Popular Stocks Will Crash (Hint: No Artificial Intelligence)
Prediction: This Stock Bubble Will Burst in 2026 and Three Popular Stocks Will Crash (Hint: No Artificial Intelligence)
News
Realme 16 Pro+ Storage Options, Colours, And Early Design Details Leak Online
Realme 16 Pro+ Storage Options, Colours, And Early Design Details Leak Online
Mobile

You Might also Like

Kubernetes Community Retires Popular Ingress NGINX Controller
News

Kubernetes Community Retires Popular Ingress NGINX Controller

6 Min Read
Prediction: This Stock Bubble Will Burst in 2026 and Three Popular Stocks Will Crash (Hint: No Artificial Intelligence)
News

Prediction: This Stock Bubble Will Burst in 2026 and Three Popular Stocks Will Crash (Hint: No Artificial Intelligence)

6 Min Read
Apple gave Vision Pro two big, very needed upgrades recently – 9to5Mac
News

Apple gave Vision Pro two big, very needed upgrades recently – 9to5Mac

4 Min Read
The Best Black Friday Apple Deals I’ve Seen: 0 Off The Latest MacBook Pro, Apple Watch SE for 9, and More
News

The Best Black Friday Apple Deals I’ve Seen: $250 Off The Latest MacBook Pro, Apple Watch SE for $129, and More

21 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?