By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: Improvements in ‘reasoning’ AI models may slow down soon, analysis finds | News
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > Improvements in ‘reasoning’ AI models may slow down soon, analysis finds | News
News

Improvements in ‘reasoning’ AI models may slow down soon, analysis finds | News

News Room
Last updated: 2025/05/12 at 9:34 PM
News Room Published 12 May 2025
Share
SHARE

An analysis by Epoch AI, a nonprofit AI research institute, suggests the AI industry may not be able to eke massive performance gains out of reasoning AI models for much longer. As soon as within a year, progress from reasoning models could slow down, according to the report’s findings.

Reasoning models such as OpenAI’s o3 have led to substantial gains on AI benchmarks in recent months, particularly benchmarks measuring math and programming skills. The models can apply more computing to problems, which can improve their performance, with the downside being that they take longer than conventional models to complete tasks.

Reasoning models are developed by first training a conventional model on a massive amount of data, then applying a technique called reinforcement learning, which effectively gives the model “feedback” on its solutions to difficult problems.

So far, frontier AI labs like OpenAI haven’t applied an enormous amount of computing power to the reinforcement learning stage of reasoning model training, according to Epoch.

That’s changing. OpenAI has said that it applied around 10x more computing to train o3 than its predecessor, o1, and Epoch speculates that most of this computing was devoted to reinforcement learning. And OpenAI researcher Dan Roberts recently revealed that the company’s future plans call for prioritizing reinforcement learning to use far more computing power, even more than for the initial model training.

But there’s still an upper bound to how much computing can be applied to reinforcement learning, per Epoch.

According to an Epoch AI analysis, reasoning model training scaling may slow downImage Credits:Epoch AI

Josh You, an analyst at Epoch and the author of the analysis, explains that performance gains from standard AI model training are currently quadrupling every year, while performance gains from reinforcement learning are growing tenfold every 3-5 months. The progress of reasoning training will “probably converge with the overall frontier by 2026,” he continues.

Epoch’s analysis makes a number of assumptions, and draws in part on public comments from AI company executives. But it also makes the case that scaling reasoning models may prove to be challenging for reasons besides computing, including high overhead costs for research.

“If there’s a persistent overhead cost required for research, reasoning models might not scale as far as expected,” writes You. “Rapid compute scaling is potentially a very important ingredient in reasoning model progress, so it’s worth tracking this closely.”

Any indication that reasoning models may reach some sort of limit in the near future is likely to worry the AI industry, which has invested enormous resources developing these types of models. Already, studies have shown that reasoning models, which can be incredibly expensive to run, have serious flaws, like a tendency to hallucinate more than certain conventional models.

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Sonic Labs Announces $10M Token Sale To Galaxy For U.S. Expansion | HackerNoon
Next Article Fallout’s second season premieres in December and will be followed by a third
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Google Introduces DolphinGemma to Support Dolphin Communication Research
News
Opera spins off MiniPay app to target Africa’s $54bn stablecoin market
Computing
Apple TV+ & Gordon Ramsey cook up ‘Knife Edge: Chasing Michelin Stars’
News
The DJI Mavic 4 Pro just flew to the top of my drone wish list | Stuff
Gadget

You Might also Like

News

Google Introduces DolphinGemma to Support Dolphin Communication Research

3 Min Read
News

Apple TV+ & Gordon Ramsey cook up ‘Knife Edge: Chasing Michelin Stars’

2 Min Read
News

Apple Expands Accessibility Features Including Live Captions, Magnifier and Sound Recognition

9 Min Read
News

CMA warns of ‘underlying weakness’ at Starling Bank – UKTN

3 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?