LOGO

AI Reasoning Improvements May Slow Down, New Analysis Shows

May 12, 2025
AI Reasoning Improvements May Slow Down, New Analysis Shows

Potential Slowdown in Reasoning AI Performance Gains

A recent analysis conducted by Epoch AI, a non-profit AI research organization, indicates that substantial performance improvements in reasoning AI models may be nearing a plateau. The report suggests that progress from these models could decelerate within the coming year.

Recent Advances and Their Costs

Reasoning models, exemplified by OpenAI’s o3, have demonstrated significant advancements on AI benchmarks, particularly in areas like mathematics and programming. These models achieve enhanced performance by applying increased computational resources to problem-solving. However, this increased processing power results in longer task completion times compared to traditional models.

The development of reasoning models typically involves initially training a conventional model on a vast dataset. Subsequently, a technique known as reinforcement learning is employed, providing the model with feedback on its solutions to complex challenges.

Shifting Focus to Reinforcement Learning

According to Epoch’s findings, leading AI labs, including OpenAI, have not yet dedicated extensive computing power to the reinforcement learning phase of reasoning model training.

This is currently changing. OpenAI has disclosed that approximately ten times more computing resources were utilized to train o3 compared to its predecessor, o1. Epoch hypothesizes that the majority of this increased computational effort was directed towards reinforcement learning. Furthermore, OpenAI researcher Dan Roberts recently revealed the company’s future strategy prioritizes reinforcement learning, allocating even greater computational power to it than to the initial model training.

Limits to Computational Scaling

However, Epoch’s analysis posits that there is an inherent limit to the amount of computing power that can be effectively applied to reinforcement learning.

improvements in ‘reasoning’ ai models may slow down soon, analysis findsJosh You, an analyst at Epoch and the author of the report, explains that performance gains from standard AI model training are currently quadrupling annually. Conversely, performance gains from reinforcement learning are increasing tenfold every 3-5 months.

He continues, predicting that the progress of reasoning training will “probably converge with the overall frontier by 2026.”

Beyond Computational Challenges

Epoch’s analysis is based on several assumptions and incorporates publicly available statements from AI company executives. It also argues that scaling reasoning models may present challenges beyond computational limitations, including substantial research overhead costs.

You writes, “If a persistent overhead cost is required for research, reasoning models might not scale as far as expected.” He emphasizes that “Rapid compute scaling is potentially a very important ingredient in reasoning model progress, so it’s worth tracking this closely.”

Implications for the AI Industry

Any indication of potential limitations in reasoning models is likely to raise concerns within the AI industry, which has made significant investments in developing these model types. Prior research has already demonstrated that reasoning models, despite their capabilities, can be costly to operate and exhibit flaws, such as a higher propensity for generating inaccurate information (hallucinations) compared to conventional models.

#AI#artificial intelligence#reasoning#AI models#machine learning#analysis