
End of Scaling Era
The Shift
After five years of "scale is all you need" (2020-2025), leading AI researchers are signaling a fundamental change: pure scaling is hitting diminishing returns. The next breakthroughs will require genuine research innovation, not just more compute.
Key Signals
Ilya Sutskever's Framework
The former OpenAI Chief Scientist frames AI history as oscillating eras:
- 2012-2020: Research era (deep learning breakthroughs)
- 2020-2025: Scaling era (bigger is better)
- 2025+: Return to research (new paradigms needed)
"Is the belief really that if you just 100x the scale everything would be transformed? I don't think that's true."
Demis Hassabis's Formula
Google DeepMind's CEO describes their current approach:
"We operate on 50% scaling, 50% innovation. Both are required for AGI."
He notes: "There's a lot of room between exponential and asymptotic" - improvements continue, but not at the pace of the scaling era.
Why Now?
Pre-training Data Is Finite
The internet contains only so much high-quality text. Models have been trained on most of it. More data requires synthetic generation or new modalities.
Compute Costs Are Enormous
Training runs cost $100M+. The economics of "just scale more" become prohibitive without guaranteed returns.
Benchmark Saturation
Models are hitting ceilings on existing benchmarks while still failing at tasks that should be easy (the "jagged intelligence" problem).
What Changes
For AI Labs
- Research hiring increases - Need novel algorithms, not just engineering
- Architectural experiments - Transformers may not be the endgame
- Multimodal and embodied AI - Text alone may not get us to AGI
For the Industry
- Smaller players can compete - Innovation matters more than scale
- Differentiation beyond size - Models compete on specialization, not just parameters
- Focus on applications - With models plateauing, product matters more
The Open Question
Is this a temporary plateau (like 2016-2019 before GPT-3) or a fundamental limit of current approaches? The answer will determine whether AGI comes from scaling, innovation, or both.
Related Reading
- Scaling Laws - The paradigm being questioned
- Application Over Training - The strategic response
- Jagged Intelligence - Why scaling alone isn't enough