Newsfeed / Yann LeCun on AMI, World Models, and Why LLMs Aren't Enough
The Information Bottleneck·December 15, 2025

Yann LeCun on AMI, World Models, and Why LLMs Aren't Enough

Turing Award winner Yann LeCun announces AMI and explains why JEPA and world models - not LLMs - are the path to human-level AI.

Yann LeCun on AMI, World Models, and Why LLMs Aren't Enough

One of the godfathers of deep learning is betting his next decade on a different path than the rest of the industry.

Why Yann LeCun Thinks LLMs Are a Dead End

This conversation with Yann LeCun is a masterclass in contrarian thinking backed by decades of technical intuition. While the industry dumps billions into scaling LLMs, LeCun is launching AMI (Advanced Machine Intelligence) with a radically different thesis: you cannot get to human-level AI through text alone.

The math is stark. Training a competitive LLM requires 30 trillion tokens - roughly 10^14 bytes of text data. That's effectively all the freely available text on the internet. Compare that to video: those same 10^14 bytes represent just 15,000 hours of video at 2MB/s. That's 30 minutes of YouTube uploads. That's what a 4-year-old child has seen in their entire waking life.

LeCun's argument isn't just about data efficiency - it's about information density and redundancy. LLMs need massive parameter counts because they're essentially memorizing isolated facts from text. World models trained on video learn abstract representations of physics, causality, and dynamics. The redundancy in visual data isn't a bug - it's what enables learning.

What makes this conversation particularly valuable is the historical arc. LeCun walks through his 20-year journey from sparse autoencoders to Siamese networks to contrastive learning to JEPA. Each iteration was solving a specific problem: how do you train a system to learn useful abstract representations without collapsing to trivial solutions?

The answer he's converged on: Joint Embedding Predictive Architectures (JEPA). Instead of predicting every pixel (which is impossible for non-deterministic futures), you predict in an abstract representation space. You eliminate all the unpredictable details - noise, irrelevant textures, quantum uncertainty - and focus on what matters for planning.

The timing of AMI is deliberate. As Meta, Google, and other big labs "clam up" and become more secretive, LeCun is doubling down on open research. His argument is practical: you can't call it research if you don't publish, because you'll just fool yourself with internal hype. Scientists need external validation, and breakthroughs require the freedom to publish.

The product strategy is ambitious but pragmatic. AMI will publish upstream research while building actual products around world models and planning systems. The bet is that agentic systems based on LLMs "really don't work very well" because they lack the ability to predict consequences and plan in abstract representation spaces.

One technical detail buried in the conversation is particularly striking: current contrastive methods (like the ones LeCun pioneered in 2005-2006) max out around 200 dimensions in their learned representations, even on ImageNet. That's the ceiling. Recent advances like Barlow Twins, VICReg, and SigReg (part of the LJEPA system) are pushing past that limit by maximizing information content rather than just using contrastive loss.

The CFD analogy is perfect: we don't simulate airflow around a plane by modeling individual molecules, let alone quantum fields. We use abstract representations at the right level of granularity. That's what world models need to do - not simulate every detail, but learn the right abstractions for planning.

8 Insights From Yann LeCun on World Models and JEPA

  • AMI's thesis: Human-level AI requires world models trained on high-dimensional continuous data (video), not just text
  • Data efficiency gap: 10^14 bytes trains an LLM on all internet text OR a vision model on 15,000 hours of video (30 min of YouTube)
  • JEPA architecture: Predict in abstract representation space, not pixel space - eliminates unpredictable details while preserving structure
  • Research strategy: AMI will publish openly because "you cannot call it research unless you publish" - internal hype creates delusion
  • Technical evolution: From contrastive learning (2005) to VICReg/SigReg (2024) - moving beyond 200-dimension ceiling
  • Planning requirement: Intelligence needs consequence prediction + optimization, not just pattern matching
  • Industry critique: Big labs (Google, Meta, OpenAI) becoming more closed despite historical benefits of open research
  • Product vision: World models for planning systems that outperform LLM-based agents on reliability and sample efficiency

What This Means for the Future of AGI Research

A Turing Award winner is betting his next decade on the thesis that text-only AI cannot reach human-level intelligence. If he's right, the industry's trillion-dollar LLM investments are building tools, not minds - and the real path to AGI runs through video, world models, and learned physics.

Related