Home
Skip to main content
xStryk™

Decision Intelligence for AI in production — guardrails, traceability & evaluation.

xTheus

AMI Labs and JEPA: LeCun's Bet on World Models Over LLMs

On December 19, 2025, Yann LeCun confirmed what the industry had been speculating for months: he launched a new company dedicated to world model development. The startup, identified as AMI Labs, reportedly seeks a valuation above $5 billion. LeCun will not serve as CEO. What he will bring is the technical thesis he has defended publicly since 2022: LLMs cannot reason, and the future of artificial intelligence runs through architectures that learn models of the physical world.

$5B+
Valuation sought
3+
Years of JEPA research at Meta FAIR
2-6×
Sample efficiency vs. diffusion (V-JEPA 2)
I·V·W
JEPA generations (Image → Video → World)
Foundational Thesis — 2022

"LLMs produce statistically plausible text. They do not understand the world. They cannot plan. They cannot learn the way animals learn — by observing the physical world. A machine that only predicts the next token will never reach human-level intelligence." — Yann LeCun, A Path Towards Autonomous Machine Intelligence (2022)

The Three Structural Limitations of LLMs

LeCun's position is not new. Since his 2022 paper he has argued that LLMs present three limitations that no scaling can resolve. These are not data limitations, parameter limitations, or compute limitations — they are paradigm limitations.

Limitation 01

No world model

An LLM predicts the next token in a sequence. It maintains no internal state of the world. It does not "know" that objects fall, doors open, or fire burns — it has merely seen those phrases co-occur many times.

Limitation 02

No continuous planning

LLMs operate in discrete token spaces. Physical and decision environments are continuous: (s, a) ∈ S × A. Chain of thought simulates planning but cannot navigate continuous state spaces or optimize trajectories with dynamic constraints.

Limitation 03

Correlation ≠ causality

Token prediction pretraining maximizes statistical coherence, not causal understanding. A model can correctly answer "if I drop the ball it falls" without having learned any causal mechanism — only text co-occurrence across trillions of documents.

The bet

JEPA → W-JEPA

Joint Embedding Predictive Architecture does not generate in observation space. It learns abstract representations and predicts future states in that space — enabling multi-step planning, causal reasoning, and physical understanding without producing pixels or tokens.

JEPA Architecture: The Mathematics of the Paradigm

The central intuition of JEPA is to minimize distances in representation space, not to reconstruct signals. The encoder learns to compress x into s_x — discarding irrelevant perceptual details. The predictor learns to predict s_y from s_x. The key: y is never reconstructed in pixel or token space.

JEPA: Forward Pass (training)
Context xx
Encoder E()E(\cdot)
sx=E(x)s_x = E(x)
Predictor P()P(\cdot)
s^y=P(sx)\hat{s}_y = P(s_x)
Loss
s^ysy2\|\hat{s}_y - s_y\|^2
Target
Target y (masked region)s_y = E(y)Stop-gradient (EMA)
Key
y is never reconstructedNo generative decoderSemantics, not pixels
JEPA Training Objective
minθ,ϕ  (x,y)D ⁣(Pθ ⁣(Eϕ(x)),  sg ⁣[Eϕ(y)])\min_{\theta,\phi}\; \sum_{(x,y)} D\!\left( P_\theta\!\left(E_\phi(x)\right),\; \mathrm{sg}\!\left[E_\phi(y)\right] \right)
θ\theta = predictor parameters · ϕ\phi = encoder parameters · sg[]\mathrm{sg}[\cdot] = stop-gradient · DD = distance in embedding space
Difference from LLMs (autoregressive)
LLM:maxθ  tlogPθ ⁣(xtx<t)\text{LLM:}\quad \max_{\theta}\;\sum_t \log P_\theta\!\left(x_t \mid x_{<t}\right)
LLM maximizes token sequence probability — discrete space, no world model, no planning

Inference: LLM vs. W-JEPA in Planning

How Each Paradigm Plans
LLM CoT
Prompt
Token 1
Token 2…n
Answer
W-JEPA
State s₀
Multi-step predictor
Latent trajectory
Optimal action

The operational difference is critical. An LLM generates tokens one by one: each token conditions the next but cannot "go back" to modify a prior decision. W-JEPA operates in a continuous state space where the predictor can optimize a complete trajectory before executing the first action — the equivalent of mentally planning the route before starting to walk. V-JEPA 2 already demonstrated emergent physical reasoning (pendulum prediction, falls, collisions) without explicit physics supervision — purely as a consequence of predicting in latent space.

The JEPA Family: From Image to Physical World

W-JEPA — continuous states, integrated physics, multi-step planning. AMI Labs target for robotics and autonomous decision systems.W-JEPA · IN DEVELOPMENT
V-JEPA 2 — temporal prediction in video latent space. Emergent physical reasoning (pendulums, falls, collisions). 2-6× efficiency vs. diffusion.V-JEPA 2 · 2024
V-JEPA 1 — spatiotemporal masking prediction. First demonstration of video semantic representations without reconstruction. Meta FAIR, 2024.V-JEPA 1 · 2024
I-JEPA — image semantic representations without supervised augmentations. Outperforms MAE and BEiT on linear benchmarks. Meta FAIR, 2023.I-JEPA · 2023

JEPA vs LLMs vs Diffusion Models: Comparative Analysis

DimensionJEPA / W-JEPALLM (GPT / Claude)Diffusion (Sora / DALL·E)
Training objectivePredict representations (embedding space)Predict next token (max log-likelihood)Reconstruct full signal (denoising)
Operation spaceContinuous (latent)Discrete (vocabulary)Continuous (pixels)
Causal reasoningStructural — in the architectureEmergent — fragile, spurious causalityAbsent
Multi-step planningNative in latent space (trajectories)Chain of thought — no continuous stateN/A
Sample efficiencyHigh — 2-6× vs. diffusion (V-JEPA 2)Low — scale (trillions of tokens) compensatesMedium
Integrated physicsCore W-JEPA objectiveNo — linguistic correlation of physicsPartial (appearance, not mechanics)
Natural applicationsRobotics, simulation, autonomous controlNLP, code, symbolic reasoningContent generation

AMI Labs: Spinout Structure and Context

LeCun's role

Not the CEO

Classic scientific spinout pattern: the technical founder leads the vision, an external operator manages the company. LeCun provides the thesis; another executive turns it into business.

Research origin

Meta FAIR

I-JEPA and V-JEPA were developed at Meta FAIR. W-JEPA applications (robotics, hardware-in-the-loop) do not fit Meta's advertising business model — hence the spinout.

Target valuation

$5B+

A round of this size implies tier-1 fund backing. The bet: that W-JEPA can be trained with efficiency comparable to current LLMs in domains where LLMs have structural limits.

Public confirmation

19 dic. 2025

Described by press as "the worst-kept secret in tech." Industry speculation anticipated the announcement by months — confirming that capital interest preceded the public confirmation.

industrial roboticsphysical simulationautonomous planningdexterous controlcausal reasoningworld modelinghardware-in-the-loopedge inference

W-JEPA Production Stack: Reference Architecture

W-JEPA · Production Stack (reference)
World data
Real video (RGB-D cameras)Physics simulations (Isaac Lab)Sensor data (IMU, LIDAR)
JEPA Pretraining
Encoder E(·) — spatiotemporal ViTPredictor P(·) — transformerDGX cluster — massive GPU
W-JEPA Planner
Trajectory optimization (latent)Multi-step causal reasoningConsequence simulation
Governance
Decision traceabilityCertified physical guardrailsPost-action audit
Edge / Actuator
Inference &lt;50ms (Jetson / FPGA)Motor action / decision

Timeline: From JEPA to AMI Labs

2022
"A Path Towards Autonomous Machine Intelligence"
LeCun publishes the foundational paper describing JEPA and the three structural limitations of LLMs. The document circulates widely in the research community.
2023
I-JEPA (Meta FAIR)
First public implementation. Image JEPA outperforms MAE and BEiT on linear benchmarks at lower computational cost. No supervised augmentations — emergent semantic representations.
2024
V-JEPA 1 y V-JEPA 2
V-JEPA 2 demonstrates 2-6× efficiency and emergent physical reasoning capabilities (pendulums, falls, collisions) without physics supervision. Paradigm validation in temporal domains.
dic. 2025
AMI Labs — public confirmation
LeCun confirms the startup on December 19. Valuation sought: $5B+. Structure: LeCun as chief scientist, external CEO. Focus: W-JEPA for physical environments and autonomous decision.
2026+
W-JEPA
Declared objective: physical world model with native planning in continuous state space. Applications: industrial robotics, physical simulation, autonomous decision systems with integrated physics.

Key Takeaways

  • JEPA does not generate: it predicts distances in abstract representation space. Eliminates reconstruction of irrelevant perceptual details and achieves 2-6× greater sample efficiency than equivalent diffusion models (V-JEPA 2).
  • LLMs' three structural limitations are paradigm limitations, not scale limitations: no world model, no continuous-space planning, statistical correlation instead of causality. No parameter or token scaling resolves them.
  • V-JEPA 2 demonstrated emergent physical reasoning (pendulums, falls, collisions) without explicit physics supervision — validating that predicting in latent space is sufficient to learn mechanics.
  • AMI Labs seeks $5B+ with LeCun as chief scientist (not CEO). The Meta FAIR spinout is natural: W-JEPA for robotics and hardware-in-the-loop is incompatible with Meta's advertising business model.
  • For production decision systems, W-JEPA enables what LLMs cannot: multi-step latent trajectories, consequence simulation, and counterfactual reasoning in continuous state spaces.