world-models

3 items

WIRED · 2026-04-28 2026-05-01-w2

The Man Behind AlphaGo Thinks AI Is Taking the Wrong Path

David Silver raised $1.1B at a $5.1B valuation on the argument that LLMs are bounded by the human-data manifold, and that the only way out is RL-trained agents operating in simulation. The architectural evidence is real: AlphaGo's Move 37 came from outside the space of human play, and Sutton's Turing Award validates the theoretical foundation Silver is building on. What this week's picks clarify is that the capability argument is almost beside the point: the OpenAI goblin postmortem shows that even current systems can't reliably control what they're optimizing for, and Karpathy's MenuGen demo shows that the harness around the model is already more consequential than the model itself. Silver's unpriced bottleneck, reliable verifiers for unbounded domains, is also the missing piece in both of those stories. The next value pool isn't in bigger models or better prompts; it's in the infrastructure that tells you whether the output was actually right.

WIRED 2026-04-28-1

The Man Behind AlphaGo Thinks AI Is Taking the Wrong Path

David Silver left DeepMind to raise $1.1B at $5.1B for Ineffable Intelligence on a thesis that says LLMs hit a ceiling defined by the human-data manifold and only RL-trained agents in simulations can break through. The architectural argument has teeth: AlphaGo's Move 37 came from outside human play, and Sutton just won the Turing Award for the foundational work. The unspoken bottleneck if Silver is right isn't compute or data, it's verifiers — reliable scoring functions for unbounded domains like science, governance, novel discovery — and that is the quiet investable category nobody's pricing yet.

Not Boring 2026-03-23-1

World Models: Computing the Uncomputable

The definitional move matters more than the technology survey: action-conditioned prediction, P(st+1 | st, at), is presented as the line separating world models from video slop. If that definition holds, the $4B+ deployed into World Labs, AMI, GI, and Decart is a bet that spatial-temporal reasoning trained on games and driving footage transfers to general embodied control. The strongest signal is Ai2's MolmoBot result: a sim-only-trained policy outperforming VLAs trained on thousands of hours of real data. If sim-to-real transfer keeps improving, the entire robotics data flywheel thesis inverts: synthetic environments become the bottleneck worth owning, not real-world demonstrations.