Google DeepMind

8 items

WIRED 2026-05-01-1

I've Covered Robots for Years. This One Is Different

None of the few dozen robot arms on the market today can screw in a light bulb; Eka can. The meaningful claim isn't the demo, though. It's that Eka and Ineffable Intelligence are now two independent labs publicly betting on pure-simulation-with-physics against the VLA consensus, and the bottleneck they're attacking lives in custom grippers that know how a key feels. Form factor follows task. The trillions flowing through the human hand don't care what's holding the chicken nugget.

Google DeepMind Blog 2026-04-15-1

Gemini Robotics-ER 1.6: Powering real-world robotics tasks through enhanced embodied reasoning

Google just revealed where robotics value accrues: the reasoning model, not the robot. ER 1.6 acts as a tool-calling orchestrator that sits above Boston Dynamics' Spot, reading industrial gauges via a multi-step agentic vision pipeline (zoom → point → code → interpret). The architecture is the text-agent pattern transplanted to physical AI: foundation model reasons and plans, specialized VLAs execute motor control. If this stack bifurcation holds, hardware makers become distribution channels for the intelligence layer — and most robotics investment theses are overweighting the wrong tier.

Quanta Magazine 2026-04-14-2

The AI Revolution in Math Has Arrived

AlphaEvolve found hypercube structures in permutation groups that mathematicians hadn't noticed in 50 years: not by answering the question posed, but by surfacing a pattern nobody thought to look for. The real capability shift isn't AI proving things faster; it's AI scanning combinatorial spaces too large for human intuition and returning structures that reframe entire research programs. Discovery is being commoditized; the scarce resource is now verification infrastructure and the human judgment to recognize which discoveries matter.

The Economist 2026-04-11-1

AI mathematicians: By devising and verifying proofs, AI is changing how maths is done

Four independent groups racing to formalize proofs in Lean, and Math Inc. translated Viazovska's sphere-packing work in weeks rather than the decade Hales needed for peer review, but DARPA's Shafto names the real bottleneck as trust, not computation. AI's primary value in mathematics is making claims auditable at scale. That separation between generation and formal verification is the architecture every enterprise AI system will eventually need.

The Twenty Minute VC (20VC) 2026-04-08-1

Demis Hassabis on 20VC: AGI Timeline, LLM Non-Commoditization, and the Algorithmic Innovation Thesis

Hassabis argues frontier models won't commoditize because algorithmic innovation, not scaling spend, is the new differentiator: only 3-4 labs can still invent. What he conspicuously omits is inference economics; collapsing costs commoditize models at the useful-capability threshold regardless of what happens at the absolute frontier. The real signal is his "jagged intelligence" admission: if foundation models remain inconsistent, the durable moat lives in application-layer reliability engineering, not model access.

The Atlantic 2026-04-05-2

The AI Industry Wants to Automate Itself

Anthropic says 90% of its code is AI-written; Amodei says that speeds up workflows 15-20%. The gap between those numbers is the story: code generation was never the bottleneck. The real race among frontier labs isn't who automates coding fastest; it's who closes the "research taste" gap between rote execution and the judgment to know what's worth building. Even the incremental version of this race compresses model generations faster than institutions can adapt.

Not Boring 2026-03-23-1

World Models: Computing the Uncomputable

The definitional move matters more than the technology survey: action-conditioned prediction, P(st+1 | st, at), is presented as the line separating world models from video slop. If that definition holds, the $4B+ deployed into World Labs, AMI, GI, and Decart is a bet that spatial-temporal reasoning trained on games and driving footage transfers to general embodied control. The strongest signal is Ai2's MolmoBot result: a sim-only-trained policy outperforming VLAs trained on thousands of hours of real data. If sim-to-real transfer keeps improving, the entire robotics data flywheel thesis inverts: synthetic environments become the bottleneck worth owning, not real-world demonstrations.

MIT Technology Review 2026-03-21-2

OpenAI's Autonomous AI Researcher: The Org Chart Is the Trade

OpenAI's "AI researcher" North Star is less about technology and more about organizational design: Pachocki's claim that 2-3 people plus a data center replaces a 500-person R&D org is a labor market thesis, not an AI capability prediction. The September 2026 "AI intern" timeline is vague enough to declare victory with any narrow demo, and the 2028 full researcher target collides with an unsolved reliability cliff that gets one paragraph in an exclusive that should have interrogated it. The real gap: coding has test suites, math has proofs, but the article scopes confidently from those verifiable domains to "business and policy dilemmas" where no ground truth exists. Everyone debates the technology; the trade is in the inference economics nobody is modeling and the evaluation frameworks nobody is building.