3 items

Three layers of the AI coding stack revealed the same structural gap this week: Cursor built its own model to get equivalent quality at 10x lower token cost; labs are subsidizing inference 5:1 to win market share; engineers are competing on internal leaderboards that track token volume, not output value. The metric that would justify the entire capital stack — useful output per token spent — has no tooling, no incentives, and no shared definition yet. Whoever builds that measurement layer has leverage over every company currently treating consumption as a proxy for progress.

Bloomberg 2026-03-22-1

Cursor Ships Composer 2: Vertical Model Independence as Margin Strategy

Cursor's Composer 2 isn't a model launch: it's a margin play. The company built a coding-only model that matches Opus 4.6 on Terminal-Bench at 10x lower token cost, because reselling Anthropic's API while competing with Claude Code was structurally terminal. The real signal is self-summarization, an RL technique that compresses 100K-token agent trajectories to 1K tokens with 50% fewer errors than prompted compaction; if this holds, it changes the economics of every long-horizon agentic workflow, not just coding.

Wall Street Journal 2026-03-22-2

The Trillion Dollar Race to Automate Our Entire Lives

WSJ's narrative arc — coding tools → life automation → trillion-dollar market — buries the only number that matters: Anthropic disclosed Claude Code at $2.5B annualized revenue while subsidizing usage at roughly 5x (offering $1,000 of compute inside $200 plans). Cursor doubling to $2B ARR in three months while both OpenAI and Anthropic burn margin to undercut it is the Uber/Lyft playbook — except the commodity being subsidized is inference, and the exit strategy is enterprise lock-in, not ride density. The sharpest buried signal: Tunguz's estimate of $36B consumer agent revenue vs. "the real money" in enterprise, combined with Codex's 8x traffic growth requiring new data centers, reveals that the AI labs are building a consumer acquisition funnel they can't yet afford to run at scale.

New York Times 2026-03-22-3

Tokenmaxxing: When AI Productivity Becomes Productivity Theater

Roose names "tokenmaxxing" — engineers competing on internal leaderboards for token consumption — but buries the only question that matters: nobody measures output quality. One OpenAI engineer burned 210 billion tokens in a week; a single Anthropic user ran up $150K in a month. The leaderboards track input volume, not output value. This is lines-of-code metrics reborn: Goodhart's Law applied to AI inference. The sharper signal is a Figma user consuming $70K in Claude tokens through a $20/month account, revealing that every SaaS platform offering AI at flat rate is running a margin time bomb. The companies that win this cycle won't consume the most tokens; they'll have the best ratio of useful output to tokens spent. That measurement layer doesn't exist yet.