agent-gating

7 items

The New York Times 2026-04-27-2

Can an A.I. Company Ever Be Good?

OpenAI publicly calls for regulation while privately lobbying against liability, and the NYT opinion piece is right that this is structural, not situational. But the prescription stops short: the piece skips regulatory capture, GDPR-style implementation theater, and the near-zero track record of omnibus tech bills. The more useful frame for builders is that regulation is coming regardless, and most enterprise AI governance won't survive a hostile audit — the companies that build governance that actually holds are the ones that own the next cycle.

Wall Street Journal 2026-04-26-3

AI Is Cannibalizing Human Intelligence (Vivienne Ming, WSJ)

Ming's Polymarket experiment splits human-AI usage into three measurable patterns: oracle (use the answer), validator (use AI to confirm priors), cyborg (use AI as sparring partner). Validators perform worse than AI alone — sycophancy laundered as evidence — while the 5-10% of cyborgs match or beat prediction-market consensus. The unbuilt premium category is AI that disagrees with you on purpose; today's benchmarks measure what AI does alone, not whether the product is building human capacity or consuming it.

Financial Times 2026-04-20-1

Who is liable when artificial intelligence makes mistakes?

Insurers whose entire business is pricing unpredictable outcomes are declining to price AI, which is the strongest external validation yet that reliability, not capability, is the binding constraint on enterprise agent deployment. AIG is filing exclusions; Aon's risk chief is calling autonomous agents uninsurable. Same playbook as cyber insurance two decades ago: the carrier that builds AI loss data first captures the $10B-plus standalone category that emerges on the other side.

Wall Street Journal 2026-04-20-2

Marc Benioff Says the Software Bears Are All Wrong About Salesforce

Salesforce just disclosed 2.4 billion Agentic Work Units growing 57% quarter over quarter, with no dollar anchor attached and revenue still crawling at 10%. CEOs don't write op-eds when they're winning; 15.3% Agentforce penetration after 18 months reads as a chasm signal, not acceleration, and Kimbarovsky sold shares from the exact article Benioff sanctioned. The scaffolding moat is real for regulated enterprise, but the AWU-without-price pattern is stage one of a per-seat-to-per-action transition Salesforce hasn't finished pricing yet.

The Verge / Decoder 2026-04-20-3

Canva's Big Pivot to AI: Editable Output as Agentic SaaS Moat

Perkins named the taxonomy that will split agentic SaaS winners from losers: AI 1.0 is one-shot, AI 2.0 is iterative. The real bet isn't the model or the generation quality; it's where the output lands. Canva's decade of interoperable layered-format investment is the scaffolding that lets the agent hand you back an editable file instead of a dead-end artifact, which is how the ServiceNow/Salesforce playbook plays out one tier down in the consumer-to-enterprise funnel. Architecture, token economics, and platform-encroachment risk all got deflected; the format moat is the one claim that survived scrutiny.

Latent Space 2026-04-07-2

Extreme Harness Engineering for Token Billionaires: 1M LOC, 0% Human Code, 0% Human Review

OpenAI's Frontier team built a 1M-line Electron app with zero human-authored code: the competitive advantage wasn't the model, it was six skills encoding what "good" looks like as text. The real shift here isn't AI writing code; it's AI inheriting engineering culture. Ghost libraries (distributing specs instead of code) and Symphony (an Elixir orchestrator the model chose for its process supervision primitives) point to a future where the scarce resource is institutional knowledge distillation, not developer headcount.

Fortune 2026-03-23-2

The Karpathy Loop: Autonomous Agent Optimization as Research Pattern

Karpathy's autoresearch ran 700 experiments in two days on a 630-line codebase: the result matters less than the pattern. The Karpathy Loop (agent + single file + testable metric + time limit) is the atomic unit of constrained autonomous optimization, and it generalizes to any problem with a measurable output and a modifiable code surface. The real competitive shift isn't building better agents; it's designing better constraints, metrics, and stopping criteria: taste becomes the bottleneck, not compute.