defensibility

6 items

tanyaverma.sh 2026-04-13-1

The Closing of the Frontier

Two-thirds of MATS symposium research posters ran on Chinese open-source models because Anthropic's Mythos restrictions closed off Western frontier access to independent safety researchers. The safety case for restricted access is degrading the safety research pipeline it claims to protect. The policy question isn't content moderation: it's whether frontier model access needs due process obligations the way utilities do.

OpenAI · 2026-03-09 2026-03-13-w2

Codex Security: now in research preview

Codex Security shipped with receipts: 15 named CVEs, published noise-reduction curves showing 84% improvement, and false positive rates cut by over 50%, giving enterprise buyers metrics to evaluate rather than claims to trust. The structurally interesting detail is the threat model architecture, which builds an editable intermediate artifact before scanning, making the agent's reasoning inspectable before execution. That pattern generalizes well beyond security, but it sits in direct tension with the cognitive load data surfacing elsewhere this week: if inspecting the agent's intermediate state is what makes it trustworthy, the oversight burden migrates rather than shrinks. Broad tier access from Pro through Edu maximizes adoption velocity while quietly undermining any dual-use containment argument either lab has made. The CISO budget is the Trojan horse for the engineering budget, and both labs are through the door.

Reuters / The Information 2026-03-11-1

OpenAI Building GitHub Competitor

The outage origin story is cover for the real move: at $840B, OpenAI needs platform economics, not API margins. Owning where AI agents commit code is more defensible than selling tokens. The buried signal is "considered making it available for purchase" — you don't leak commercialization plans for an internal workaround. The Microsoft relationship tension (49% owner's crown jewel being targeted) is the governance story nobody is writing.

Pirate Wires 2026-03-11-2

Inside the Culture Clash That Tore Apart the Pentagon's Anthropic Deal

Michael's account reveals the structural impossibility of scenario-by-scenario AI usage carveouts at military scale — but his sabotage hypothetical (lasers intentionally defective) exposes that the 'supply-chain risk' designation is built on speculation, not evidence. The real signal: 'all lawful use' is becoming the default for defense AI contracts, forcing every AI company to choose between the defense market and the safety brand. Anthropic is implicitly betting the commercial market is larger — and the blacklisting may accidentally prove them right by strengthening enterprise trust.

Anthropic 2026-03-09-1

Making frontier cybersecurity capabilities available to defenders

Product announcement dressed as research disclosure. Claude Code Security uses multi-stage self-verification to scan codebases beyond pattern-matching SAST. The 500-vuln claim has no CVEs, no false positive rates, and no comparison to existing tools. Zero external validation in the announcement itself -- the WSJ/Firefox piece did that work. The real play: security scanning as a loss-leader wedge for enterprise platform deals. Neither lab announced pricing.

OpenAI 2026-03-09-2

Codex Security: now in research preview

Same-day competitive counter to Anthropic with stronger receipts: 15 named CVEs in the appendix (GnuTLS heap overflows, GnuPG stack buffer overflow, GOGS 2FA bypass), published improvement curves (84% noise reduction, 90%+ severity over-reporting reduction, 50%+ false positive reduction). The threat model architecture -- building an editable intermediate artifact before scanning -- is the most interesting pattern: it generalizes as "make the agent's understanding inspectable before execution." Broader tier access (Pro through Edu) weakens the dual-use containment narrative but maximizes adoption velocity.