ai-security

6 items

The New Yorker 2026-04-26-1

When Your Digital Life Vanishes

DriveSavers' ransomware recoveries went 6x in two years: under 50 in 2023, nearly 300 in 2025, with the firm's ransomware lead naming AI directly as the multiplier turning unsophisticated IT operators into sophisticated attackers. Buried in the same New Yorker piece: data center proliferation is wildly inflating storage costs, AI agents are now "notorious" for accidental deletions, and HDD lifespan stays flat at seven years even as Seagate ships 44TB drives. The cloud-abundance narrative has the order book pointed the wrong way — the AI revolution is also a data destruction revolution, and the recovery industry is the only place reading the signal correctly.

tanyaverma.sh 2026-04-13-1

The Closing of the Frontier

Two-thirds of MATS symposium research posters ran on Chinese open-source models because Anthropic's Mythos restrictions closed off Western frontier access to independent safety researchers. The safety case for restricted access is degrading the safety research pipeline it claims to protect. The policy question isn't content moderation: it's whether frontier model access needs due process obligations the way utilities do.

UK AI Security Institute 2026-04-13-3

AISI Evaluation of Claude Mythos Preview's Cyber Capabilities

A UK government lab confirmed Mythos can autonomously execute a 32-step corporate network attack end-to-end, outperforming every tested model including GPT-5, with performance still scaling at the 100M token ceiling. The evaluation tested capability against undefended ranges, so what AISI validated is threat potential, not operational impact against a real defended environment. The structural shift is that government evaluation infrastructure is becoming the third-party verification layer for frontier AI claims, sitting between self-reported lab benchmarks and the market the way FDA trials sit between pharma and prescribers.

Alex Kim's Blog 2026-04-04-2

Claude Code Source Leak: Anti-Distillation DRM, KAIROS Autonomous Mode, and the Defensive Architecture

The Claude Code source leak is most interesting for what the defensive architecture reveals: anti-distillation via fake tool injection, Zig-level client attestation below the JS runtime, and undercover mode that strips AI attribution from open-source commits — each individually bypassable within hours by anyone who reads the activation logic. The more significant find is KAIROS, an unreleased autonomous daemon with GitHub webhooks, nightly memory distillation, and cron-scheduled refresh every five minutes, showing Anthropic is building always-on background agents, not session-based assistants. The leak itself was a known Bun bug left unpatched for 20 days — the gap between what Anthropic built and what it shipped is the operational risk signal, not the defensive code.

Sockpuppet.org 2026-04-01-3

Vulnerability Research Is Cooked

Every IT department runs on a hidden subsidy: the scarcity of people smart enough to hack them. Anthropic's Frontier Red Team just demonstrated 500 validated high-severity vulnerabilities from a trivial bash script and Claude Opus 4.6, no fuzzers, no specialized tooling, just raw model inference. The Bitter Lesson is about to hit security like a brick: 80% of exploit development was jigsaw-puzzle grinding, and now everyone has a universal solver. The scarce resource isn't intelligence anymore; it's the ability to patch faster than agents can find what's broken.

WIRED 2026-03-18-3

Justice Department Says Anthropic Can't Be Trusted With Warfighting Systems

The DOJ's filing reveals a dependency it was supposed to prevent: Claude is currently the only AI model cleared for classified DOD systems, which means the supply-chain risk designation is partly a self-inflicted wound. The government's argument that Anthropic "could" sabotage warfighting systems conflates a vendor's contractual right to set usage terms with criminal sabotage, and the distinction matters for every AI company negotiating enterprise AUPs. The real signal is structural: safety restrictions are now priced as commercial liability in the defense market, and the replacement vendors inheriting these contracts gain not just revenue but classified use-case intelligence that compounds for years.