ai-defensibility

2 items

Morningstar 2026-03-18-2

Morningstar's Largest-Ever Moat Review: 37 Downgrades and the Two Upgrades That Matter More

Morningstar halved its moat duration horizon for application-layer software from 20 years to 10, triggering 37 downgrades in the largest review since the firm started rating moats. The fair value cuts (Adobe at 32%, ServiceNow at 18%, Salesforce at 7%) are a lagging indicator: these stocks were already down 20-30% before the methodology caught up. The buried signal is in the two upgrades: CrowdStrike and Cloudflare both went to wide moat because AI expands the attack surface and network traversal that security infrastructure must handle. When 37 moats narrow and two widen, the widening tells you where the new toll bridges are.

WIRED 2026-03-18-3

Justice Department Says Anthropic Can't Be Trusted With Warfighting Systems

The DOJ's filing reveals a dependency it was supposed to prevent: Claude is currently the only AI model cleared for classified DOD systems, which means the supply-chain risk designation is partly a self-inflicted wound. The government's argument that Anthropic "could" sabotage warfighting systems conflates a vendor's contractual right to set usage terms with criminal sabotage, and the distinction matters for every AI company negotiating enterprise AUPs. The real signal is structural: safety restrictions are now priced as commercial liability in the defense market, and the replacement vendors inheriting these contracts gain not just revenue but classified use-case intelligence that compounds for years.