ai-regulation
2 items · chronological order
Does A.I. Need a Constitution?
Lepore traces Claude's Constitution from the Capitol insurrection through Anthropic's founding to its 30,000-word moral framework: corporate governance filling a vacuum left by democratic failure. Five constitutional law professors independently critique the borrowed-legitimacy play: calling it a "constitution" creates expectations the document can't meet. The piece's biggest gap is also its most revealing: Lepore never asks whether character-based training actually works, because her thesis requires it not to matter. For enterprises, the real signal is upstream: every AI vendor choice now inherits a governance framework as a liability, and the next regulatory window will punish self-regulation as insufficient regardless of sincerity.
Anthropic Opposes the Extreme AI Liability Bill That OpenAI Backed
Illinois SB 3444 would grant AI developers blanket liability immunity for catastrophic harm if they publish their own safety framework — no external audit, no enforcement. OpenAI backs it; Anthropic is lobbying to kill it. Self-certification has never survived contact with high-consequence outcomes: aviation, pharma, and nuclear all tried it and produced catastrophic failures before external verification became mandatory. AI labs are now writing the legal architecture that determines whether they face accountability at all.