Anthropic: No "kill switch" for AI in classified settings

2 min read Original article ↗

Anthropic says it has no way to control or shut down its AI models once they're deployed by the Pentagon, according to a new court filing.

Why it matters: The Pentagon designated Anthropic a supply chain risk, contending the AI firm is inappropriately getting involved in how its technology can be used in sensitive military operations.

What's inside: Anthropic argues in the filing to a federal appeals court in D.C. that it has no visibility, technical ability or any kind of "kill switch" for its technology once it's deployed.

Catch up quick: The company's usage policies include no Claude for autonomous weapons or mass surveillance, red lines that the Pentagon dismissed as red herrings and led to the dispute.

Friction point: The Pentagon is arguing in court that Anthropic is a supply chain risk as the Trump administration moves to deploy its new Mythos model across the federal government.

What's next: A hearing is scheduled for May 19.