Everyone's worried that AI's newest models are a hacker's dream weapon

3 min read Original article ↗

Top AI and government officials tell Axios CEO Jim VandeHei that Anthropic, OpenAI and other tech giants will soon release new models that are scary good at hacking sophisticated systems at scale.

The one to watch: Anthropic is privately warning top government officials that its not-yet-released model — currently branded "Mythos" — makes large-scale cyberattacks much more likely in 2026.

The model allows agents to work on their own with wild sophistication and precision to penetrate corporate, government and municipal systems. It's a hacker's dream weapon.

Fortune got its hands on an unpublished Anthropic blog post describing Mythos. The post said the model is "currently far ahead of any other AI model in cyber capabilities."

The threat is no longer theoretical, and will be exacerbated by employees testing agents without realizing they're making it easier for cybercriminals to hack their company.

Flashback: Late last year, Anthropic disclosed the first documented case of a cyberattack largely executed by AI — a Chinese state-sponsored group that used AI agents to autonomously hack roughly 30 global targets, with the AI handling 80–90% of tactical operations independently.

Here's why this is different: The new models are even better at powering agents to think, act, reason and improvise on their own without rest or pause or limitation.

At the same time, systems are more vulnerable because so many employees are firing up Claude, Copilot or other agentic models — often at home — and creating agents of their own.

The bottom line: Everyone working at every company in America needs to know right now the dangers of using agents, especially unsupervised, anywhere near sensitive information. Leaders need to hammer this home.

Go deeper: How Anthropic's Pentagon deal could get revived.

📈 If you're a CEO or on a CEO's team: Sign up now to join Jim's new Axios C-Suite weekly newsletter.