What Anthropic’s Mythos Means for the Future of Cybersecurity

Two weeks ago, Anthropic announced that its new model, Claude Mythos Preview, can autonomously find and weaponize software vulnerabilities, turning them into working exploits without expert guidance. These were vulnerabilities in key software like operating systems and internet infrastructure that thousands of software developers working on those systems failed to find. This capability will have major security implications, compromising the devices and services we use every day. As a result, Anthropic is not releasing the model to the general public, but instead to a limited number of companies.

The news rocked the internet security community. There were few details in Anthropic’s announcement, angering many observers. Some speculate that Anthropic doesn’t have the GPUs to run the thing, and that cybersecurity was the excuse to limit its release. Others argue Anthropic is holding to its AI safety mission. There’s hype and counterhype, reality and marketing. It’s a lot to sort out, even if you’re an expert.

We see Mythos as a real but incremental step, one in a long line of incremental steps. But even incremental steps can be important when we look at the big picture.

How AI Is Changing Cybersecurity

We’ve written about shifting baseline syndrome, a phenomenon that leads people—the public and experts alike—to discount massive long-term changes that are hidden in incremental steps. It has happened with online privacy, and it’s happening with AI. Even if the vulnerabilities found by Mythos could have been found using AI models from last month or last year, they couldn’t have been found by AI models from five years ago.

The Mythos announcement reminds us that AI has come a long way in just a few years: The baseline really has shifted. Finding vulnerabilities in source code is the type of task that today’s large language models excel at. Regardless of whether it happened last year or will happen next year, it’s been clear for a while this kind of capability was coming soon. The question is how we adapt to it.

We don’t believe that an AI that can hack autonomously will create permanent asymmetry between offense and defense; it’s likely to be more nuanced than that. Some vulnerabilities can be found, verified, and patched automatically. Some vulnerabilities will be hard to find but easy to verify and patch—consider generic cloud-hosted web applications built on standard software stacks, where updates can be deployed quickly. Still others will be easy to find (even without powerful AI) and relatively easy to verify, but harder or impossible to patch, such as IoT appliances and industrial equipment that are rarely updated or can’t be easily modified.

Then there are systems whose vulnerabilities will be easy to find in code but difficult to verify in practice. For example, complex distributed systems and cloud platforms can be composed of thousands of interacting services running in parallel, making it difficult to distinguish real vulnerabilities from false positives and to reliably reproduce them.

So we must separate the patchable from the unpatchable, and the easy to verify from the hard to verify. This taxonomy also provides us guidance for how to protect such systems in an era of powerful AI vulnerability-finding tools.

Unpatchable or hard to verify systems should be protected by wrapping them in more restrictive, tightly controlled layers. You want your fridge or thermostat or industrial control system behind a restrictive and constantly updated firewall, not freely talking to the internet.

Distributed systems that are fundamentally interconnected should be traceable and should follow the principle of least privilege, where each component has only the access it needs. These are bog-standard security ideas that we might have been tempted to throw out in the era of AI, but they’re still as relevant as ever.

Rethinking Software Security Practices

This also raises the salience of best practices in software engineering. Automated, thorough, and continuous testing was always important. Now we can take this practice a step further and use defensive AI agents to test exploits against a real stack, over and over, until the false positives have been weeded out and the real vulnerabilities and fixes are confirmed. This kind of VulnOps is likely to become a standard part of the development process.

Documentation becomes more valuable, as it can guide an AI agent on a bug-finding mission just as it does developers. And following standard practices and using standard tools and libraries allows AI and engineers alike to recognize patterns more effectively, even in a world of individual and ephemeral instant software—code that can be generated and deployed on demand.

Will this favor offense or defense? The defense eventually, probably, especially in systems that are easy to patch and verify. Fortunately, that includes our phones, web browsers, and major internet services. But today’s cars, electrical transformers, fridges, and lampposts are connected to the internet. Legacy banking and airline systems are networked.

Not all of those are going to get patched as fast as needed, and we may see a few years of constant hacks until we arrive at a new normal: where verification is paramount and software is patched continuously.

This essay was written with Barath Raghavan, and originally appeared in IEEE Spectrum.

Posted on April 28, 2026 at 7:06 AM10 Comments

Comments

mw April 28, 2026 7:32 AM

Most of the unpatchable Systems are of bad design. It does not matter if these systems are a couple of years old or from today. The bad design is everywhere. Best are very old systems, they do not have the ability to communicate. Internet technology is very helpful to develop, deploy and debug industrial systems. But the designers should disconnect such systems from the public networks. Secure and safe devices are possible, but expensive. As long as device and system designer are not liable for their bad designs nothing will change.

Clive Robinson April 28, 2026 11:45 AM

@ Bruce, ALL,

With regards,

“Even if the vulnerabilities found by Mythos could have been found using AI models from last month or last year, they couldn’t have been found by AI models from five years ago.”

Hmmm when was the transformer paper “Attention Is All You Need ” published by Google?

The reality is that Expert Systems and Fuzzy Logic were doing “pattern matching” quite some time before that.

What we don’t know is who was using the equivalent of “Digital Signal Processing”(DSP) “Matched adaptive filters” prior to that.

I was certainly talking about how this could be done on this blog with “Known Knowns” etc quite a while before Anthropic went public.

And that is the point realy,

We only knew it when someone went public about it.

It could easily have been more than a decade ago, but nobody chose to say anything…

There is something known as the “Defence Spending Conundrum” of,

You never know when you’ve spent to much on defence, but you tend to find out when you not spent enough… because somebody attacks you.

That applies as much if not more so to “Defence Research” as it does to deployable weapons.

It’s a hard issue to come to terms with because it’s also how “arms races” start, and they are usually a disaster in the making for every one who gets into one.

dave April 28, 2026 11:55 AM

What I don’t think anyone is facing is the psychological impact this is having on the end user. Before AI the individual could depend on the fact that finding and exploiting bugs was a resource intensive endeavor and that this fact represented a barrier to entry. AI significantly reduces that barrier to entry.

where verification is paramount and software is patched continuously.

Because no one, including me, is ever going to want to do that. Just like the dead internet theory I am coming to think of it as the dead phone theory. To quote Mad Eye Moody, “constant vigilance”. If that is the price of entry I don’t want to play. “Patch Tuesdays” were a hassle but patching everything everyday? I’m out.

Rontea April 28, 2026 2:23 PM

This development is both inevitable and unsurprising. Mythos is simply the latest in a series of incremental advances that reflect where AI-assisted vulnerability research has been heading for years.
The lesson is the same one: secure systems through good architecture, least privilege, proper documentation, and continuous testing. AI isn’t going to change that. It just raises the cost of ignoring it.

lurker April 28, 2026 2:52 PM

@Dave

“patching everything everyday” should not be necessary when AI is used in the development and testing chain. IF (big IF) this AI is so good at finding bugs (and some people think it is not, see counterhype, reality, above), then it should be compulsory for software vendors to use it on their products before releasing them to the public.

Oh, but of course software isn’t a “product”. You can’t hold it in your hand, or break it with a hammer, so none of the usual consumer protection laws apply to it.

somebody April 28, 2026 3:39 PM

AI may be good at finding some bugs and bad at finding all bugs. This favours the offence, and depending on the exact details developers deploying AI may not help very much at all. If AI can find a random bug with 1% effectiveness the offense always wins.

It remains much better to try to design and write code without bugs than to create random text strings and make changes till it compiles, call it code and then run it through AI and make more changes.

It’s best to go back and take Hoare’s path not taken: So simple there are obviously no errors.

Clive Robinson April 28, 2026 5:32 PM

@ Dave,

It appears to be your turn in the barrel, not that you really deserve it…

With regards,

“Before AI the individual could depend on the fact that finding and exploiting bugs was a resource intensive endeavor and that this fact represented a barrier to entry. “

A “human resource” is what you should have said… AI is in many respects a way way more expensive resource, and after a few early gains it’s “productivity will tank”.

The reason is it’s finding variations on the simplest of instances in well defined classes of vulnerabilities, ie “known knowns”. It will only find “unknown knowns” by the probability equivalent to current fuzzing which is to ve honest quite low. Also Current AI LLM and ML systems provably lack the ability to find “unknown unknowns” or the “Black Swan” type attacks that can only be reasoned out as there is no pattern to match against in the training data set…

These are things I’ve pointed out on a number of occasions going back quite some time, but “they are facts that some don’t want known”…

But moving on, you note,

“Because no one, including me, is ever going to want to do that. Just like the dead internet theory I am coming to think of it as the dead phone theory.”

Well… To be brutally honest, I could truthfully say,

“It’s your own fault for not being an engineer, but being an artisan.”

But that is not asking the real question that should be asked,

Has management and marketing ever allowed software developers to be real engineers?

To which the answer is of course “NO” and the excuse they give is “productivity” or similar nonsense based on screwing maximum lines of code / person / day with no regard to “quality” or “safety” of which “security / Privacy” are just subsets.

People moan that “hardware development is slow” with “software development being fast”…

The reason of course is “legislation and regulation” have a look at all of that which “hardware has to comply with” yet “software mostly does not” because aside from niche activities there is no real legislation or regulation that effects software development or the developers… That is of course not true for hardware development and developers, where very real criminal penalties / consequences apply.

Are you prepared to put in the training and take the consequent 50% pay cut?

Yup have a look at pay / conditions of specialised EE jobs like R.F., Telecommunications, Industrial Safety, Aerospace Safety, Medical Safety, and many more “safety domains”.

I’m trained in many of them that makes me fairly unique, even more so when you throw in Telecoms and Radio Frequency(RF). In return if I was employed in the way most software developers are, I’d be paid less than a third of that of an also ran software team lead…

Am I griping about it,

“being “unfair?”

Not really it was always my choice where I went and how I got there. But increasingly Management wanted me to be the equivalent of not just a Software Team Lead, Project Manager, Contracts Manager, etc yet also be an expert engineer prepared to put my entire future career on a legal dotted line to absolve them of any responsibility… yet be paid next to nothing…

It’s just one reason why I just don’t do that nonsense any more.

GottaChimeInHere April 28, 2026 6:51 PM

I just have to chime in here:

Note the dates, about 6.5 weeks apart.

For now, AI is better at finding security flaws than producing working exploits from those flaws. But AI, like breaking encryption, just keeps improving with time.

At the very least, Open Source is proving convenient for demonstrating AI capabilities. Firefox is now far more secure. And commercial software vendors badly want access.

Magnus April 28, 2026 11:03 PM

<

blockquote>dave • April 28, 2026 11:55 AM
finding and exploiting bugs was a resource intensive endeavor and that this fact represented a barrier to entry

<

blockquote>

It still is, as everyone knows generative AI bots use a huge amount of electricity and computing resources, especially for big PR stunts like this. The big AI companies have been secretive about how much energy and computation time have been put into the results they’ve hyped (maths olympiads, coding competitions, high profile mathematical proofs, etc).

What’s even worse is it’s by design; the more power they need, the better. It’s a moat against competitors and customers doing it themselves.

On the topic in general, I think it’s laughable that an industry with so much hype and misleading bull****, and that doesn’t give a crap about the harm its products do (like talking teenagers into suicide), all of a sudden is so concerned about system vulnerabilities. Especially when the rest of the industry has paid so little attention to it in the first place.

And OpenAI putting its hand up and exclaiming “Ooh! Ooh! Our new system is also so powerful and dangerous we won’t be releasing it to the public either!” Can these people hear themselves?

Clear to me that Mythos isn’t ready for the public, which would have gotten another big laugh at the lack of substance compared to the hype; that it’s still full of problems that the 50 “exclusive” companies are in a beta-testing programme for.

And if the system really can churn out exploits, do we really think that not one single employee, among who knows how many employees at 50 (50!) companies won’t pass on a nasty exploit to hacker friends if they find one?

Leave a comment

Blog moderation policy

Login

Allowed HTML <a href="URL"> • <em> <cite> <i> • <strong> <b> • <sub> <sup> • <ul> <ol> <li> • <blockquote> <pre> Markdown Extra syntax via https://michelf.ca/projects/php-markdown/extra/

Sidebar photo of Bruce Schneier by Joe MacInnis.