Using LLMs to Exploit Vulnerabilities
Interesting research: “Teams of LLM Agents can Exploit Zero-Day Vulnerabilities.”
Abstract: LLM agents have become increasingly sophisticated, especially in the realm of cybersecurity. Researchers have shown that LLM agents can exploit real-world vulnerabilities when given a description of the vulnerability and toy capture-the-flag problems. However, these agents still perform poorly on real-world vulnerabilities that are unknown to the agent ahead of time (zero-day vulnerabilities).
In this work, we show that teams of LLM agents can exploit real-world, zero-day vulnerabilities. Prior agents struggle with exploring many different vulnerabilities and long-range planning when used alone. To resolve this, we introduce HPTSA, a system of agents with a planning agent that can launch subagents. The planning agent explores the system and determines which subagents to call, resolving long-term planning issues when trying different vulnerabilities. We construct a benchmark of 15 real-world vulnerabilities and show that our team of agents improve over prior work by up to 4.5×.
The LLMs aren’t finding new vulnerabilities. They’re exploiting zero-days—which means they are not trained on them—in new ways. So think about this sort of thing combined with another AI that finds new vulnerabilities in code.
These kinds of developments are important to follow, as they are part of the puzzle of a fully autonomous AI cyberattack agent. I talk about this sort of thing more here.
Subscribe to comments on this entry
Clive Robinson • June 17, 2024 7:52 AM
@ Bruce, ALL,
That is somewhat ambiguous.
If it’s actually a “zero day” then by the definition it’s unknown thus not in the LLM “weights”.
In part that is why the word “toy” appears in,
And the zero-day definition of “unknown” in
But
“How unknown?”
Is an important question.
As I’ve noted in the past there are
1, “Instances of vulnerability”
2, “Classes of vulnerability”.
Thus if the LLM knows sufficient “Instances”, to find the “Class” they are in. Then finding new “Instances” in the “Class” is well within the “stochastic parrot” description.
Especially if other “instances” in “other classes” that have some commonality are within the LLM weightings.
It’s actually not a new idea. Most cyber-attacks are not actually new, and some go back hundreds of years in conventional physical world attacks, that have just been ported across.
So old sour wine in new shinny bottles.
Thus a secondary question arises,
“Can an LLM trained up with physical world attacks cross them over to information world attacks?”
I’m reasonably certain the answer is yes.
Because although it might look like the LLM has invented a new vulnerability attack, in reality it has not, just found commonality and added a little randomisation.