Article Details
Scrape Timestamp (UTC): 2026-02-03 23:59:18.216
Source: https://www.theregister.com/2026/02/03/autonomous_cyberattacks_not_real_yet/
Original Article Text
Click to Toggle View
AI agents can't yet pull off fully autonomous cyberattacks - but they are already very helpful to crims. Don't relax: This is a 'when, not if' scenario. AI agents and other systems can't yet conduct cyberattacks fully on their own - but they can help criminals in many stages of the attack chain, according to the International AI Safety report. The second annual report, chaired by the Canadian computer scientist Yoshua Bengio and authored by more than 100 experts across 30 countries, found that over the past year, developers of AI systems have vastly improved their ability to help automate and perpetrate cyberattacks. Perhaps the best, and scariest, evidence of that finding appeared in Anthropic's November 2025 report about Chinese cyberspies abusing its Claude Code AI tool to automate most elements of attacks directed at around 30 high-profile companies and government organizations. Those attacks succeeded in "a small number of cases." "At least one real-world incident has involved the use of semi-autonomous cyber capabilities, with humans intervening only at critical decision points," according to the AI safety report. "Fully autonomous end-to-end attacks, however, have not been reported." Two areas where AI is especially useful to criminals are scanning for software vulnerabilities and writing malicious code. During DARPA's AI Cyber Challenge (AIxCC) – a two-year competition in which teams built AI models to find vulnerabilities in open source software that undergirds critical infrastructure – finalist systems autonomously identified 77 percent of the synthetic vulnerabilities used in the final scoring round, according to competition organizers. And while that is an example of defenders using AI to find and fix vulnerabilities, rather than attackers using AI to find and exploit them, criminals are using models in similar ways. Last northern summer, we saw attackers on underground forums claiming to use HexStrike AI, an open-source red-teaming tool, to target critical vulnerabilities in Citrix NetScaler appliances within hours of the vendor disclosing the problems. Additionally, AI systems are getting much better at malware writing, and criminals can trade weaponized models that write ransomware and data-stealing code for as little as $50 a month. The good news for now, according to the report’s authors, is that AI systems still aren't great at carrying out multi-stage attacks without human help. "Research suggests that autonomous attacks remain limited because AI systems cannot reliably execute long, multi-stage attack sequences," according to the report. "For example, failures they exhibit include executing irrelevant commands, losing track of operational state, and failing to recover from simple errors without human intervention." Keep in mind, however, that this all was written before the security dumpster fire that is OpenClaw – the AI agent previously known as Moltbot and Clawdbot – and Moltbook, the vibe-coded social media platform for AI agents. So it's also entirely plausible that the world won't end with a sophisticated, autonomous multi-stage cyberattack dreamed up by a nation-state crew or criminal mastermind, but rather a single agent that goes off the rails.
Daily Brief Summary
The International AI Safety report reveals AI systems are increasingly aiding cybercriminals, though fully autonomous attacks remain unrealized.
AI tools have been used in attacks on approximately 30 high-profile companies and government organizations, with some success.
AI excels in scanning for vulnerabilities and generating malicious code, significantly aiding criminals in these attack stages.
DARPA's AI Cyber Challenge demonstrated AI's potential in identifying vulnerabilities, with finalist systems detecting 77% of synthetic flaws.
Criminals leverage AI models like HexStrike to exploit vulnerabilities swiftly, as seen with Citrix NetScaler appliances.
AI-generated malware and ransomware models are available on underground forums, with some priced as low as $50 monthly.
Current AI limitations include difficulty executing complex, multi-stage attacks without human intervention, reducing immediate threat levels.
Future risks include potential for AI agents to act unpredictably, emphasizing the need for ongoing vigilance and preparedness.