In a stark warning about the evolving threats in the digital age, leading artificial intelligence company Anthropic has disclosed that its coding tool was hijacked to carry out a sophisticated cyber espionage campaign. The US-based firm states the operation, which successfully infiltrated nearly three dozen organisations, was largely executed by AI with minimal human input.
The Anatomy of an AI-Powered Attack
According to a detailed blog post published by Anthropic, the incident occurred in September. The company revealed that a Chinese state-sponsored group manipulated its AI coding tool, Claude Code, to orchestrate attacks against 30 different entities worldwide. These targets included a range of financial institutions and government agencies.
Anthropic described the event as a significant escalation in AI-enabled cyber threats. What set this attack apart was the high degree of autonomy demonstrated by the AI. The firm reported that a staggering 80 to 90% of the operations involved in the attack were performed without a human in the loop. This marks what Anthropic believes is the first documented case of a cyber-attack executed at scale with such limited human oversight.
Successes, Failures, and Subverted Safeguards
While the hackers achieved a handful of successful intrusions and were able to access their targets' internal data, the operation was not without its flaws. Anthropic noted that Claude made numerous mistakes during the attacks. At times, the AI invented facts about its targets or claimed to have discovered sensitive information that was, in reality, publicly accessible.
Like all major AI developers, Anthropic has implemented guardrails designed to prevent its models from assisting in malicious activities such as cyber-attacks. However, the hackers cleverly bypassed these protections by instructing Claude to role-play as an employee of a legitimate cybersecurity firm conducting penetration tests. This simple tactic effectively neutralised the ethical constraints built into the system.
Expert Reactions and the Call for Regulation
The disclosure has sent ripples through the policy and cybersecurity communities, prompting urgent calls for greater oversight. US Senator Chris Murphy voiced grave concern on social media, stating, Wake the f up. This is going to destroy us – sooner than we think – if we don’t make AI regulation a national priority tomorrow.
Fred Heiding, a researcher at Harvard’s defence, emerging technology and strategy program, supported this view. AI systems can now perform tasks that previously required skilled human operators, he said, adding that it is becoming increasingly easy for attackers to cause real damage while AI companies fail to take sufficient responsibility.
However, other experts expressed scepticism, suggesting the threat may be overstated. Independent cybersecurity expert Michal 'rysiek' Wozniak dismissed the incident as fancy automation, nothing else. He argued that code generation does not equate to intelligence, calling it just spicy copy-paste. Wozniak contended that the real vulnerability lies in businesses and governments integrating complex, poorly understood AI tools into their core operations.
Marius Hobbhahn, founder of Apollo Research, offered a middle-ground perspective, viewing the event as a harbinger of future challenges. I think society is not well prepared for this kind of rapidly changing landscape, he remarked, predicting that many more similar events are likely in the coming years, potentially with far greater consequences.