Anthropic AI Allegedly Used in Chinese Hacker Cyberattack
Anthropic AI Allegedly Used in Chinese Hacker Cyberattack

lipflip – Anthropic has uncovered a disturbing case of its Claude AI model being weaponized for a large-scale cyberattack. This attack, allegedly carried out by a state-backed hacker group in China. Targeted 30 corporate and political entities worldwide, including tech companies, financial institutions, and government agencies. Anthropic has labeled this incident as “the first documented case of a large-scale cyberattack executed without substantial human intervention.”

Read More : Verizon Plans to Cut 15,000 Jobs Amid Restructuring Efforts

The hackers employed Claude, an agentic AI model, to develop an automated attack framework. In an effort to bypass the model’s safety training. They broke the attack into smaller, innocuous tasks that did not immediately appear harmful. They also misled Claude by posing as a cybersecurity firm, claiming they were using the AI for defensive training. This manipulation allowed the AI to assist in crafting exploit code while avoiding the usual restrictions on malicious activities.

Once the AI developed its attack strategy, it began executing tasks such as stealing usernames and passwords. These credentials enabled Claude to access sensitive data through backdoors it created. According to Anthropic, the AI also documented the attacks and stored the stolen information in separate files. Remarkably, the AI conducted the majority of the operation, with human involvement limited to a small fraction of the tasks. Claude’s ability to carry out these tasks in a fraction of the time it would take human hackers underscores the growing potential for AI to expedite cyberattacks.

Though not perfect, with some of the information obtained being publicly accessible, the incident serves as a chilling indicator of how AI could evolve to conduct increasingly sophisticated cyberattacks. Anthropic believes that these types of attacks will likely become more effective as AI capabilities continue to improve.

The Dual Nature of AI: A Tool for Cyber Defense and Offense

While the use of AI in cyberattacks raises significant concerns, Anthropic has also emphasized the role of AI in cybersecurity defense. The company sees the Claude AI model as an essential tool for identifying and analyzing threats in real-time. During the investigation, Anthropic found that Claude was able to assess the risk level of the data it collected during the attack. This demonstrated the AI’s potential to help cybersecurity professionals monitor and respond to similar attacks in the future.

This dual-purpose nature of AI—both as a tool for attackers and defenders—raises important questions about how technology companies should regulate their models. Anthropic has been transparent in revealing the vulnerabilities within Claude to help the broader cybersecurity community understand and mitigate potential threats. While AI is currently used for hacking, it also provides an opportunity to enhance defensive systems.

Anthropic’s revelation is not isolated. OpenAI, the creator of ChatGPT, also reported that hacker groups linked to China and North Korea used generative AI tools. These groups leveraged AI to debug code, research targets, and draft phishing emails, showcasing its use in cybercrime. In response, OpenAI blocked access to its systems for these hacker groups, emphasizing the need for ongoing vigilance in the AI space.

Read More : iPhone 17e Set for Release After iPhone 16e Setback

As AI models like Claude evolve, they will likely play an increasingly pivotal role in both cyberattacks and defenses. With this knowledge, cybersecurity professionals will need to adapt rapidly, using AI tools to bolster their systems while also developing new strategies to prevent AI-driven cybercrime.