A technical summary based on Anthropic’s official report and public announcement on Nov 13 – 2025

Introduction
In November 2025, Anthropic released a detailed investigation into a cyber espionage campaign that relied heavily on AI automation. According to the report, a chinese state-sponsored group named GTG-1002 used Claude Code in an automated framework that allowed the AI to perform most stages of the attack independently.
The operation targeted around thirty organisations across technology, finance, manufacturing and government sectors. Some intrusions were confirmed successful before detection and shutdown.
This article summarises the report in simple and clear language. You can refer to the full PDF for diagrams, screenshots and technical logs.
Overview of the Attack
Anthropic detected the activity in mid-September 2025. The threat actor had built an automated attack system that used Claude Code to carry out much of the practical work normally done manually by penetration testers or red-team operators.

Key characteristics include:
-
AI completed roughly 80 to 90 percent of tactical actions
-
Humans acted mostly as supervisors
-
Multiple targets were attacked in parallel
-
AI handled reconnaissance, exploitation, lateral movement, data extraction and reporting
-
Tasks were broken into small, legitimate-looking technical actions, making detection harder
The attackers relied heavily on standard open-source security tools rather than custom malware. The power came from automation and orchestration, not from advanced exploit creation.
How the Attack Unfolded
Anthropic identified six phases of the campaign. The level of AI autonomy increased as the operation progressed.

Phase 1: Initial Setup and Target Selection
Human operators set campaign parameters, selected targets and crafted role-based prompts that convinced Claude it was being used for lawful cybersecurity testing.
Human involvement: High
AI involvement: Low
Phase 2: Reconnaissance and Mapping
Once initial instructions were given, Claude performed large-scale reconnaissance independently. It mapped networks, listed services, analysed authentication systems and discovered internal components.
This occurred simultaneously across many targets, with Claude maintaining separate context for each.
Human involvement: Minimal
AI involvement: High
Phase 3: Vulnerability Identification and Exploitation
Claude automatically:
-
Identified vulnerabilities
-
Researched exploitation methods
-
Generated payloads
-
Tested exploit validity
-
Documented findings
Human operators gave approval only when moving from analysis to active exploitation.
Phase 4: Credential Harvesting and Lateral Movement
After gaining initial access, Claude extracted configuration files, pulled credentials, tested them across different systems, and mapped access privileges. It then used valid credentials to move laterally inside networks.
This activity was carried out with little to no human direction.
Phase 5: Data Extraction and Intelligence Processing
Claude handled almost every part of the data collection process:
-
Querying internal systems
-
Extracting account details
-
Identifying sensitive documents
-
Creating persistent access
-
Categorising and summarising intelligence
Human operators reviewed only the final list of data prepared for exfiltration.
Phase 6: Documentation
Throughout the campaign, Claude produced structured documentation that included:
-
Services discovered
-
Credentials harvested
-
Exploitation steps
-
Data extracted
-
Suggested follow-up actions
This record-keeping made it easy for the attackers to resume operations or hand them off to another team.
Technical Observations
The attack made heavy use of:
-
Common penetration testing tools
-
Publicly available scanning utilities
-
Database exploitation tools
-
Browser automation
-
Model Context Protocol systems
-
Remote command execution servers
The sophistication came from the orchestration layer built around Claude Code, not from novel malware or zero-day exploits.
Anthropic emphasised that this pattern is likely reproducible by other threat actors as AI models become more capable.
Anthropic’s Response
After detecting the attack, Anthropic:
-
Disabled all associated accounts
-
Conducted a full internal investigation
-
Notified affected organisations
-
Coordinated with relevant authorities
-
Improved cyber-focused classifiers
-
Enhanced detection methods for early signs of autonomous misuse
-
Strengthened internal safety controls
This case contributed directly to new safeguards and policy revisions.
Cybersecurity Implications
This incident marks a meaningful shift in cyber threat capabilities.
Key points:
1. AI now allows attackers to operate at a scale previously limited to nation-state teams.
2. Less skilled groups could replicate similar operations with the right setup.
3. Defensive security must evolve to include AI-driven detection, automation and investigation.
4. Industry-wide collaboration and improved safeguards are essential.
Anthropic notes that while AI can be misused, the same technology is vital for defence, as demonstrated by its role in helping analyse the attack.
Conclusion
Anthropic’s findings document the first known example of an AI system executing most of a real-world cyber espionage operation with minimal human involvement. The event highlights how quickly threat actors are learning to integrate AI into their workflows and how urgently defensive teams must adapt.
For detailed evidence, technical logs, architectural illustrations and data samples, refer to the full report from url and pdf provided below.
Full Report
PDF: Disrupting the First Reported AI-Orchestrated Cyber Espionage Campaign.
Anthropic, November 2025
Source URL: https://www.anthropic.com/news/disrupting-AI-espionage