The burgeoning field of artificial intelligence faces a critical and escalating threat: industrial-scale “distillation attacks.” Recent findings from Anthropic expose sophisticated campaigns by rival AI laboratories illicitly extracting advanced capabilities from frontier models like Claude. These attacks don’t just violate terms of service; they pose significant national security risks, undermine export controls, and endanger the responsible development of powerful AI. Understanding and combating these pervasive threats requires rapid, coordinated action across the global AI ecosystem.
What Exactly Are AI Distillation Attacks?
AI distillation is a technique where a less capable AI model is trained on the outputs of a stronger, more advanced one. In its legitimate form, distillation is a common and valuable training method. Frontier AI labs, for instance, often distill their own large models into smaller, more efficient versions for commercial deployment or customer use. This process reduces computational cost and latency.
However, the technique can be weaponized for illicit purposes. Competitors can use distillation to rapidly acquire powerful AI capabilities from other labs. This allows them to bypass the immense time, resources, and expertise required for independent development. Such illicit AI distillation attacks represent a significant form of intellectual property theft and unfair competition.
The Alarming Scale of Illicit Extraction
Anthropic’s investigation uncovered industrial-scale campaigns orchestrated by three specific AI laboratories: DeepSeek, Moonshot, and MiniMax. These labs generated over 16 million interactions with Claude through approximately 24,000 fraudulent accounts. This extensive activity blatantly violated Anthropic’s terms of service and regional access restrictions. The sheer volume and organized nature of these exchanges highlight a deliberate effort to extract core AI capabilities rather than engage in legitimate research or usage.
The Grave Risks Posed by Illicit Distillation
The implications of widespread AI distillation attacks stretch far beyond corporate competition. They introduce severe risks that impact national security, global stability, and the integrity of AI development.
National Security Implications
One of the most pressing concerns is the lack of necessary safeguards in illicitly distilled models. Leading US companies like Anthropic embed robust safety mechanisms into their AI systems. These safeguards prevent misuse by state and non-state actors, such as developing bioweapons or orchestrating malicious cyber activities. Models built through illicit distillation, however, are unlikely to retain these critical protections. This means dangerous AI capabilities could proliferate globally with many of their essential safeguards stripped away.
Foreign labs that illicitly distill American models can then integrate these unprotected capabilities into military, intelligence, and surveillance systems. This empowers authoritarian governments to deploy frontier AI for offensive cyber operations, large-scale disinformation campaigns, and pervasive mass surveillance. If these distilled models are subsequently open-sourced, the risk multiplies exponentially. Such dangerous capabilities can then spread freely, falling beyond any single government’s control and increasing global instability.
Undermining Export Controls
Anthropic has consistently advocated for robust export controls to help maintain America’s lead in advanced AI technologies. AI distillation attacks directly undermine these crucial controls. They allow foreign labs, including those subject to the control of entities like the Chinese Communist Party, to close the competitive advantage that export controls are designed to preserve. These labs achieve this through surreptitious means, circumventing legal frameworks.
Without clear visibility into these sophisticated attacks, seemingly rapid advancements by certain foreign labs can be mistakenly attributed to independent innovation. This perception might incorrectly suggest that export controls are ineffective or easily bypassed. In reality, such advancements often depend significantly on capabilities covertly extracted from American models. Executing this extraction at scale further requires access to advanced chips. Therefore, AI distillation attacks reinforce the fundamental rationale for export controls: restricting chip access limits both direct model training and the scale of illicit distillation.
A Deep Dive into the Attack Playbook
The three identified distillation campaigns followed a strikingly similar pattern. They used fraudulent accounts and commercial proxy services to access Claude at an industrial scale while actively evading detection.
Sophisticated Access Mechanisms
For national security reasons, Anthropic does not offer commercial access to Claude in China, nor to subsidiaries of Chinese companies located outside the country. To bypass these restrictions, malicious labs employ commercial proxy services. These services resell access to Claude and other frontier AI models. They operate what Anthropic terms “hydra cluster” architectures. These sprawling networks consist of thousands of fraudulent accounts, distributing traffic across Anthropic’s API and various third-party cloud platforms.
The immense breadth of these networks ensures there are no single points of failure. If one account is banned, another swiftly takes its place. In one instance, a single proxy network managed over 20,000 fraudulent accounts simultaneously. To further complicate detection, they often mixed distillation traffic with unrelated, legitimate customer requests.
Targeted Capability Extraction
Once illicit access is secured, the labs generate massive volumes of meticulously crafted prompts. These prompts are specifically designed to extract particular capabilities from the target model. The goal is either to collect high-quality responses for direct model training or to generate tens of thousands of unique tasks necessary for reinforcement learning.
What differentiates an AI distillation attack from normal usage is the distinct pattern of interaction. A single prompt, such as “You are an expert data analyst combining statistical rigor with deep domain knowledge…”, might seem harmless. However, when variations of this prompt arrive tens of thousands of times across hundreds of coordinated accounts, all targeting the same narrow capability, the illicit pattern becomes undeniably clear. Massive volume concentrated in specific areas, highly repetitive structures, and content directly valuable for training AI models are the defining hallmarks of a distillation attack. The focus is always on harvesting what is most valuable.
Case Studies: Labs Identified by Anthropic
Anthropic’s investigation attributed each campaign to a specific lab with high confidence. This attribution relied on IP address correlation, detailed request metadata, infrastructure indicators, and corroboration from industry partners.
DeepSeek’s Reasoning and Redaction Tactics
DeepSeek’s operation involved over 150,000 exchanges. Their primary targets included Claude’s reasoning capabilities across diverse tasks and rubric-based grading. They effectively used Claude as a reward model for reinforcement learning. Notably, DeepSeek also focused on creating censorship-safe alternatives to politically sensitive queries. This likely aimed to train their own models to steer conversations away from restricted topics, such as those about dissidents, party leaders, or authoritarianism.
DeepSeek generated synchronized traffic across accounts, using identical patterns and shared payment methods. Coordinated timing suggested a “load balancing” strategy to increase throughput and evade detection. A significant technique involved prompts asking Claude to articulate its internal reasoning step-by-step. This effectively generated “chain-of-thought” training data at scale for DeepSeek’s models.
Moonshot AI’s Agentic and Coding Focus
Moonshot AI (known for its Kimi models) executed a much larger operation, involving over 3.4 million exchanges. This campaign employed hundreds of fraudulent accounts across multiple access pathways. The varied account types made the coordinated operation harder to detect. Moonshot targeted Claude’s agentic reasoning, tool use, coding, data analysis, and even computer vision capabilities.
Initially, attribution relied on request metadata matching public profiles of senior Moonshot staff. In a later phase, Moonshot shifted to a more targeted approach. This involved attempting to extract and reconstruct Claude’s intricate reasoning traces, seeking deeper insight into the model’s cognitive processes.
MiniMax’s Rapid Adaptation and Coding Theft
MiniMax conducted the largest observed campaign, generating over 13 million exchanges. Their operation specifically targeted Claude’s agentic coding and orchestration capabilities. Anthropic attributed this campaign through request metadata and infrastructure indicators, confirming timings against MiniMax’s public product roadmap.
Remarkably, this campaign was detected while still active, prior to MiniMax’s model release. This provided Anthropic with unprecedented visibility into the full lifecycle of a distillation attack. When Anthropic released a new model during MiniMax’s active campaign, MiniMax swiftly pivoted within 24 hours. They redirected nearly half their traffic to immediately capture capabilities from the updated system, demonstrating extreme responsiveness and intent.
Anthropic’s Proactive Defense Strategy
Anthropic continues to invest heavily in robust defenses. These measures aim to make such AI distillation attacks harder to execute and easier to identify. Their multi-faceted approach includes:
Advanced Detection Systems: Anthropic has developed sophisticated classifiers and behavioral fingerprinting systems. These tools are designed to identify distinct distillation attack patterns in API traffic. This includes detecting “chain-of-thought” elicitation, a common technique for constructing reasoning training data. They also use detection tools for identifying coordinated activity across large numbers of accounts.
Intelligence Sharing: Crucially, Anthropic is sharing technical indicators and insights with other AI labs, major cloud providers, and relevant authorities. This collaborative approach provides a more holistic and accurate picture of the evolving distillation landscape.
Strengthened Access Controls: Verification processes for educational accounts, security research programs, and startup organizations have been significantly strengthened. These pathways were most commonly exploited for setting up fraudulent accounts.
Innovative Countermeasures: Anthropic is actively developing product, API, and model-level safeguards. These are specifically designed to reduce the efficacy of model outputs for illicit distillation, without degrading the experience for legitimate customers.
A Collective Call to Action
No single company can effectively solve the pervasive challenge of AI distillation attacks alone. The scale and sophistication of these campaigns necessitate a rapid, coordinated response. Industry players, cloud providers, policymakers, and the broader global AI community must collaborate. Sharing evidence and developing unified strategies are paramount to safeguarding frontier AI models. Without this collective effort, the risks of unchecked proliferation of dangerous AI capabilities will only continue to grow.
Frequently Asked Questions
What are AI distillation attacks and why are they a significant threat?
AI distillation attacks involve training a less capable AI model on the outputs of a stronger, more advanced one to illicitly extract its capabilities. They are a significant threat because they enable competitors to acquire powerful AI capabilities quickly and cheaply, bypassing independent development costs. More critically, these attacks circumvent crucial safeguards built into legitimate models, posing national security risks by allowing dangerous AI functionalities (e.g., for bioweapons or cyberattacks) to proliferate without protection. They also undermine international export controls designed to maintain competitive advantages and control AI proliferation.
How do malicious actors typically carry out AI distillation attacks to extract capabilities?
Malicious actors execute AI distillation attacks using sophisticated methods. They often create “hydra cluster” architectures, which are sprawling networks of thousands of fraudulent accounts, often using commercial proxy services to bypass regional access restrictions (e.g., from China to US models). These networks distribute traffic across APIs and cloud platforms, mixing illicit activity with legitimate requests to evade detection. Once access is secured, they generate massive volumes of carefully crafted prompts, sometimes tens of thousands of times, to elicit specific, high-value capabilities like reasoning, coding, or tool use, often generating chain-of-thought data for training.
What coordinated actions are necessary to effectively combat large-scale AI distillation attacks?
Combating large-scale AI distillation attacks requires a multi-pronged, coordinated effort. Key actions include:
- Enhanced Detection: AI labs must continually develop advanced behavioral fingerprinting and classification systems to identify attack patterns.
- Intelligence Sharing: AI labs, cloud providers, and government authorities need to share technical indicators and insights to gain a holistic view of the threat landscape.
- Strengthened Access Controls: Improving verification for API access and platform accounts helps prevent the creation of fraudulent profiles.
- Proactive Countermeasures: Developing product, API, and model-level safeguards can reduce the utility of model outputs for illicit distillation.
- Policy and Enforcement: Policymakers must reinforce export controls and consider new legal frameworks to address digital intellectual property theft in AI. This collective action is essential to protect the integrity and security of frontier AI development.
Conclusion
The revelations about industrial-scale AI distillation attacks by DeepSeek, Moonshot, and MiniMax serve as a stark warning. The illicit extraction of AI capabilities poses multifaceted threats, from intellectual property theft to grave national security concerns. Anthropic’s findings underscore the urgent need for a unified defense strategy across the AI industry, governmental bodies, and cloud infrastructure providers. By enhancing detection, sharing intelligence, bolstering access controls, and developing innovative countermeasures, the global AI community can collectively safeguard the integrity and responsible development of frontier AI models. Protecting these advanced systems from illicit exploitation is not just a business imperative; it is a critical step towards securing a safe and trustworthy AI future.