Table of Contents Show
In a significant disclosure that underscores the escalating stakes of global AI competition, Anthropic has revealed three coordinated, industrial scale campaigns to illicitly extract the capabilities of its Claude AI model. The operations, attributed with high confidence to Chinese AI laboratories DeepSeek, Moonshot AI, and MiniMax, collectively generated over 16 million exchanges through approximately 24,000 fraudulent accounts in direct violation of Anthropic’s terms of service and regional access restrictions.
Understanding Distillation and Its Illicit Application
Distillation refers to training a less capable model on the outputs of a more advanced one – a legitimate and widely accepted industry practice. However, when executed without authorization, it becomes a mechanism for capability theft, allowing competitors to acquire sophisticated AI capabilities at a fraction of the time and cost of independent development. What distinguishes an attack from legitimate use is not the technique itself, but the intent, scale, and coordinated pattern behind it.
The Three Campaigns: Scale, Tactics, and Attribution
DeepSeek generated over 150,000 exchanges targeting Claude’s reasoning capabilities and prompting the model to produce censorship safe alternatives to politically sensitive queries. Moonshot AI amassed over 3.4 million exchanges focused on agentic reasoning, tool use, coding, and computer vision with Anthropic tracing the campaign to specific senior staff through request metadata. MiniMax conducted the most extensive operation, generating over 13 million exchanges targeting agentic coding and tool orchestration. Anthropic detected this campaign while still active, observing MiniMax pivot within 24 hours of a new Claude model release to target the latest system.
National Security and Export Control Implications
Models built through illicit distillation are unlikely to retain safety guardrails protecting against bioweapons development, malicious cyber operations, and mass surveillance. These stripped capabilities can be integrated into military and intelligence infrastructure or multiplied further if open sourced. Critically, these attacks undermine export controls – apparent foreign advancements achieved through distillation have incorrectly suggested that chip restrictions are ineffective. Anthropic argues the opposite: executing distillation at scale requires advanced chips, reinforcing the rationale for maintaining those controls.
Anthropic’s Response and the Path Forward
Anthropic has deployed behavioral fingerprinting classifiers, chain of thought elicitation detection, intelligence sharing with laboratories and authorities, and strengthened access verification across commonly exploited account pathways. Model and API level countermeasures are also in development. The company has been clear that no single organization can address this challenge alone – rapid, coordinated action across the AI industry, cloud providers, and policymakers is essential. The window to act, Anthropic warns, is narrow.
Also Read: Portkey Secures $15 Million in Series A Funding Led by Elevation Capital to Scale LLMOps Platform