Three Chinese AI Labs Ran Industrial-Scale Distillation Attacks on Claude, Extracting 16 Million+ Exchanges
Anthropic reveals that DeepSeek, Moonshot AI, and MiniMax used approximately 24,000 fraudulent accounts to conduct over 16 million exchanges with Claude, systematically extracting agentic reasoning, tool use, and coding capabilities for their own model training.
Anthropic disclosed on February 23 that three Chinese AI companies — DeepSeek, Moonshot AI, and MiniMax — conducted industrial-scale distillation campaigns against Claude. Using approximately 24,000 fraudulent accounts, the three firms collectively executed more than 16 million exchanges designed to extract Claude’s agentic reasoning, tool use, and coding capabilities for use in training their own models.
The Scale of Each Operation
MiniMax ran the largest campaign by far, accounting for over 13 million exchanges. The operation focused on extracting agentic coding and tool use capabilities, and was detected ahead of a new MiniMax model release. Anthropic observed a telling behavioral pattern: whenever a new Claude model was released, MiniMax redirected roughly half its traffic to the new model within 24 hours — a clear indicator of systematic capability harvesting.
Moonshot AI, the company behind the Kimi model series, conducted more than 3.4 million exchanges. Its operation targeted a broader range of capabilities: agentic reasoning, coding, computer-use agent development, and computer vision. In a particularly notable finding, Anthropic’s analysis of request metadata matched profiles of senior Moonshot AI staff, directly linking the company’s employees to the distillation infrastructure.
DeepSeek ran the smallest operation at over 150,000 exchanges, but its approach was methodologically distinct. Rather than simply extracting outputs, DeepSeek focused on reasoning capabilities and rubric-based scoring. It prompted Claude to “verbalize the internal reasoning of completed answers step by step” — effectively generating chain-of-thought training data at scale. DeepSeek also used Claude to generate censorship-safe alternative responses to politically sensitive queries.
How They Did It
All three companies used fraudulent accounts combined with commercial proxy services to access Claude at scale. Anthropic’s investigation found that a single proxy network was managing more than 20,000 fraudulent accounts simultaneously — suggesting a well-resourced, professionally operated infrastructure behind the distillation campaigns.
How Anthropic Caught Them
Detection relied on a combination of IP address correlation, request metadata analysis, infrastructure indicators, and intelligence from industry partners. The multi-signal approach allowed Anthropic to not only identify the distillation activity but also attribute it to specific organizations.
In response, Anthropic has deployed classifier systems and behavioral fingerprinting tools designed to detect and block distillation attempts in real time. The company indicated these systems are now running in production.
OpenAI Reports Similar Attacks
Anthropic isn’t alone. OpenAI has claimed that DeepSeek employed similar distillation techniques against ChatGPT, suggesting this is not an isolated incident but a systematic industry practice.
Model distillation — harvesting a model’s outputs at scale to train a competing model — is emerging as one of the most significant intellectual property challenges in the AI industry. Traditional software IP protections don’t map cleanly onto the problem: the “theft” happens through legitimate API calls, making detection and enforcement fundamentally different from conventional code piracy.
For AI providers, the implications are clear. Distillation defense is no longer optional — it’s a core infrastructure requirement. And as models become more capable, the incentive for competitors to extract that capability through distillation only grows.
Official announcement: https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks
Related Articles
Anthropic Launches Claude Code Security, an AI-Powered Vulnerability Scanner
Anthropic unveils Claude Code Security, a new feature built into Claude Code on the web that reasons about code like a human security researcher. Available as a limited research preview for Enterprise and Team customers.
Claude Code v2.1.47 Released: 40+ Bug Fixes, Windows Overhaul, Memory Improvements, and Faster Startup
Anthropic ships Claude Code v2.1.47 with over 40 bug fixes targeting Windows rendering bugs, long-session memory leaks, a ~500ms startup improvement, plan mode preservation after compaction, and a revamped background agent kill shortcut.
Claude Code v2.1.49 Released: Background Agent Kill Fix, Startup Speedups, and Yoga WASM Memory Leaks Squashed
Anthropic ships Claude Code v2.1.49 with roughly 20 changes, including a fix for Ctrl+C and ESC being silently ignored during background agent execution, multiple startup performance improvements via MCP batching and auth-failure caching, an end to Yoga WASM memory growth in long sessions, file editing in simple mode, and new SDK capability fields.
Popular Articles
Claude Code v2.1.93 Released - Deferred Permission Decisions, Flicker-Free Rendering, and More
Anthropic releases Claude Code v2.1.93 with deferred permission decisions for PreToolUse hooks, flicker-free rendering option, PermissionDenied hook, and named subagent typeahead support.
Claude Code v2.1.92 Released - forceRemoteSettingsRefresh, Bedrock Setup Wizard, and More
Anthropic releases Claude Code v2.1.92 with forceRemoteSettingsRefresh policy setting, AWS Bedrock setup wizard, /cost command improvements, and numerous bug fixes.
Claude Code v2.1.84 Release - PowerShell Tool Preview and Environment Configuration Enhancements
Claude Code v2.1.84 introduces PowerShell tool for Windows, new environment variable overrides for model selection, idle session handling improvements, and various stability fixes.
Latest Articles
Claude Code v2.1.93 Released - Deferred Permission Decisions, Flicker-Free Rendering, and More
Anthropic releases Claude Code v2.1.93 with deferred permission decisions for PreToolUse hooks, flicker-free rendering option, PermissionDenied hook, and named subagent typeahead support.
Claude Code v2.1.92 Released - forceRemoteSettingsRefresh, Bedrock Setup Wizard, and More
Anthropic releases Claude Code v2.1.92 with forceRemoteSettingsRefresh policy setting, AWS Bedrock setup wizard, /cost command improvements, and numerous bug fixes.
Claude Code v2.1.91 Released - MCP Tool Result Persistence and Improved Edit Tool
Claude Code v2.1.91 introduces MCP tool result persistence override, improved shell execution controls, and enhanced Edit tool efficiency.