Anthropic Exposes Industrial-Scale AI Distillation Attacks by DeepSeek, Moonshot AI, and MiniMax
Original: Anthropic Exposes Industrial-Scale AI Model Distillation Attacks by DeepSeek, Moonshot AI, and MiniMax View original →
Industrial-Scale Distillation Attacks Discovered
On February 24, 2026, Anthropic publicly disclosed that major Chinese AI companies had been conducting large-scale distillation attacks against its Claude models. DeepSeek, Moonshot AI, and MiniMax were identified as the perpetrators.
Scale and Method
The attack involved:
- Creation of over 24,000 fraudulent accounts
- Generation of more than 16 million exchanges with Claude
- Using that conversation data to train and improve their own competing AI models
Why Illicit Distillation Is Dangerous
Anthropic distinguishes between legitimate and illicit distillation. While AI labs legitimately use distillation to create smaller, cheaper models for their customers, foreign labs that illicitly distill American models can remove safety guardrails and feed extracted capabilities into their military, intelligence, and surveillance systems.
Call for Coordinated Action
Anthropic warned that these attacks are growing in both intensity and sophistication, calling for rapid, coordinated action from industry players, policymakers, and the broader AI community to address the threat.
Full details are available in Anthropic's official report: Detecting and Preventing Distillation Attacks.
Related Articles
Anthropic revealed that Chinese AI labs DeepSeek, Moonshot AI, and MiniMax created over 24,000 fraudulent accounts and generated 16+ million Claude exchanges to extract its capabilities and improve their own competing models.
Anthropic revealed that Chinese AI labs DeepSeek, Moonshot AI, and MiniMax created over 24,000 fraudulent accounts and generated 16+ million Claude exchanges to extract its capabilities and improve their own competing models.
Anthropic published a March 6, 2026 case study showing how Claude Opus 4.6 authored a working test exploit for Firefox vulnerability CVE-2026-2796. The company presents the result as an early warning about advancing model cyber capabilities, not as proof of reliable real-world offensive automation.
Comments (0)
No comments yet. Be the first to comment!