Chinese AI labs created 24K accounts and «distilled» 16 million messages from Claude

Chinese attacks risk bypassing the safeguards Anthropic builds into its models. (Picture: Anthropic)
Anthropic claims to have discovered industrial scale extraction of Claude data from DeepSeek, Moonshot AI and MiniMax.

The massive attacks were used to improve their own models with agentic reasoning, tool use, and coding capabilities, violating Anthropic’s Terms of Service and creating a national security risk, they say.

Distillation works by sending millions of prompts to an AI to incorporate its techniques and capabilities into their own models, drastically reducing training time and costs.

They also circumvent Anthropic’s protections for use in developing bioweapons and malicious cyber activities, Anthropic says. Once these models are open sourced, this becomes available to anyone.

OpenAI said the same just last week, accusing DeepSeek of distillation.

— These campaigns are growing in intensity and sophistication. The window to act is narrow, and the threat extends beyond any single company or region, Anthropic writes.

Read more: Anthropic’s announcement, writeups on Reuters, TechCrunch, Engadget and The Verge.