
The massive attacks were used to improve their own models with agentic reasoning, tool use, and coding capabilities, violating Anthropic’s Terms of Service and creating a national security risk, they say.
Distillation works by sending millions of prompts to an AI to incorporate its techniques and capabilities into their own models, drastically reducing training time and costs.
They also circumvent Anthropic’s protections for use in developing bioweapons and malicious cyber activities, Anthropic says. Once these models are open sourced, this becomes available to anyone.
OpenAI said the same just last week, accusing DeepSeek of distillation.
— These campaigns are growing in intensity and sophistication. The window to act is narrow, and the threat extends beyond any single company or region, Anthropic writes.
Read more: Anthropic’s announcement, writeups on Reuters, TechCrunch, Engadget and The Verge.










