Anthropic 大牌坊发力了:
We’ve identified industrial-scale distillation attacks on our models by DeepSeek, Moonshot AI, and MiniMax.
These labs created over 24,000 fraudulent accounts and generated over 16 million exchanges with Claude, extracting its capabilities to train and improve their own models.
Distillation can be legitimate: AI labs use it to create smaller, cheaper models for their customers.
But foreign labs that illicitly distill American models can remove safeguards, feeding model capabilities into their own military, intelligence, and surveillance systems.

我们发现 DeepSeek、Moonshot AI 和 MiniMax 对我们的模型发起了工业规模的蒸馏攻击。
这些实验室创建了超过 24,000 个欺诈账户,并与 Claude 进行了超过 1600 万次交易,提取了 Claude 的功能来训练和改进他们自己的模型。
蒸馏是合理的:人工智能实验室利用蒸馏为客户创建更小、更便宜的模型。
但是,非法提炼美国模型的外国实验室可以移除安全措施,将模型能力输送到他们自己的军事、情报和监视系统中。

https://x.com/AnthropicAI/status/2025997928242811253
https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks
 
 
Back to Top