«

Anthropic指控三家中国AI实验室滥用Claude模型优化自身系统。

qimuai 发布于 阅读:1 一手编译


Anthropic指控三家中国AI实验室滥用Claude模型优化自身系统。

内容来源:https://www.engadget.com/ai/anthropic-accuses-three-chinese-ai-labs-of-abusing-claude-to-improve-their-own-models-205210613.html?src=rss

内容总结:

美国人工智能公司Anthropic近日公开指控三家中国AI企业通过"蒸馏攻击"不当利用其聊天机器人Claude提升自身模型能力。据该公司披露,深度求索、月之暗面与Minimax三家公司通过约2.4万个虚假账户进行了超1600万次违规交互,以"工业级规模"提取Claude的核心能力。

Anthropic在官网声明中指出,这种系统性行为不仅侵犯其服务条款,更可能绕过AI安全护栏。该公司通过IP地址关联、元数据分析和行业协作,以"高度置信"级别锁定相关企业。此次事件距离OpenAI去年指控竞争对手类似行为仅一年之隔。

面对行业争议,Anthropic表示将升级系统防御机制以遏制此类行为。值得注意的是,该公司自身也正面临音乐出版商的集体诉讼,被指控使用非法复制歌曲训练Claude模型。这一系列事件折射出全球AI行业在技术竞争与知识产权保护之间亟待建立规范框架的现实挑战。

中文翻译:

人工智能公司Anthropic指控三家中国AI实验室滥用其Claude模型来改进自身产品,此举紧随OpenAI近期关于其他公司"蒸馏"其模型的声明。Anthropic在官网公开指控深度求索、月之暗面及MiniMax三家公司开展"工业级规模的非法能力提取行动",通过约2.4万个虚假账户与Claude进行超1600万次对话,旨在提升其自身模型性能。

AI领域的"蒸馏"通常指较弱模型借鉴强模型输出进行自我训练。Anthropic指出,虽然蒸馏技术本身具有合法性,但上述行为已构成恶意攻击。这些公司通过系统性操作将Claude作为技术捷径,不仅可能绕过安全防护机制,更直接规避了模型开发过程中的必要规范。

Anthropic表示,通过IP地址关联、元数据分析及基础设施特征比对,并结合行业伙伴提供的佐证,已"高度确信"锁定这三家实施蒸馏攻击的企业。去年初,OpenAI也曾以类似理由封禁可疑账户。作为应对措施,Anthropic宣布将升级系统以增强反蒸馏能力。

值得注意的是,Anthropic自身正面临音乐出版商的集体诉讼,被指控使用非法复制歌曲训练Claude模型。此次公开指控正值该AI公司处于法律纠纷之际。

英文来源:

Anthropic accuses three Chinese AI labs of abusing Claude to improve their own models
Anthropic's latest post follows OpenAI's recent claim of other AI companies "distilling" its AI models.
Anthropic is issuing a call to action against AI "distillation attacks," after accusing three AI companies of misusing its Claude chatbot. On its website, Anthropic claimed that DeepSeek, Moonshot and MiniMax have been conducting "industrial-scale campaigns…to illicitly extract Claude’s capabilities to improve their own models."
Distillation in the AI world refers to when less capable models lean on the responses of more powerful ones to train themselves. While distillation isn't a bad thing across the board, Anthropic said that these types of attacks can be used in a more nefarious way. According to Anthropic, these three Chinese AI firms were responsible for more than "16 million exchanges with Claude through approximately 24,000 fraudulent accounts." From Anthropic's perspective, these competing companies were using Claude as a shortcut to develop more advanced AI models, which could also lead to circumventing certain safeguards.
Anthropic said in its post that it was able to link each of these distilling attack campaigns to the specific companies with "high confidence" thanks to IP address correlation, metadata requests and infrastructure indicators, along with corroborating with others in the AI industry who have noticed similar behaviors.
Early last year, OpenAI made similar claims of rival firms distilling its models and banned suspected accounts in response. As for Anthropic, the company behind Claude said it would upgrade its system to make distillation attacks harder to do and easier to identify. While Anthropic is pointing fingers at these other firms, it's also facing a lawsuit from music publishers who accused the AI company of using illegal copies of songs to train its Claude chatbot.

Engadget

文章目录


    扫描二维码,在手机上阅读