Anthropic Says Chinese AI Firms Used 16 Million Claude Queries to Copy Model
嗯,用户让我用中文帮他总结一下这篇文章的内容,控制在一百个字以内,而且不需要以“文章内容总结”或者“这篇文章”这样的开头,直接写描述即可。首先,我需要理解这篇文章讲的是什么。 文章主要讲的是Anthropic公司发现有三家中国的人工智能公司——DeepSeek、Moonshot AI和MiniMax,通过非法手段从Claude模型中提取能力来提升自己的模型。他们使用了大约24,000个欺诈账户,进行了超过1600万次的交互,违反了服务条款和区域访问限制。 接下来,文章解释了什么是蒸馏技术,即用较弱的模型训练较强的AI系统的输出。虽然蒸馏是合法的,但竞争对手非法使用它来获取其他公司的能力是违法的。这不仅涉及国家安全风险,还可能导致这些模型被用于恶意活动。 此外,文章提到这些攻击利用了商业代理服务来大规模访问Claude模型,并且这些服务使用复杂的网络结构来避免被检测到。Anthropic公司已经采取了一些措施来应对这些威胁。 用户的需求是希望得到一个简洁的总结,控制在100字以内。所以我要抓住关键点:三家公司非法提取Claude的能力,使用欺诈账户和代理服务进行蒸馏攻击,涉及国家安全风险,并且Anthropic采取了应对措施。 现在我需要把这些信息浓缩成一句话或几句话,确保不超过100字,并且直接描述内容,不加开头语。 可能的结构是:Anthropic指控三家公司非法提取其AI模型的能力用于训练自己的模型,并详细描述了攻击手段及其带来的安全风险。 这样既涵盖了主要事件、手段和后果,又符合字数限制。 </think> Anthropic指控三家人造智能公司非法提取其AI模型的能力用于训练自身模型,并详细描述了通过欺诈账户和代理服务进行的大规模蒸馏攻击及其带来的国家安全风险。 2026-2-24 06:4:0 Author: thehackernews.com(查看原文) 阅读量:8 收藏

Artificial Intelligence / Anthropic

Anthropic on Monday said it identified "industrial-scale campaigns" mounted by three artificial intelligence (AI) companies, DeepSeek, Moonshot AI, and MiniMax, to illegally extract Claude's capabilities to improve their own models.

The distillation attacks generated over 16 million exchanges with its large language model (LLM) through about 24,000 fraudulent accounts in violation of its terms of service and regional access restrictions. All three companies are based in China, where the use of its services is prohibited due to "legal, regulatory, and security risks."

Distillation refers to a technique where a less capable model is trained on the outputs generated by a stronger AI system. While distillation is a legitimate way for companies to produce smaller, cheaper versions of their own frontier models, it's illegal for competitors to leverage it to acquire such capabilities from other AI companies at a fraction of the time and cost that would take them if they were to develop them on their own.

"Illicitly distilled models lack necessary safeguards, creating significant national security risks," Anthropic said. "Models built through illicit distillation are unlikely to retain those safeguards, meaning that dangerous capabilities can proliferate with many protections stripped out entirely."

Foreign AI companies that distill American models can weaponize these unprotected capabilities to facilitate malicious activities, cyber-related or otherwise, thereby serving as a foundation for military, intelligence, and surveillance systems that authoritarian governments can deploy for offensive cyber operations, disinformation campaigns, and mass surveillance.

The campaigns detailed by AI upstart entail the use of fraudulent accounts and commercial proxy services to access Claude at scale while avoiding detection. Anthropic said it was able to attribute each campaign to a specific AI lab based on request metadata, IP address correlation, request metadata, and infrastructure indicators.

The details of the three distillation attacks are below -

  • DeepSeek, which targeted Claude's reasoning capabilities, rubric-based grading tasks, and sought its help in generating censorship-safe alternatives to politically sensitive queries like questions about dissidents, party leaders, or authoritarianism across over 150,000 exchanges.
  • Moonshot AI, which targeted Claude's agentic reasoning and tool use, coding capabilities, computer-use agent development, and computer vision across over 3.4 million exchanges.
  • MiniMax, which targeted Claude's agentic coding and tool use capabilities across over 13 million exchanges.

"The volume, structure, and focus of the prompts were distinct from normal usage patterns, reflecting deliberate capability extraction rather than legitimate use," Anthropic added. "Each campaign targeted Claude's most differentiated capabilities: agentic reasoning, tool use, and coding."

The company also pointed out that the attacks relied on commercial proxy services that resell access to Claude and other frontier AI models at scale. These services are powered by "hydra cluster" architectures that contain massive networks of fraudulent accounts to distribute traffic across their API.

The access is then used to generate large volumes of carefully crafted prompts that are designed to extract specific capabilities from the model for the purpose of training their own models by harvesting the high-quality responses. 

"The breadth of these networks means that there are no single points of failure," Anthropic said. "When one account is banned, a new one takes its place. In one case, a single proxy network managed more than 20,000 fraudulent accounts simultaneously, mixing distillation traffic with unrelated customer requests to make detection harder."

To counter the threat, Anthropic said it has built several classifiers and behavioral fingerprinting systems to identify suspicious distillation attack patterns in API traffic, strengthened verification for educational accounts, security research programs, and startup organizations, and implemented enhanced safeguards to reduce the efficacy of model outputs for illicit distillation.

The disclosure comes weeks after Google Threat Intelligence Group (GTIG) disclosed it identified and disrupted distillation and model extraction attacks aimed at Gemini's reasoning capabilities through more than 100,000 prompts.

"Model extraction and distillation attacks do not typically represent a risk to average users, as they do not threaten the confidentiality, availability, or integrity of AI services," Google said earlier this month. "Instead, the risk is concentrated among model developers and service providers."

Found this article interesting? Follow us on Google News, Twitter and LinkedIn to read more exclusive content we post.


文章来源: https://thehackernews.com/2026/02/anthropic-says-chinese-ai-firms-used-16.html
如有侵权请联系:admin#unsafe.sh