Anthropic accuses DeepSeek, Moonshot and MiniMax of unauthorised distillation of the Claude model
Anthropic this week claimed that Chinese competitors were stealing its AI technology by launching so-called distillation attacks on its models.
The accusation is that of distillation, but spirits have nothing to do with it. Making the accusation is Anthropic, the American company that developed the Claude artificial intelligence model. And ending up in the crosshairs are three Chinese competitors, DeepSeek, Moonshot and MiniMax, who allegedly used their LLM to create 24,000 accounts and engage in more than 16 million conversations with the US AI.
The aim is precisely to distil knowledge, a kind of training whereby a model converses with a more powerful one to learn how to generate better and better answers. This is, Anthropic explains, a normal practice in the industry, but one that is usually carried out internally: one distils one's own model in order to create a smaller, cheaper version to put on the market. The problem arises precisely when the coach, i.e. the LLM questioned, is that of another company.
Claude's developers, in a post published on the company blog, announced that they had encountered more than 150 thousand interactions between their model and that of DeepSeek. Conversations that would have sought to make the latter understand Claude's reasoning capabilities, asking him to articulate and explain them step by step. But there were also requests to provide alternative answers, able to go beyond the meshes of censorship, on politically sensitive topics such as dissidents, party leaders, authoritarianism.
On the other hand, more than 3.4 million conversations were orchestrated by Moonshot AI, this time targeting the use of agents, coding and data analysis skills, and computer vision. Finally, most of the training sessions, over 13 million, involved MiniMax. Again, interaction focused on agent development.
The issue, Claude's creators emphasise, is not only about using other people's models to train their own. Models distilled in this way, they say, 'lack the necessary safety measures'. What does that mean? 'Anthropic and other American companies build systems that prevent artificial intelligence from being used, for example, to develop biological weapons or to manage cyber attacks'. Prohibitions, however, are lost through distillation. At stake, in other words, is a national security issue.

