{"id":184699,"date":"2025-01-29T04:19:28","date_gmt":"2025-01-29T04:19:28","guid":{"rendered":"https:\/\/globetimeline.com\/ar\/tech\/rewrite-this-title-in-arabic-openai-says-it-has-evidence-chinas-deepseek-used-its-model-to-train-competitor\/"},"modified":"2025-01-29T04:19:28","modified_gmt":"2025-01-29T04:19:28","slug":"rewrite-this-title-in-arabic-openai-says-it-has-evidence-chinas-deepseek-used-its-model-to-train-competitor","status":"publish","type":"post","link":"https:\/\/globetimeline.com\/ar\/tech\/rewrite-this-title-in-arabic-openai-says-it-has-evidence-chinas-deepseek-used-its-model-to-train-competitor\/","title":{"rendered":"rewrite this title in Arabic OpenAI says it has evidence China\u2019s DeepSeek used its model to train competitor"},"content":{"rendered":"<p>Summarize this content to 2000 words in 6 paragraphs in Arabic Stay informed with free updatesSimply sign up to the Artificial intelligence myFT Digest &#8212; delivered directly to your inbox.OpenAI says it has found evidence that Chinese artificial intelligence start-up DeepSeek used the US company\u2019s proprietary models to train its own open-source competitor, as concerns grow over a potential breach of intellectual property.The San-Francisco-based ChatGPT maker told the Financial Times it had seen some evidence of \u201cdistillation\u201d, a technique used by developers to obtain better performance on smaller models by using outputs from larger, more capable models. This allows them to achieve similar results on specific tasks at a much lower cost.OpenAI declined to comment further on details of its evidence. Its terms of service state users cannot \u201ccopy\u201d any of its services or \u201cuse output to develop models that compete with OpenAI\u201d.DeepSeek\u2019s release of its R1 reasoning model has surprised markets, as well as investors and technology companies in Silicon Valley, due to its impressive performance at cognitive tasks. Its models have attained high rankings and comparable results to leading US models.One person close to OpenAI said that distillation was a common practice in the industry and highlighted that the company offers developers a way to do this using its own platform, but said: \u201cThe issue is when you are doing it to create your own model for your own purposes.\u201dDeepSeek did not immediately respond to a request for comment.Earlier, President Donald Trump\u2019s AI and crypto tsar David Sacks said \u201cit is possible\u201d that IP theft had occurred.\u201cThere\u2019s a technique in AI called distillation\u2009.\u2009.\u2009.\u2009when one model learns from another model [and] kind of sucks the knowledge out of the parent model,\u201d Sacks told Fox News on Tuesday.\u201cAnd there\u2019s substantial evidence that what DeepSeek did here is they distilled the knowledge out of OpenAI models, and I don\u2019t think OpenAI is very happy about this,\u201d Sacks added, although he did not provide evidence.DeepSeek said it used just 2,048 Nvidia H800 graphics cards and $5.6mn to train its V3 model with 671bn parameters, a fraction of what OpenAI and Google spent to train comparably sized models. Some experts pointed out how the model generated responses that indicated it had been trained on outputs from OpenAI\u2019s GPT-4, which would violate its terms of service.\u00a0Industry insiders say that, in reality, it is common practice for AI labs, both in China and the US, to use outputs from leading companies such as OpenAI.Industry leaders such as OpenAI have invested in hiring people to teach their models how to produce responses that sound more human. This is expensive and labour-intensive, and industry insiders say it is common for smaller players to piggyback off their work. \u201cIt is a very common practice for start-ups and academics to use outputs from human-aligned commercial LLMs, like ChatGPT, to train another model,\u201d said Ritwik Gupta, a PhD candidate in AI at the University of California, Berkeley.\u201cThat means you get this human feedback step for free. It is not surprising to me that DeepSeek supposedly would be doing the same. If they were, stopping this practice precisely may be difficult,\u201d he added. The practice also points to an emerging financial conundrum for frontier companies that are doing cutting-edge research in AI on how they defend their technical edge when other groups can piggyback off their models.\u00a0Chinese companies have quickly absorbed lessons from their US counterparts while innovating approaches to maximise their limited number of chips, making it cheaper to train and run the models.\u00a0\u201cWe know [China]-based companies \u2014 and others \u2014 are constantly trying to distil the models of leading US AI companies,\u201d OpenAI added in a statement. \u201cWe engage in countermeasures to protect our IP, including a careful process for which frontier capabilities to include in released models, and believe as we go forward that it is critically important that we are working closely with the US government to best protect the most capable models from efforts by adversaries and competitors to take US technology.\u201dOpenAI is currently battling allegations of its own copyright infringement from newspapers and content creators, including lawsuits from The New York Times and prominent authors, who accuse the company of training their models on their articles and books without permission.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Summarize this content to 2000 words in 6 paragraphs in Arabic Stay informed with free updatesSimply sign up to the Artificial intelligence myFT Digest &#8212; delivered directly to your inbox.OpenAI says it has found evidence that Chinese artificial intelligence start-up DeepSeek used the US company\u2019s proprietary models to train its own open-source competitor, as concerns<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[63],"tags":[],"class_list":{"0":"post-184699","1":"post","2":"type-post","3":"status-publish","4":"format-standard","6":"category-tech"},"_links":{"self":[{"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/posts\/184699","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/comments?post=184699"}],"version-history":[{"count":0,"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/posts\/184699\/revisions"}],"wp:attachment":[{"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/media?parent=184699"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/categories?post=184699"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/globetimeline.com\/ar\/wp-json\/wp\/v2\/tags?post=184699"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}