{"id":3281,"date":"2025-01-27T17:28:00","date_gmt":"2025-01-27T15:28:00","guid":{"rendered":"https:\/\/leonardonetwork.eu\/?p=3281"},"modified":"2025-01-27T17:28:01","modified_gmt":"2025-01-27T15:28:01","slug":"la-nascita-di-una-stella-deepseek-lai-cinese","status":"publish","type":"post","link":"https:\/\/leonardonetwork.eu\/en\/the-birth-of-a-star-deepseek-lai-chinese\/","title":{"rendered":"The Birth of a Star, DeepSeek, the Chinese AI"},"content":{"rendered":"<p>Liang Wenfeng, once the head of a quantitative hedge fund in China, transformed his career by betting on artificial intelligence research. He invested considerable resources, acquiring 10,000 Nvidia chips and assembling a team of brilliant young researchers. This bold project led, two years later, to the launch of DeepSeek, an initiative that quickly captured the attention of the global tech industry.<\/p>\n\n\n\n<p>On January 20, the DeepSeek lab launched an open-source model that has attracted the attention of Silicon Valley. According to a company document, DeepSeek-R1 outperforms leading models like OpenAI in mathematics and reasoning, posing a serious challenge to Western AI giants.<\/p>\n\n\n\n<p>DeepSeek&#039;s success illustrates an unexpected effect of US technological restrictions. With limited access to advanced chips, many Chinese companies have focused on applications rather than basic models. However, DeepSeek has taken a different approach: improving the architecture of its AI models to more efficiently utilize limited resources.<\/p>\n\n\n\n<p>\u201cDeepSeek excels at software-driven resource optimization, an approach that fosters collaborative innovation,\u201d says Marina Zhang, associate professor at the University of Technology Sydney.<\/p>\n\n\n\n<p><strong>The Origin of DeepSeek<\/strong><\/p>\n\n\n\n<p>DeepSeek began as Fire-Flyer, a research arm of High-Flyer, a successful Chinese quantitative hedge fund. Founded in 2015, High-Flyer accumulated GPUs for financial analytics until Liang decided in 2023 to redirect those resources to building advanced AI models.<\/p>\n\n\n\n<p>DeepSeek has adopted a unique recruitment strategy, targeting young graduates from prestigious universities like Beijing and Tsinghua University, eager to demonstrate their worth. &quot;We chose researchers with no industry experience but with an innovative mindset,&quot; said Liang.<\/p>\n\n\n\n<p><strong>Innovation under Pressure<\/strong><\/p>\n\n\n\n<p>US restrictions on advanced chips have pushed DeepSeek to develop more efficient training methods. &quot;They&#039;ve optimized inter-chip communication and implemented mix-of-model strategies,&quot; says Wendy Chang of the Mercator Institute for China Studies. Their latest model is so efficient that it required only a tenth of the computing power needed to train Meta&#039;s Llama 3.1.<\/p>\n\n\n\n<p>Sharing DeepSeek&#039;s innovations openly has strengthened its global reputation. &quot;They demonstrate that advanced models can be built with fewer resources by optimizing training methods,&quot; Chang concludes.<\/p>\n\n\n\n<p>US restrictions may prove ineffective in containing the advancement of Chinese AI, as alternative strategies like DeepSeek&#039;s are emerging with success.<\/p>\n\n\n\n<p>Source Wired<\/p>","protected":false},"excerpt":{"rendered":"<p>Liang Wenfeng, once the head of a quantitative hedge fund in China, transformed his career by betting on artificial intelligence research. He invested considerable resources, stockpiling 10,000 Nvidia chips and assembling a team of brilliant young researchers. This bold project led, two years later, to the launch of DeepSeek, an initiative that quickly gained traction. <\/p>","protected":false},"author":2,"featured_media":3284,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[38],"tags":[],"class_list":["post-3281","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-blog"],"acf":[],"_links":{"self":[{"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/posts\/3281","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/comments?post=3281"}],"version-history":[{"count":0,"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/posts\/3281\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/media\/3284"}],"wp:attachment":[{"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/media?parent=3281"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/categories?post=3281"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/leonardonetwork.eu\/en\/wp-json\/wp\/v2\/tags?post=3281"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}