DeepSeek’s success learning from bigger AI models raises questions about the billions being spent on the most advanced technology.
Since Chinese artificial intelligence (AI) start-up DeepSeek rattled Silicon Valley and Wall Street with its cost-effective models, the company has been accused of data theft through a practice that is common across the industry.
Whether it's ChatGPT since the past couple of years or DeepSeek more recently, the field of artificial intelligence (AI) has seen rapid advancements, with models becoming increasingly large and complex.
One possible answer being floated in tech circles is distillation, an AI training method that uses bigger "teacher" models to train smaller but faster-operating "student" models.
Top White House advisers this week expressed alarm that China's DeepSeek may have benefited from a method that allegedly piggybacks off the advances of US rivals called "distillation."
Microsoft and OpenAI are investigating whether DeepSeek, a Chinese artificial intelligence startup, illegally copying proprietary American technology, sources told Bloomberg
OpenAI believes DeepSeek used a process called “distillation,” which helps make smaller AI models perform better by learning from larger ones.
If there are elements that we want a smaller AI model to have, and the larger models contain it, a kind of transference can be undertaken, formally known as knowledge distillation since you ...
Since the Chinese AI startup DeepSeek released its powerful large language model R1, it has sent ripples through Silicon Valley and the U.S. stock market, sparking widespread discussion and debate.
The new AI app DeepSeek disrupted global markets this week after releasing a model that could compete with US models like ChatGPT but was more cost-effective. View on euronews
NPR's Mary Louise Kelly talks with Chris Lehane, chief global affairs officer of OpenAI, about Stargate, DeepSeek and the future of AI development.