Whether it’s ChatGPT since the past couple of years or DeepSeek more recently, the field of artificial intelligence (AI) has seen rapid advancements, with models becoming increasingly large and ...
Distillation, also known as model or knowledge distillation, is a process where knowledge is transferred from a large, complex AI ‘teacher’ model to a smaller and more efficient ‘student’ model. Doing ...
A woman holds a cell phone in front of a computer screen displaying the DeepSeek logo (Photo by Artur Widak, NurPhoto via Getty Images) At this month’s Paris AI Summit, the global conversation around ...
The AI industry is witnessing a transformative trend: the use of distillation to make AI models smaller and cheaper. This shift, spearheaded by companies like DeepSeek and OpenAI, is reshaping the AI ...
The original version of this story appeared in Quanta Magazine. The Chinese AI company DeepSeek released a chatbot earlier this year called R1, which drew a huge amount of attention. Most of it ...