Distillation: Making AI Models More Efficient and Affordable

1 min read
Source: Quanta Magazine
Distillation: Making AI Models More Efficient and Affordable
Photo: Quanta Magazine
TL;DR Summary

DeepSeek's use of knowledge distillation, a widely used AI technique that involves training smaller models using the outputs of larger ones, has sparked controversy but is a common practice in AI development. Originally developed in 2015 at Google to make ensemble models more efficient, distillation helps create smaller, cheaper, and faster AI models by transferring 'dark knowledge' from a teacher to a student model. It has become a fundamental tool in AI, enabling companies like Google, OpenAI, and Amazon to deploy powerful models more efficiently, and continues to be an active area of research and application.

Share this article

Reading Insights

Total Reads

0

Unique Readers

1

Time Saved

5 min

vs 6 min read

Condensed

90%

1,00396 words

Want the full story? Read the original article

Read on Quanta Magazine