AI Distillation: The Race for Cheaper, More Efficient AI Models

AI Firms Embrace Distillation for Cheaper, More Efficient Models
Leading artificial intelligence companies, including OpenAI, Microsoft, and Meta, are increasingly adopting a technique known as "distillation" to develop more affordable and efficient AI models. This strategy is a direct response to the global race for AI dominance and aims to make advanced AI more accessible to consumers and businesses.
DeepSeek's Breakthrough and Industry Impact
The technique gained significant traction after China's DeepSeek successfully utilized distillation to create powerful yet efficient AI models. These models were built upon open-source systems previously released by competitors Meta and Alibaba. DeepSeek's achievement reportedly shook the confidence of Silicon Valley's AI leadership, causing a notable dip in the market value of major US tech stocks.
Understanding AI Distillation
AI distillation is a process where a large, sophisticated AI model, referred to as a "teacher" model, is used to train a smaller, more specialized "student" model. The teacher model generates data, which the student model then learns from. This effectively transfers the knowledge and predictive capabilities of the larger model to the smaller one, enabling faster and more cost-effective execution of specific tasks.
Olivier Godement, head of product for OpenAI's platform, described distillation as "magical," highlighting its ability to create highly capable yet inexpensive and fast AI models for specific applications.
The Cost of Large Language Models (LLMs)
Developing and maintaining large language models like OpenAI's GPT-4, Google's Gemini, and Meta's Llama requires immense data and computational resources. While exact figures are not publicly disclosed, the cost of training these foundational models is estimated to be in the hundreds of millions of dollars.
Distillation's Benefits for Developers and Businesses
Distillation offers a significant advantage by allowing developers and businesses to access the capabilities of these powerful models at a fraction of the cost. This enables the rapid deployment of AI models on devices such as laptops and smartphones, making AI-powered applications more accessible.
Industry Adoption and Partnerships
OpenAI provides its platform for distillation, allowing developers to learn from the LLMs that power products like ChatGPT. Microsoft, a major investor in OpenAI, has leveraged GPT-4 to distill its own smaller language models, such as Phi, as part of their commercial partnership.
Challenges and Limitations of Distillation
While distillation offers numerous benefits, experts note that it also comes with limitations. Ahmed Awadallah of Microsoft Research points out that reducing model size inevitably impacts overall capability. A distilled model might excel at a specific task, like summarizing emails, but would likely perform poorly on other, unrelated tasks.
David Cox, vice-president for AI models at IBM Research, suggests that most businesses do not require massive models for their operations. Distilled models are often sufficient for tasks like customer service chatbots or running on mobile devices.
Business Model Implications
The widespread adoption of distillation poses a challenge to the business models of leading AI firms. While distilled models are cheaper to run and create, they also generate less revenue. Companies like OpenAI may charge less for distilled models due to their reduced computational requirements.
However, OpenAI's Godement maintains that large language models will remain essential for complex, high-stakes tasks where accuracy and reliability command a premium. He also notes that large models are crucial for discovering new capabilities that can later be distilled into smaller ones.
Protecting Intellectual Property in the Age of Distillation
OpenAI is actively working to prevent its large models from being used to train competitors. The company monitors usage and can revoke access for users suspected of exporting data for rival model training, as reportedly done with accounts linked to DeepSeek. However, detecting and preventing such misuse is challenging.
Douwe Kiela, CEO of Contextual AI, acknowledges the difficulty in completely preventing distillation, stating, "OpenAI has been trying to protect against distillation for a long time, but it is very hard to avoid it altogether."
Open Models and the Future of AI
Distillation also benefits proponents of open-source AI models, where technology is freely shared for further development. DeepSeek has made its recent models available to developers, aligning with the open-source philosophy.
Yann LeCun, Meta's chief AI scientist, expressed enthusiasm for distillation, stating, "We're going to use [distillation] and put it in our products right away. That's the whole idea of open source. You profit from everyone and everyone else's progress as long as those processes are open."
The Shifting Landscape of AI Innovation
Distillation accelerates the pace of AI development, allowing competitors to quickly replicate the capabilities of advanced models. DeepSeek's recent releases exemplify how companies can catch up rapidly, raising questions about the long-term advantage of being a first-mover in LLM development.
IBM's Cox notes the dynamic nature of the AI landscape, where significant investment in developing models the "hard way" can be quickly matched by competitors leveraging more efficient methods like distillation, creating a "tricky business landscape."
Additional reporting by Michael Acton in San Francisco.
© 2025 The Financial Times Ltd. All rights reserved. Not to be redistributed, copied, or modified in any way.
Original article available at: https://arstechnica.com/ai/2025/03/ai-firms-follow-deepseeks-lead-create-cheaper-models-with-distillation/