Close Menu
    What's Hot

    Uber Eats, Freight Could Be Edge for Robotaxis: CEO Dara Khosrowshahi

    February 4, 2026

    ‘$60K–$65K Looks Realistic’, Analysts Warn

    February 4, 2026

    Amazon Q4 Earnings Preview: AWS margins, AI capex in focus (AMZN:NASDAQ)

    February 4, 2026
    Facebook X (Twitter) Instagram
    Hot Paths
    • Home
    • News
    • Politics
    • Money
    • Personal Finance
    • Business
    • Economy
    • Investing
    • Markets
      • Stocks
      • Futures & Commodities
      • Crypto
      • Forex
    • Technology
    Facebook X (Twitter) Instagram
    Hot Paths
    Home»Business»AI companies race to use ‘distillation’ to produce cheaper models
    Business

    AI companies race to use ‘distillation’ to produce cheaper models

    Press RoomBy Press RoomMarch 2, 2025No Comments5 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Leading artificial intelligence firms including OpenAI, Microsoft and Meta are turning to a process called “distillation” in the global race to create AI models that are cheaper for consumers and businesses to adopt.

    The technique caught widespread attention after China’s DeepSeek used it to build powerful and efficient AI models based on open-source systems released by competitors Meta and Alibaba. The breakthrough rocked confidence in Silicon Valley’s AI leadership, leading Wall Street investors to wipe billions of dollars of value from US Big Tech stocks.

    Through distillation, companies take a large language model — dubbed a “teacher” model — which generates the next likely word in a sentence. The teacher model generates data which then trains a smaller “student” model, helping to quickly transfer knowledge and predictions of the bigger model to the smaller one.

    While distillation has been widely used for years, recent advances have led industry experts to believe the process will increasingly be a boon for start-ups seeking cost-effective ways to build applications based on the technology.

    Some content could not load. Check your internet connection or browser settings.

    “Distillation is quite magical,” said Olivier Godement, head of product for OpenAI’s platform. “It’s the process of essentially taking a very large smart frontier model and using that model to teach a smaller model . . . very capable in specific tasks that is super cheap and super fast to execute.”

    Large language models such as OpenAI’s GPT-4, Google’s Gemini and Meta’s Llama require massive amounts of data and computing power to develop and maintain. While the companies have not revealed precise figures for how much it costs to train large models, it is likely to be hundreds of millions of dollars.

    Thanks to distillation, developers and businesses can access these models’ capabilities at a fraction of the price, allowing app developers to run AI models quickly on devices such as laptops and smartphones. 

    Recommended

    Joe Cummings illustration of Person in the News, Liang Wenfeng

    Developers can use OpenAI’s platform for distillation, learning from the large language models that underpin products like ChatGPT. OpenAI’s largest backer, Microsoft, used GPT-4 to distil its small language family of models Phi as part of a commercial partnership after investing nearly $14bn into the company.

    However, the San Francisco-based start-up has said it believes DeepSeek distilled OpenAI’s models to train its competitor, a move that would be against its terms of service. DeepSeek has not commented on the claims.

    While distillation can be used to create high-performing models, experts add they are more limited.

    “Distillation presents an interesting trade-off; if you make the models smaller, you inevitably reduce their capability,” said Ahmed Awadallah of Microsoft Research, who said a distilled model can be designed to be very good at summarising emails, for example, “but it really would not be good at anything else.”

    David Cox, vice-president for AI models at IBM Research, said most businesses do not need a massive model to run their products, and distilled ones are powerful enough for purposes such as customer service chatbots or running on smaller devices like phones.

    “Anytime you can [make it less expensive] and it gives you the right performance you want, there is very little reason not to do it,” he added.

    That presents a challenge to many of the business models of leading AI firms. Even if developers use distilled models from companies like OpenAI, they cost far less to run, are less expensive to create, and, therefore, generate less revenue. Model-makers like OpenAI often charge less for the use of distilled models as they require less computational load.

    Yet, OpenAI’s Godement argued that large language models will still be required for “high intelligence and high stakes tasks” where “businesses are willing to pay more for a high level of accuracy and reliability”. He added that large models will also be needed to discover new capabilities that can then be distilled into smaller ones.

    Still, the company aims to prevent its large models from being distilled to train a competitor. OpenAI has teams monitoring usage and can remove access to users it suspects are generating vast amounts of data to export and train a rival, as it has apparently done with accounts it believes were linked to DeepSeek. Yet much of this action happens retroactively.

    “OpenAI has been trying to protect against distillation for a long time, but it is very hard to avoid it altogether,” said Douwe Kiela, chief executive of Contextual AI, a start-up building information retrieval tools for enterprises.

    Recommended

    A montage of the logos of DeepSeek, Meta, Anthropic and OpenAI

    Distillation is also a victory for advocates of open models, where the technology is made freely available for developers to build upon. DeepSeek has made its recent models also open for developers.

    “We’re going to use [distillation] and put it in our products right away,” said Yann LeCun, Meta’s chief AI scientist. “That’s the whole idea of open source. You profit from everyone and everyone else’s progress as long as those processes are open.”

    Distillation also means that model-makers can spend billions of dollars to advance the capabilities of AI systems but still face competitors that often catch up quickly, as DeepSeek’s recent releases demonstrate. This raises questions about the first-mover advantage in building LLMs when their capabilities can be replicated in a matter of months.

    “In a world where things are moving so fast . . . you could actually spend a lot of money, doing it the hard way, and then the rest of the field is right on your heels,” IBM’s Cox said. “So it is an interesting and tricky business landscape.”

    Additional reporting Michael Acton in San Francisco

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Press Room

    Related Posts

    City fears mount that Budget will target banks to help fill £20bn fiscal hole

    August 29, 2025

    Renewable food is on the horizon

    August 28, 2025

    Bankers learn of firings via premature email to hand back their laptops

    August 28, 2025
    Leave A Reply Cancel Reply

    LATEST NEWS

    Uber Eats, Freight Could Be Edge for Robotaxis: CEO Dara Khosrowshahi

    February 4, 2026

    ‘$60K–$65K Looks Realistic’, Analysts Warn

    February 4, 2026

    Amazon Q4 Earnings Preview: AWS margins, AI capex in focus (AMZN:NASDAQ)

    February 4, 2026

    Meet the Billionaire Owners Behind Every NFL Team

    February 4, 2026
    POPULAR
    Business

    The Business of Formula One

    May 27, 2023
    Business

    Weddings and divorce: the scourge of investment returns

    May 27, 2023
    Business

    How F1 found a secret fuel to accelerate media rights growth

    May 27, 2023
    Advertisement
    Load WordPress Sites in as fast as 37ms!

    Archives

    • February 2026
    • January 2026
    • December 2025
    • November 2025
    • October 2025
    • September 2025
    • August 2025
    • July 2025
    • June 2025
    • May 2025
    • April 2025
    • March 2025
    • February 2025
    • January 2025
    • December 2024
    • November 2024
    • April 2024
    • March 2024
    • February 2024
    • January 2024
    • December 2023
    • November 2023
    • October 2023
    • September 2023
    • May 2023

    Categories

    • Business
    • Crypto
    • Economy
    • Forex
    • Futures & Commodities
    • Investing
    • Market Data
    • Money
    • News
    • Personal Finance
    • Politics
    • Stocks
    • Technology

    Your source for the serious news. This demo is crafted specifically to exhibit the use of the theme as a news site. Visit our main page for more demos.

    We're social. Connect with us:

    Facebook X (Twitter) Instagram Pinterest YouTube

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • Home
    • Buy Now
    © 2026 ThemeSphere. Designed by ThemeSphere.

    Type above and press Enter to search. Press Esc to cancel.