logo
Small Language Models Are the New Rage, Researchers Say

Small Language Models Are the New Rage, Researchers Say

WIRED13-04-2025

Apr 13, 2025 2:00 AM Larger models can pull off a wider variety of feats, but the reduced footprint of smaller models makes them attractive tools. Illustration: Celsius Pictor for Quanta Magazine
The original version of this story appeared in Quanta Magazine.
Large language models work well because they're so large. The latest models from OpenAI, Meta, and DeepSeek use hundreds of billions of 'parameters'—the adjustable knobs that determine connections among data and get tweaked during the training process. With more parameters, the models are better able to identify patterns and connections, which in turn makes them more powerful and accurate.
But this power comes at a cost. Training a model with hundreds of billions of parameters takes huge computational resources. To train its Gemini 1.0 Ultra model, for example, Google reportedly spent $191 million. Large language models (LLMs) also require considerable computational power each time they answer a request, which makes them notorious energy hogs. A single query to ChatGPT consumes about 10 times as much energy as a single Google search, according to the Electric Power Research Institute.
In response, some researchers are now thinking small. IBM, Google, Microsoft, and OpenAI have all recently released small language models (SLMs) that use a few billion parameters—a fraction of their LLM counterparts.
Small models are not used as general-purpose tools like their larger cousins. But they can excel on specific, more narrowly defined tasks, such as summarizing conversations, answering patient questions as a health care chatbot, and gathering data in smart devices. 'For a lot of tasks, an 8 billion–parameter model is actually pretty good,' said Zico Kolter, a computer scientist at Carnegie Mellon University. They can also run on a laptop or cell phone, instead of a huge data center. (There's no consensus on the exact definition of 'small,' but the new models all max out around 10 billion parameters.)
To optimize the training process for these small models, researchers use a few tricks. Large models often scrape raw training data from the internet, and this data can be disorganized, messy, and hard to process. But these large models can then generate a high-quality data set that can be used to train a small model. The approach, called knowledge distillation, gets the larger model to effectively pass on its training, like a teacher giving lessons to a student. 'The reason [SLMs] get so good with such small models and such little data is that they use high-quality data instead of the messy stuff,' Kolter said.
Researchers have also explored ways to create small models by starting with large ones and trimming them down. One method, known as pruning, entails removing unnecessary or inefficient parts of a neural network—the sprawling web of connected data points that underlies a large model.
Pruning was inspired by a real-life neural network, the human brain, which gains efficiency by snipping connections between synapses as a person ages. Today's pruning approaches trace back to a 1989 paper in which the computer scientist Yann LeCun, now at Meta, argued that up to 90 percent of the parameters in a trained neural network could be removed without sacrificing efficiency. He called the method 'optimal brain damage.' Pruning can help researchers fine-tune a small language model for a particular task or environment.
For researchers interested in how language models do the things they do, smaller models offer an inexpensive way to test novel ideas. And because they have fewer parameters than large models, their reasoning might be more transparent. 'If you want to make a new model, you need to try things,' said Leshem Choshen, a research scientist at the MIT-IBM Watson AI Lab. 'Small models allow researchers to experiment with lower stakes.'
The big, expensive models, with their ever-increasing parameters, will remain useful for applications like generalized chatbots, image generators, and drug discovery. But for many users, a small, targeted model will work just as well, while being easier for researchers to train and build. 'These efficient models can save money, time, and compute,' Choshen said.
Original story reprinted with permission from Quanta Magazine, an editorially independent publication of the Simons Foundation whose mission is to enhance public understanding of science by covering research developments and trends in mathematics and the physical and life sciences.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

OpenAI and Jony Ive's ‘io' brand has vanished, but their AI hardware deal remains
OpenAI and Jony Ive's ‘io' brand has vanished, but their AI hardware deal remains

The Verge

timean hour ago

  • The Verge

OpenAI and Jony Ive's ‘io' brand has vanished, but their AI hardware deal remains

OpenAI has scrubbed mentions of io, the hardware startup co-founded by famous Apple designer Jony Ive, from its website and social media channels. The sudden change closely follows their recent announcement of OpenAI's nearly $6.5 billion acquisition and plans to create dedicated AI hardware. OpenAI tells The Verge the deal is still happening, but it scrubbed mentions due to a trademark lawsuit from Iyo, the hearing device startup spun out of Google's moonshot factory. The announcement blog post and a nine-minute video featuring Ive and OpenAI CEO Sam Altman are no longer available. The blog post from Ive and Altman announcing the deal said, 'The io team, focused on developing products that inspire, empower and enable, will now merge with OpenAI to work more intimately with the research, engineering and product teams in San Francisco.' OpenAI spokesperson Kayla Wood: This page is temporarily down due to a court order following a trademark complaint from iyO about our use of the name 'io.' We don't agree with the complaint and are reviewing our options.

OpenAI pulls promotional materials around Jony Ive deal
OpenAI pulls promotional materials around Jony Ive deal

TechCrunch

timean hour ago

  • TechCrunch

OpenAI pulls promotional materials around Jony Ive deal

In Brief OpenAI appears to have pulled a much-discussed video promoting the friendship between CEO Sam Altman and legendary Apple designer Jony Ive (plus, incidentally, OpenAI's $6.5 billion deal to acquire Ive and Altman's device startup io) from its website and YouTube page. Does that suggest something is amiss with the acquisition, or with plans for Ive to lead design work at OpenAI? Not exactly, according to Bloomberg's Mark Gurman, who reports that the 'deal is on track and has NOT dissolved or anything of the sort.' Instead, he said a judge has issued a restraining order over the io name, forcing the company to pull all materials that used it. And indeed, an earlier report at Bloomberg Law noted that AI device maker IYO has brought a trademark lawsuit against OpenAI, with the judge suggesting she's open to IYO's argument that OpenAI's promotional video might already be creating consumer confusion. TechCrunch has reached out to OpenAI for comment. For now, at least, the video remains viewable on X.

Jony Ive Deal Removed from OpenAI Site Over Trademark Suit
Jony Ive Deal Removed from OpenAI Site Over Trademark Suit

Bloomberg

timean hour ago

  • Bloomberg

Jony Ive Deal Removed from OpenAI Site Over Trademark Suit

Marketing materials and video related to a blockbuster partnership between former Apple Inc. designer Jony Ive and OpenAI Inc. were removed from the web due to a trademark dispute. Social media users noticed Sunday that a video and website hosted by OpenAI that announced the artificial intelligence company's $6.5 billion acquisition of Ive's secretive AI hardware startup IO Products were no longer online, prompting speculation about a hiccup in the agreement.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store