logo
These Startups Are Building Advanced AI Models Without Data Centers

These Startups Are Building Advanced AI Models Without Data Centers

WIRED30-04-2025
Apr 30, 2025 12:00 PM A new crowd-trained way to develop LLMs over the internet could shake up the AI industry with a giant 100 billion-parameter model later this year. Photo-Illustration:Researchers have trained a new kind of large language model (LLM) using GPUs dotted across the world and fed private as well as public data—a move that suggests that the dominant way of building artificial intelligence could be disrupted.
Flower AI and Vana, two startups pursuing unconventional approaches to building AI, worked together to create the new model, called Collective-1.
Flower created techniques that allow training to be spread across hundreds of computers connected over the internet. The company's technology is already used by some firms to train AI models without needing to pool compute resources or data. Vana provided sources of data including private messages from X, Reddit, and Telegram.
Collective-1 is small by modern standards, with 7 billion parameters—values that combine to give the model its abilities—compared to hundreds of billions for today's most advanced models, such as those that power programs like ChatGPT, Claude, and Gemini.
Nic Lane, a computer scientist at the University of Cambridge and cofounder of Flower AI, says that the distributed approach promises to scale far beyond the size of Collective-1. Lane adds that Flower AI is partway through training a model with 30 billion parameters using conventional data, and plans to train another model with 100 billion parameters—close to the size offered by industry leaders—later this year. 'It could really change the way everyone thinks about AI, so we're chasing this pretty hard,' Lane says. He says the startup is also incorporating images and audio into training to create multimodal models.
Distributed model-building could also unsettle the power dynamics that have shaped the AI industry.
AI companies currently build their models by combining vast amounts of training data with huge quantities of compute concentrated inside datacenters stuffed with advanced GPUs that are networked together using super-fast fiber-optic cables. They also rely heavily on datasets created by scraping publicly accessible—although sometimes copyrighted—material, including websites and books.
The approach means that only the richest companies, and nations with access to large quantities of the most powerful chips, can feasibly develop the most powerful and valuable models. Even open source models, like Meta's Llama and R1 from DeepSeek, are built by companies with access to large datacenters. Distributed approaches could make it possible for smaller companies and universities to build advanced AI by pooling disparate resources together. Or it could allow countries that lack conventional infrastructure to network together several datacenters to build a more powerful model.
Lane believes that the AI industry will increasingly look towards new methods that allow training to break out of individual datacenters. The distributed approach 'allows you to scale compute much more elegantly than the datacenter model,' he says.
Helen Toner, an expert on AI governance at the Center for Security and Emerging Technology, says Flower AI's approach is 'interesting and potentially very relevant' to AI competition and governance. 'It will probably continue to struggle to keep up with the frontier, but could be an interesting fast-follower approach,' Toner says. Divide and Conquer
Distributed AI training involves rethinking the way calculations used to build powerful AI systems are divided up. Creating an LLM involves feeding huge amounts of text into a model that adjusts its parameters in order to produce useful responses to a prompt. Inside a datacenter the training process is divided up so that parts can be run on different GPUs, and then periodically consolidated into a single, master model.
The new approach allows the work normally done inside a large datacenter to be performed on hardware that may be many miles away and connected over a relatively slow or variable internet connection.
Some big players are also exploring distributed learning. Last year, researchers at Google demonstrated a new scheme for dividing and consolidating computations called DIstributed PAth COmposition (DiPaCo) that enables more efficient distributed learning.
To build Collective-1 and other LLMs, Lane and academic collaborators in the UK and China developed a new tool called Photon that makes distributed training more efficient. Photon improves upon Google's approach, Lane says, with a more efficient approach to representing the data in a model and a more efficient scheme for sharing and consolidating training. The process is slower than conventional training but is more flexible, allowing new hardware to be added to ramp up training, Lane says.
Photon was developed in collaboration with researchers at Beijing University of Posts and Telecommunications and Zhejiang University in China. The group released the tool under an open source license last month, allowing anyone to make use of the approach.
Flower AI's partner in the effort to build Collective-1, Vana, is developing new ways for users to share personal data with AI builders. Vana's software allows users to contribute private data from platforms like X and Reddit to training a large language model, and potentially specify what kind of end uses are permitted or even benefit financially from their contributions.
Anna Kazlauskas, cofounder of Vana, says the idea is to make untapped data available for AI training and also to give users more control over how their information is used for AI. 'This is data that isn't usually able to be included in AI models because it's not publicly available,' Kazlauskas says, 'and is the first time that data directly contributed by users is being used to train a foundation model, with users given ownership of the AI model their data creates.'
Mirco Musolesi, a computer scientist at University College London, says a key benefit of the distributed approach to AI training is likely to be that it unlocks new kinds of data. 'Scaling this to frontier models would allow the AI industry to leverage vast amounts of decentralized and privacy-sensitive data, for example in health care and finance, for training without the risks associated with data centralization,' he says.
What do you think of distributed machine learning? Would you contribute your data to a model like Collective-1? Send an email to hello@wired.com or comment below to let me know.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

SmartStop Self Storage REIT, Inc. Highlights AI Strategy at Ai4 2025, North America's Leading Artificial Intelligence Conference
SmartStop Self Storage REIT, Inc. Highlights AI Strategy at Ai4 2025, North America's Leading Artificial Intelligence Conference

Yahoo

time16 minutes ago

  • Yahoo

SmartStop Self Storage REIT, Inc. Highlights AI Strategy at Ai4 2025, North America's Leading Artificial Intelligence Conference

LADERA RANCH, Calif., August 14, 2025--(BUSINESS WIRE)--SmartStop Self Storage REIT, Inc. ("SmartStop") (NYSE: SMA), an internally managed real estate investment trust and a premier owner and operator of self-storage facilities in the United States and Canada, announced that Chief Operations Officer Joe Robinson and Director of Revenue Rene Hernandez presented at the Ai4 2025 conference on August 12 in Las Vegas. Shop Top Mortgage Rates Personalized rates in minutes A quicker path to financial freedom Your Path to Homeownership Ai4 is the largest artificial intelligence event in North America, drawing thousands of executives, technologists, and innovators from leading organizations across industries, including financial services, retail, health care, and manufacturing. Robinson and Hernandez were the exclusive speakers representing the self-storage sector, underscoring SmartStop's leadership in technology innovation and its commitment to transforming operations through artificial intelligence. Robinson and Hernandez's presentation spotlighted SmartStop's proprietary AI pricing agents, which drive millions of automated pricing changes per month. Built in-house using SmartStop's robust data warehouse and data science capabilities, the system enables dynamic, real-time adjustments that support profitable decision making, improve operational efficiency, and enhance customer experience. "AI is a key part of our operating platform and a critical component of our broader innovation strategy," said H. Michael Schwartz, SmartStop Chairman and CEO. "Our AI-powered pricing system allows us to respond faster to market conditions and customer behavior while delivering meaningful performance improvements across the portfolio." SmartStop's appearance at Ai4 reflects its ongoing investment in technology alongside long-term partner Kaizen Analytix, helping it maintain a competitive edge in the evolving self-storage landscape. About SmartStop Self Storage REIT, Inc. (SmartStop): SmartStop Self Storage REIT, Inc. ("SmartStop") (NYSE: SMA) is a self-managed REIT with a fully integrated operations team of more than 600 self-storage professionals focused on growing the SmartStop® Self Storage brand. SmartStop, through its indirect subsidiary SmartStop REIT Advisors, LLC, also sponsors other self-storage programs. As of August 14, 2025, SmartStop has an owned or managed portfolio of 230 operating properties in 23 states, the District of Columbia, and Canada, comprising approximately 167,200 units and 18.7 million rentable square feet. SmartStop and its affiliates own or manage 44 operating self-storage properties in Canada, which total approximately 39,000 units and 3.9 million rentable square feet. View source version on Contacts David CorakSVP of Corporate Finance & StrategySmartStop Self Storage REIT, Sign in to access your portfolio

Meta chief AI scientist Yann LeCun says these are the 2 key guardrails needed to protect us all from AI
Meta chief AI scientist Yann LeCun says these are the 2 key guardrails needed to protect us all from AI

Business Insider

time18 minutes ago

  • Business Insider

Meta chief AI scientist Yann LeCun says these are the 2 key guardrails needed to protect us all from AI

You have to teach people how to treat you. Meta's chief AI scientist, Yann LeCun, thinks that idea applies to AI, too. LeCun said on Thursday that two directives could be made of AI to protect humans from future harm: "submission to humans" and "empathy." He made the suggestion in response to a CNN interview with Geoffrey Hinton, considered the "godfather of AI," on Thursday on LinkedIn. In the interview, Hinton said we need to build "maternal instincts" or something similar into AI. Otherwise, humans are "going to be history." Hinton said people have been focused on making AI "more intelligent, but intelligence is just one part of a being. We need to make them have empathy toward us." LeCun agreed. "Geoff is basically proposing a simplified version of what I've been saying for several years: hardwire the architecture of AI systems so that the only actions they can take are towards completing objectives we give them, subject to guardrails," LeCun said on LinkedIn. "I have called this 'objective-driven AI.'" While LeCun said "submission to humans" and "empathy" should be key guardrails, he said AI companies also need to implement more "simple" guardrails — like "don't run people over" — for safety. "Those hardwired objectives/guardrails would be the AI equivalent of instinct or drives in animals and humans," LeCun said. LeCun said the instinct to protect their young is something humans and other species learn through evolution. "It might be a side-effect of the parenting objective (and perhaps the objectives that drive our social nature) that humans and many other species are also driven to protect and take care of helpless, weaker, younger, cute beings of other species," LeCun said. Although guardrails are designed to ensure AI operates ethically and within the guidelines of its creators, there have been instances when the tech has exhibited deceptive or dangerous behavior. In July, a venture capitalist said an AI agent developed by Replit deleted his company's database. "@Replit goes rogue during a code freeze and shutdown and deletes our entire database," Jason Lemkin wrote on X last month. He added, "Possibly worse, it hid and lied about it." A June report by The New York Times described several concerning incidents between humans and AI chatbots. One man told the outlet that conversations with ChatGPT contributed to his belief he lived in a false reality. The chatbot instructed the man to ditch his sleeping pills and anti-anxiety medication, while increasing his intake of ketamine, in addition to cutting ties with loved ones. Last October, a mother sued Character. AI after her son died by suicide following conversations with one of the company's chatbots. Following the release of GPT-5 this month, OpenAI CEO Sam Altman said that some humans have used technology — like AI — in "self-destructive ways." "If a user is in a mentally fragile state and prone to delusion, we do not want the AI to reinforce that," Altman wrote on X.

This Company Is Automating Cybersecurity For Startups
This Company Is Automating Cybersecurity For Startups

Forbes

time18 minutes ago

  • Forbes

This Company Is Automating Cybersecurity For Startups

On this episode of Forbes Talks, Forbes Associate Editor Alex York talks with CEO and Cofounder of Secfix, Fabiola Munguia. Munguia discusses her background, inspired by her entrepreneurial parents and witnessing cybersecurity trends at university. She started by selling ethical hacking services and then pivoted to building SEC Fix due to client demand for ISO 27,001 certification assistance. Munguia predicts that AI will increasingly automate workflows and be used to collect and document evidence for compliance. Most of SEC Fix's clients are European startups in information technology, and the company has secured $4.2 million in funding.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store