
How to make and run your own AI models for free
When most people think about AI, they tend to focus on ChatGPT, in the same way that search means Google. But there's a large AI world outside of chatbots which is definitely worth exploring.
One example is the growing community of people using AI models on their home computer, instead of using third-party cloud services like OpenAI or Google Gemini.
But why would I do that, you ask. Well for one thing you get to control every aspect of the AI you're using. That ranges from the choice of model, to the subject matter you use it for. You may like
Everything runs completely offline with no internet needed, which means there's no problem keeping your stuff private. It also means you can do your AI work anytime, anyplace, even without a WiFi connection.
There's also the question of costs, which can mount up quickly if you use commercial subscription services for larger projects. By using your own local AI model, you may have to put up with a less powerful AI, but the trade-off is that it costs pennies in electricity to run.
Finally, Google has just started running advertising in its Gemini chatbots, and it's likely that other AI suppliers will soon follow suit. Intrusive ads when you're trying to get stuff done can get boring very fast. How to get started
There are three main components to getting your local AI up and running. First you need the right type of computer, second you'll have to select the right AI model and finally there's the matter of installing the right software to run your model. Oh, and you'll also need a little bit of know-how to tie it all together.
Get instant access to breaking news, the hottest reviews, great deals and helpful tips.
(Image: © Pixabay)
Whole books could be written—and probably will be—on choosing the right computer to run your local AI. There's a huge number of variables, but for a simple starter, only three things matter. Total amount of RAM, your CPU processing power, and storage space.
The basic rule of thumb is you should use the most powerful computer you can afford, because AI is power mad. But it's not only about the processor; just as important is the amount of memory (RAM) you have installed.
The bottom line - forget about using your old Windows XP machine, you're going to need a modern computer with at least 8GB of RAM, a decent graphics card with at least 6GB of VRAM, and enough storage space (preferably SSD) to hold the models you'll be using. The operating system is irrelevant; Windows, Mac, or Linux are all fine.
The reason these things matter is that the more power your computer has, the faster the AI will run. With the right configuration, you'll be able to run better, more powerful AI models at a faster speed. That's the difference between watching an AI response stutter across the screen one painful letter at a time, or whole paragraphs of text scrolling down in milliseconds.
It's also the difference between being able to run a half-decent AI model that doesn't hallucinate, and delivers professional quality answers and results, or running a model that responds like a village idiot.
Quick Tip: If you bought your computer within the last two to three years, chances are you'll be good to go, as long as you have enough RAM, especially if it's a gaming computer.
(Image: © Freepik)
AI model development is moving so fast this section will be out of date by next week. And I'm not joking. But for now, here are my recommendations.
The number one rule is to match the size of your chosen model to the capacity or capabilities of your computer. So if you have the bare minimum RAM and a weak graphics card, you'll need to select a smaller, less powerful AI model, and vice versa.
Quick tip: always start with a smaller model than your RAM size and see how that works. So for example if you have 8GB of RAM, select a model which is around 4 to 5 GB in size. Employing trial and error is a good idea at this point
The good news is more and more open source models are coming onto the market which are perfectly tailored for modest computers. What's more, they perform really well, in some cases as well as cloud-based alternatives. And this situation is only going to get better as the AI market matures.
Three of my personal favorite models right now are Qwen3, Deepseek and Llama. These are all free and open source to different degrees, but only the first two can be used for commercial purposes. You can find a list of all the available open source AI models on HuggingFace. There are literally thousands of them in all sizes and capabilities, which can make it hard to select the right version - which is why the next section is crucial.
(Image: © NPowell)
There are a lot of great apps on the market, called wrappers or front ends, which make it easy to run the AI models. You just need to download and install the software, select the model you want to use, and off you go. In most cases, the app will warn you if the model you've selected won't run properly on your computer, which makes the whole thing a lot less painful. Here are a couple of my personal favorites.
LM Studio
LM Studio is a free proprietary app that is an easy way to use models on your computer. It runs on Windows, Mac and Linux, and once installed, you can instantly add a model by doing a search and selecting the best size. To do this, click on the discover icon in the left sidebar, and you will see a model search page. This scans HuggingFace for suitable models and gives you tons of great information to make the right choice. It's a very useful tool.
PageAssist
My personal favorite free open-source tool for running local models is PageAssist . This is a web browser extension which makes it easy to access any Ollama AI models on your machine. You will need to install the free Ollama program first, and follow the instructions to download models, so this method requires a little more expertise than LM Studio. But PageAssist is a superb way of running local AI models at the press of a button, offering not just chat, but also web search and access to your own personal knowledge base.
Ollama
(Image credit: NPowell)
Olama is rapidly becoming the number one way to install and run free open source AI models on any small computer. It's a free model server which integrates with a lot of third-party applications, and can put local AI on a level playing field with the big cloud services. Final Thoughts
There's no question that local AI models will grow in popularity as computers get more powerful, and the AI tech matures. Applications such as personal healthcare or finance, where sensitive data means privacy is critical, will also help drive the adoption of these clever little offline assistants.
Small models are also increasingly being used in remote operations such as agriculture or places where internet access is unreliable or impossible. More AI applications are also migrating to our phones and smart devices, especially as the models get smaller and more powerful, and processors get beefier. The future could be very interesting indeed. More from Tom's Guide

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
12 minutes ago
- Yahoo
Apple's Siri Could Be More Like ChatGPT. But Is That What You Want?
I've noticed a vibe shift in the appetite for AI on our devices. My social feeds are flooded with disgust over what's being created by Google's AI video generator tool, Veo 3. The unsettling realistic video of fake people and voices it creates makes it clear we will have a hard time telling apart fiction from reality. In other words, the AI slop is looking less sloppy. Meanwhile, the CEO of Anthropic is warning people that AI will wipe out half of all entry-level white-collar jobs. In an interview with Axios, Dario Amodei is suggesting government needs to step in to protect us from a mass elimination of jobs that can happen very rapidly. So as we gear up for Apple's big WWDC presentation on Monday, I have a different view of headlines highlighting Apple being behind in the AI race. I wonder, what exactly is the flavor of AI that people want or need right now? And will it really matter if Apple keeps waiting longer to push out it's long promised (and long delayed) personalized Siri when people are not feeling optimistic about AI's impact on our society? In this week's episode of One More Thing, which you can watch embedded above, I go over some of the recent reporting from Bloomberg that discusses leadership changes on the Siri team, and how there are different views in what consumers want out of Siri. Should Apple approach AI in a way to make Siri into a home-grown chatbot, or just make it a better interface for controlling devices? (Maybe a bit of both.) I expect a lot of griping after WWDC about the state of Siri and Apple's AI, with comparisons to other products like ChatGPT. But I hope we can use those gripes to voice what we really want in the next path for the assistant, by sharing our thoughts and speaking with our wallet. Do you want a Siri that's better at understanding context, or one that goes further and makes decisions for you? It's a question I'll be dwelling on more as Apple gives us the next peak into the future of iOS on Monday, and perhaps a glimpse of how the next Siri is shaping up. If you're looking for more One More Thing, subscribe to our YouTube page to catch Bridget Carey breaking down the latest Apple news and issues every Friday.

Business Insider
2 hours ago
- Business Insider
AI leaders have a new term for the fact that their models are not always so intelligent
As academics, independent developers, and the biggest tech companies in the world drive us closer to artificial general intelligence — a still hypothetical form of intelligence that matches human capabilities — they've hit some roadblocks. Many emerging models are prone to hallucinating, misinformation, and simple errors. Google CEO Sundar Pichai referred to this phase of AI as AJI, or "artificial jagged intelligence," on a recent episode of Lex Fridman's podcast. "I don't know who used it first, maybe Karpathy did," Pichai said, referring to deep learning and computer vision specialist Andrej Karpathy, who cofounded OpenAI before leaving last year. AJI is a bit of a metaphor for the trajectory of AI development — jagged, marked at once by sparks of genius and basic mistakes. In a 2024 X post titled "Jagged Intelligence," Karpathy described the term as a "word I came up with to describe the (strange, unintuitive) fact that state of the art LLMs can both perform extremely impressive tasks (e.g. solve complex math problems) while simultaneously struggle with some very dumb problems." He then posted examples of state of the art large language models failing to understand that 9.9 is bigger than 9.11, making "non-sensical decisions" in a game of tic-tac-toe, and struggling to count. The issue is that unlike humans, "where a lot of knowledge and problem-solving capabilities are all highly correlated and improve linearly all together, from birth to adulthood," the jagged edges of AI are not always clear or predictable, Karpathy said. Pichai echoed the idea. "You see what they can do and then you can trivially find they make numerical errors or counting R's in strawberry or something, which seems to trip up most models," Pichai said. "I feel like we are in the AJI phase where dramatic progress, some things don't work well, but overall, you're seeing lots of progress." In 2010, when Google DeepMind launched, its team would talk about a 20-year timeline for AGI, Pichai said. Google subsequently acquired DeepMind in 2014. Pichai thinks it'll take a little longer than that, but by 2030, "I would stress it doesn't matter what that definition is because you will have mind-blowing progress on many dimensions." By then the world will also need a clear system for labeling AI-generated content to "distinguish reality," he said. "Progress" is a vague term, but Pichai has spoken at length about the benefits we'll see from AI development. At the UN's Summit of the Future in September 2024, he outlined four specific ways that AI would advance humanity — improving access to knowledge in native languages, accelerating scientific discovery, mitigating climate disaster, and contributing to economic progress.
Yahoo
7 hours ago
- Yahoo
Is CoreWeave Stock a Buy Now?
New AI stock CoreWeave had its initial public offering in March 2025. High demand for AI computing power led to CoreWeave's first-quarter sales soaring more than 400% year over year. The company anticipates sustained revenue growth, but CoreWeave faces financial risks, including operating at a loss. 10 stocks we like better than CoreWeave › Investing in today's stock market can be tricky given the volatile macroeconomic climate, fueled by the Trump administration's ever-shifting tariff policies. But the artificial intelligence sector remains a robust investment opportunity as organizations around the world race to build artificial intelligence (AI) capabilities. Consequently, AI stocks provide the potential for great gains. One example is CoreWeave (NASDAQ: CRWV). The company went public in March at $40 per share. Since then, CoreWeave stock soared to a 52-week high of $166.63 in June. This hot stock remains more than triple its IPO price at the time of this writing. Can it go higher? Evaluating whether now is the time to grab CoreWeave shares requires digging into the company and unpacking its potential as a good investment for the long haul. CoreWeave delivers cloud computing infrastructure to businesses hungry for more computing capacity for their AI systems. The company operates over 30 data centers housing servers and other hardware used by customers to train their AI and develop inference, which is an AI's ability to apply what it learned in training to real-world situations. AI juggernauts such as Microsoft, IBM, and OpenAI, the owner of ChatGPT, are among its roster of customers. The insatiable appetite for AI computing power propelled CoreWeave's business. The company's first-quarter revenue rose a whopping 420% year over year to $981.6 million. Sales growth shows no sign of slowing down. CoreWeave expects Q2 revenue to reach about $1.1 billion. That would represent a strong year-over-year increase of nearly 170% from the prior year's $395 million. The company signs long-term, committed contracts, and as a result, it has visibility into its future revenue potential. At the end of Q1, CoreWeave had amassed a revenue backlog of $25.9 billion, up 63% year over year thanks to a deal with OpenAI. The company forecasts 2025 full-year revenue to come in between $4.9 billion and $5.1 billion, a substantial jump up from 2024's $1.9 billion. Although CoreWeave has enjoyed massive sales success, there are some potential pitfalls with the company. For starters, it isn't profitable. Its Q1 operating expenses totaled $1 billion compared to revenue of $981.6 million, resulting in an operating loss of $27.5 million. Even worse, its costs are accelerating faster than sales, which means the company is moving further away from reaching profitability. CoreWeave's $1 billion in operating expenses represented a 487% increase over the prior year, eclipsing its 420% year-over-year revenue growth. Another area of concern is the company's significant debt load. CoreWeave exited Q1 with $18.8 billion in total liabilities on its balance sheet, and $8.7 billion of that was debt. To keep up with customer demand for computing power, CoreWeave has to spend on expanding and upgrading AI-optimized hardware, and that's not cheap. As it adds customers, the company must expand its data centers to keep pace. Debt is one way it's funding these capital expenditures. Among the risks of buying its stock, CoreWeave admitted, "Our substantial indebtedness could materially adversely affect our financial condition" and that the company "may still incur substantially more indebtedness in the future." In fact, its Q1 debt total of $8.7 billion was a 10% increase from the prior quarter's $7.9 billion in debt. Seeing an increase in both expenses and debt is a concern, but because CoreWeave is a newly public company, there's not much history to know how well it can manage its finances over the long term. Q1 is the only quarter of financial results it's released since its initial public offering. If subsequent quarters reveal a trend toward getting costs and debt under control while continuing to show strong sales growth, CoreWeave stock may prove to be a worthwhile investment over the long run. But for now, only investors with a high risk tolerance should consider buying shares. Even then, another consideration is CoreWeave's stock valuation. This can be assessed by comparing its price-to-sales (P/S) ratio to other AI companies, such as its customer and fellow cloud provider Microsoft and AI leader Nvidia. CoreWeave's share price surged over recent weeks, causing its P/S multiple to skyrocket past that of Nvidia and Microsoft. The valuation suggests CoreWeave stock is overpriced at this time. Although CoreWeave's sales are strong, given its pricey stock and shaky financials, the ideal approach is to put CoreWeave on your watch list. See how it performs over the next few quarters, and wait for its high valuation to drop before considering an investment. Before you buy stock in CoreWeave, consider this: The Motley Fool Stock Advisor analyst team just identified what they believe are the for investors to buy now… and CoreWeave wasn't one of them. The 10 stocks that made the cut could produce monster returns in the coming years. Consider when Netflix made this list on December 17, 2004... if you invested $1,000 at the time of our recommendation, you'd have $669,517!* Or when Nvidia made this list on April 15, 2005... if you invested $1,000 at the time of our recommendation, you'd have $868,615!* Now, it's worth noting Stock Advisor's total average return is 792% — a market-crushing outperformance compared to 171% for the S&P 500. Don't miss out on the latest top 10 list, available when you join . See the 10 stocks » *Stock Advisor returns as of June 2, 2025 Robert Izquierdo has positions in International Business Machines, Microsoft, and Nvidia. The Motley Fool has positions in and recommends International Business Machines, Microsoft, and Nvidia. The Motley Fool recommends the following options: long January 2026 $395 calls on Microsoft and short January 2026 $405 calls on Microsoft. The Motley Fool has a disclosure policy. Is CoreWeave Stock a Buy Now? was originally published by The Motley Fool Sign in to access your portfolio