
Why AI is still making things up
AI makers could do more to limit chatbots' penchant for "hallucinating," or making stuff up — but they're prioritizing speed and scale instead.
Why it matters: High-profile AI-induced gaffes keep embarrassing users and the technology's unreliability continues to cloud its adoption.
The big picture: Hallucinations aren't quirks — they're a foundational feature of generative AI that some researchers say will never be fully fixed.
AI models predict the next word based on patterns in their training data and the prompts a user provides. They're built to try to satisfy users, and if they don't "know" the answer they guess.
Chatbots' fabrications were a problem when OpenAI CEO Sam Altman introduced ChatGPT in 2022, and the industry continues to remind users that they can't trust every fact a chatbot asserts.
Every week brings painful new evidence that users are not listening to these warnings.
Last week it was a report from Robert F. Kennedy Jr.'s Health and Human Services department citing studies that didn't exist. Experts found evidence suggesting OpenAI's tools were involved.
A week earlier, The Chicago Sun-Times published a print supplement with a summer reading list full of real authors, but hallucinated book titles.
AI legal expert Damien Charlotin tracks legal decisions in which lawyers have used evidence that featured AI hallucinations. His database details 30 instances in May 2025. Legal observers fear the total number is likely much higher.
Yes, but: AI makers are locked in fierce competition to top benchmarks, capture users and wow the media. They'd love to tamp down hallucinations, but not at the expense of speed.
Chatbots could note how confident the language model is in the accuracy of the response, but they have no real incentive for that, Tim Sanders, executive fellow at Harvard Business School and VP at software marketplace G2, told Axios. "That's the dirty little secret. Accuracy costs money. Being helpful drives adoption."
Between the lines: AI companies are making efforts to reduce hallucinations, mainly by trying to fill in gaps in training data.
Retrieval augmentation generation (RAG) is one process for grounding answers in contextually relevant documents or data.
RAG connects the model to trusted data so it can retrieve relevant information before generating a response, producing more accurate answers.
AWS offers Amazon Bedrock, a cloud service that allows customers to use various AI providers and responsible AI capabilities (including reducing hallucinations) to build generative AI applications.
AWS says its Bedrock Guardrails can filter over 75% of hallucinated responses.
Researchers from Google's DeepMind, along with Stanford University and University of Illinois at Urbana-Champaign, are working on a Search-Augmented Factuality Evaluator (SAFE), which uses AI to fact check AI.
Anthropic offers a guide to help developers limit hallucinations — including allowing the model to answer, "I don't know."
OpenAI's developer guide includes a section on how much accuracy is "good enough" for production, from both business and technical perspectives.
Researchers inside AI companies have raised alarms about hallucinations, but those warnings aren't always front of mind in organizations hell-bent on a quest for "superintelligence."
Raising money for building the next big model means companies have to keep making bigger promises about chatbots replacing search engines and AI agents replacing workers. Focusing on the technology's unreliability only undercuts those efforts.
The other side: Some AI researchers insist that the hallucination problem is overblown or at least misunderstood, and that it shouldn't discourage speedy adoption of AI to boost productivity.
"We should be using [genAI] twice as much as we're using it right now," Sanders told Axios.
Sanders disputes a recent New York Times article suggesting that as AI models get smarter, their hallucination rates get worse.
OpenAI's o3 and similar reasoning models are designed to solve more complex problems than regular chatbots. "It reasons. It iterates. It guesses over and over until it gets a satisfactory answer based on the prompt or goal," Sanders told Axios. "It's going to hallucinate more because, frankly, it's going to take more swings at the plate."
Hallucinations would be less of a concern, Sanders says, if users understood that genAI is designed to make predictions, not verify facts. He urges users to take a "trust, but verify" approach.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


CNBC
24 minutes ago
- CNBC
Filmmaker Justine Bateman on AI's impact on Hollywood and film
Justine Bateman, filmmaker and author, joins 'Power Lunch' to discuss the impact of AI on Hollywood.


Bloomberg
24 minutes ago
- Bloomberg
Anysphere CEO on Cursor Being Valued at $9.9 Billion
Anysphere, the company behind AI coding assistant Cursor, has raised its valuation to $9.9 billion, as some call the three-year-old startup the fastest-growing software company. Michael Truell, CEO of Anysphere, joins Caroline Hyde and Ed Ludlow on 'Bloomberg Technology.' (Source: Bloomberg)


The Verge
28 minutes ago
- The Verge
The Ray-Ban Meta smart glasses are on sale for their best price to date
With summer just around the corner, sunglasses make a practical and timely Father's Day gift. If you want to gift a pair that'll truly impress, right now you can buy the latest Ray-Ban Meta Smart Glasses starting at $239.20 (about $61 off) at Amazon, Best Buy, and Target. Beyond just protecting dad's eyes, these stylish smart glasses will make his life a lot easier. The built-in 12MP camera lets him snap hands-free photos and 1080p videos that are surprisingly good, and even livestream directly to Facebook or Instagram. They also feature solid sound and excellent call quality, thanks to five built-in microphones, so he can listen to music and take calls without ever pulling out his phone. The latest Meta Smart Glasses improve on the original with new AI features that make them even more useful. With just his voice, Dad can take photos, record videos, send messages, ask questions, or get translations in multiple languages. The AI can also remember objects, landmarks, and parking spots – and even help him scan QR codes with ease.