logo
How Retrieval-Augmented Generation Could Stop AI Hallucinations

How Retrieval-Augmented Generation Could Stop AI Hallucinations

Forbes23-06-2025
Sagar Gupta, EST03 Inc., is an ERP Implementation Leader with over 20 years of experience in enterprise-scale technology transformations.
Large language models (LLMs) like OpenAI's GPT-4 and Google's PaLM have captured the imagination of industries ranging from healthcare to law. Their ability to generate human-like text has opened the doors to unprecedented automation and productivity. But there's a problem: Sometimes, these models make things up. This phenomenon—known as hallucination—is one of the most pressing issues in the AI space today.
The Hallucination Challenge
At its core, an LLM generates responses based on statistical associations learned from massive datasets. It's like a parrot with access to all the books ever written—but no real understanding of what's true or relevant. That's why hallucinations happen: The model is trained to sound plausible, not necessarily be accurate.
Researchers classify hallucinations into two main types:
• Intrinsic: These contradict known facts or include logical inconsistencies.
• Extrinsic: These are unverifiable, meaning there's no reliable source to back them up.
The root causes lie in incomplete training data, ambiguous prompts and the lack of real-time access to reliable information.
The RAG Solution
Retrieval-augmented generation (RAG) enriches traditional LLMs with a system that fetches relevant documents from a trusted database in real time. The model then uses these documents to generate responses grounded in actual content, rather than relying solely on what it 'remembers' from training.
The architecture typically includes:
• A retriever, often based on technologies like dense passage retrieval (DPR) or best matching 25 (BM25)
• A generator, usually a transformer-based model that crafts the response based on the retrieved data
This combination essentially transforms the LLM into an open-book test-taker rather than a guesser.
RAG In Action
Real-world experiments show promise. A 2021 study reported a 35% reduction in hallucinations in question-answering tasks using RAG. Similarly, models like DeepMind's RETRO and Meta's Atlas demonstrate significantly better factual accuracy by incorporating retrieval systems.
Innovations like the fusion-in-decoder (FiD) and REPLUG models take this further by improving how the model processes multiple retrieved documents or integrates them into frozen models for faster deployment.
But even RAG has its limits. If the retriever pulls the wrong information or the generator misinterprets it, hallucinations can still occur. And there's an added trade-off: Retrieval increases system complexity and inference time—no small issue in real-time applications.
Rethinking Evaluations
Evaluating hallucinations is another hurdle. Existing metrics like FactCC and FEVER try to measure factual consistency, but they often miss nuances. Human evaluations remain the gold standard, but they're costly and slow.
Researchers are now exploring reference-free factuality metrics and better ways to assess whether the retrieved documents actually support the generated answer.
What's Next?
Three exciting directions could further improve how we tackle hallucinations:
1. Differentiable Retrieval: Instead of separating the retriever and generator, future systems might train both components together in a fully end-to-end fashion. This could tighten the alignment between what's retrieved and what's generated.
2. Memory-Augmented Models: Some experts are exploring how AI can maintain long-term memory internally, reducing the need for external retrieval or complementing it when appropriate.
3. Fact-Aware Training: By incorporating factual correctness into the training objective itself—via techniques like reinforcement learning from human feedback—models might learn to prioritize truth over plausibility.
How RAG Helps Enforce Departmental Private Policies
Here's how RAG systems can support department-specific policies in real enterprise environments:
With RAG, AI assistants can answer employee questions about HR policies using only internal documents—like the company's official handbook or compliance playbook—ensuring no public or outdated data leaks into responses.
Examples: Confidential grievance reporting, DEI guidelines and code of conduct.
Use Case: An employee asks about the process for reporting harassment. Instead of guessing or fabricating, the AI pulls directly from the current internal grievance protocol.
Financial departments are governed by strict rules, often tailored to the business and changing frequently. RAG systems can help ensure AI-generated summaries, reports or answers reflect the latest finance policies pulled from internal financial controls documents or regulatory compliance handbooks.
Examples: Internal audit procedures, expense reimbursement rules and compliance with SOX (Sarbanes–Oxley).
Use Case: A junior accountant asks, 'Can I reimburse a client dinner without itemized receipts?' The AI retrieves the latest expense policy and provides an accurate, compliance-approved response.
LLMs trained on public data should never guess legal advice. RAG enables law departments to control which internal documents are used, like NDAs, internal counsel memos or state-specific guidelines.
Examples: Confidentiality agreements, IP handling protocols and litigation hold instructions.
Use Case: A manager asks if they can share a prototype with a vendor. The AI accesses the legal department's approved NDA workflow and provides the required preconditions for IP protection.
RAG helps enforce brand consistency and confidentiality. AI writing assistants can generate content only using approved brand tone documents, messaging guidelines or embargoed launch timelines.
Examples: Brand tone guidelines, embargoed campaign details and competitive comparison policies.
Use Case: A content writer asks, 'What's our positioning against competitor X?' Instead of hallucinating risky comparisons, the AI references an internal competitive intelligence deck.
Sales reps often operate on tight timelines and ambiguous inputs. RAG-equipped AI assistants can ground responses in the official sales playbook, quoting rules and commission policies.
Examples: Discount approval thresholds, territory conflict resolution and lead qualification rules.
Use Case: A rep asks, 'Can I offer a 25% discount to a client in EMEA?' The AI checks the discount matrix and responds based on regional approval flows.
Security-related queries are risky when answered with public data. RAG ensures internal policies guide responses.
Examples: Data access controls, employee onboarding/offboarding protocols and acceptable use policy.
Use Case: An employee asks how to report a phishing attempt. The AI retrieves and relays the internal incident response protocol and contact escalation path.
Final Word
In an age where trust, privacy and compliance are business-critical, RAG doesn't just reduce hallucinations—it helps operationalize private knowledge safely across departments. For enterprises betting big on generative AI, grounding outputs in real, governed data isn't optional—it's the foundation of responsible innovation.
Forbes Technology Council is an invitation-only community for world-class CIOs, CTOs and technology executives. Do I qualify?
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

M&T Bank Selects Amperity to Accelerate Customer Data Strategy
M&T Bank Selects Amperity to Accelerate Customer Data Strategy

Yahoo

time17 minutes ago

  • Yahoo

M&T Bank Selects Amperity to Accelerate Customer Data Strategy

Amperity enables M&T Bank to address customer-centricity challenges and deepen relationships with its customers SEATTLE, August 05, 2025--(BUSINESS WIRE)--Amperity, the AI-powered customer data cloud, announced M&T Bank has selected its platform to unify customer data across the bank's operations. This collaboration will enable M&T Bank, a leading full-service financial institution, to build more comprehensive customer profiles and deliver more personalized banking experiences. The relationship between M&T Bank and Amperity represents a significant advancement in the bank's customer data capabilities. With Amperity's platform, M&T Bank can now more effectively consolidate fragmented customer information, streamline data management, and deploy insights across channels. This collaboration enables the bank to gain a more comprehensive understanding of customer needs, preferences and behaviors, ultimately leading to enhanced customer experiences across all touchpoints. "In today's competitive banking landscape, understanding our customers across all business lines is essential," said Kim Nupp, Director of Customer 360 Management at M&T Bank. "Amperity's platform will enable us to further consolidate customer interactions into a unified view of the customer, helping to deliver more relevant, timely communications to meet our customers' financial needs." The implementation of Amperity will support M&T Bank's data-driven strategy by: Better unifying online and offline customer interactions Supporting growth through targeted service offerings Helping to streamline the banking experience across all touchpoints "M&T Bank is setting the pace for customer-centric innovation in modern banking," said Tony Owens, CEO of Amperity. "With Amperity's identity resolution at the core of their data strategy, they've built a modern foundation that gives them a complete view of each customer. This kind of transformation isn't possible without solving identity first - and it's what will separate the leaders from the rest as the industry moves into the next era of personalization." To learn how companies like M&T Bank are partnering with Amperity visit About Amperity Amperity's Customer Data Cloud empowers brands to transform raw customer data into strategic business assets with unprecedented speed and accuracy. Through AI-powered identity resolution, customizable data models, and intelligent automation, Amperity helps technologists eliminate data bottlenecks and accelerate business impact. More than 400 leading brands worldwide, including Virgin Atlantic, Dr Martens, Wyndham Hotels & Resorts and Reckitt rely on Amperity to drive customer insights and revenue growth. Founded in 2016, Amperity operates globally with offices in Seattle, New York City, London, and Melbourne. For more information, visit or follow us on LinkedIn, X, Facebook and Instagram. View source version on Contacts Press Contact Andrea Mochermanpress@ +1 (206) 432-8302 Error while retrieving data Sign in to access your portfolio Error while retrieving data Error while retrieving data Error while retrieving data Error while retrieving data

Merging deep tech and biomedical science expertise, ChronicleBio aims to accelerate breakthroughs for millions living with complex chronic conditions
Merging deep tech and biomedical science expertise, ChronicleBio aims to accelerate breakthroughs for millions living with complex chronic conditions

Yahoo

time17 minutes ago

  • Yahoo

Merging deep tech and biomedical science expertise, ChronicleBio aims to accelerate breakthroughs for millions living with complex chronic conditions

New venture harnesses multi-omic data and AI to develop the largest platform on complex chronic illnesses to drive breakthroughs in POTS, ME/CFS, Long COVID, and other neuroimmune disorders SAN FRANCISCO, Aug. 5, 2025 /PRNewswire/ -- ChronicleBio, a techbio startup focused on radically accelerating insights on neuroimmune disorders, officially launched Tuesday with a bold mission: to build the world's richest, AI-ready data platform for historically overlooked conditions like POTS, ME/CFS, Long COVID, and other complex chronic conditions. ChronicleBio is uniquely led by a founding team that unites elite tech, healthcare, and life sciences: Fidji Simo (Co-founder), current CEO of Instacart and the incoming CEO of OpenAI's applications division, brings deep expertise in consumer-scale technology, product innovation, and health systems transformation. Rohit Gupta (Co-founder & CEO), a biobanking and precision medicine leader who previously oversaw centers at Stanford and UCSF, is the leading scientific and strategic force behind ChronicleBio. Rishi Reddy (Co-founder & Executive Chairman), of Tarsadia Investments, brings an entrepreneurial and venture-backed perspective and a track record in health and life sciences investing. Together, they are tackling a critical gap in biomedical research. ChronicleBio is creating a scalable, ethically driven data ecosystem, integrating clinical, molecular, biospecimen, and patient-reported insights to drive biomarker discovery, patient stratification, and precision therapeutic development. "Entire communities of patients have been ignored for too long, not because their conditions aren't debilitating and prevalent, but because the systems weren't built to understand them," said Fidji Simo, co-founder of ChronicleBio. "ChronicleBio is developing a future where patients can enlist in research, and where the data we collect from them goes directly towards better diagnostics and therapeutics. By building these critical missing datasets and leveraging the massive advances in AI, we aim to accelerate the path to cures." These complex chronic conditions impact nearly 300 million people worldwide, but remain woefully understudied, leaving patients hopeless and without meaningful treatments. It's a disease space desperate for data, and ChronicleBio is offering the scalable solutions to provide it. "Neuroimmune disorders remain some of the most poorly understood in medicine and not for lack of interest, but because the underlying data simply doesn't exist at the scale or structure required for serious breakthroughs," said Rohit Gupta, co-founder and CEO of ChronicleBio. "We're empowering patients and clinicians to get involved in this effort to grow our research-grade biobank and AI-ready data platform so we can close that gap and provide the tools to finally interrogate these diseases with scientific rigor and precision, bringing an end to complex chronic conditions." ChronicleBio partners directly with patients, clinics, and research collaborators to generate longitudinal, multi-modal datasets, including genomics, immune profiling, and real-world outcomes designed for exploration and clinical trial readiness. Its platform is already being used in multiple high-integrity clinics in the U.S. and India, as well as through research collaborations with thousands of patients contributing data and biosamples. "We're creating an ecosystem with global reach, designed not just to generate data, but to fundamentally shift how we discover, develop, and deliver treatments at scale. The opportunity for impact—scientifically, medically, and societally—is enormous," said Rishi Reddy, co-founder and Executive Chairman of ChronicleBio. "We're building the foundational infrastructure for a massive medical need that has been hiding in plain sight." ChronicleBio's founding vision isn't just ambitious. It's personal. Both Simo and Reddy have been diagnosed with at least one of the very conditions the company is focused on researching. Gupta's family also suffers from neuroimmune disorders. The group's world-class expertise is outshined only by its heartfelt determination to rewrite the scientific narrative for the millions of patients overlooked by traditional research and transform lived experience into actionable discovery. ChronicleBio's approach includes: A global biobank platform collecting and connecting data on neuroimmune illnesses A proprietary, AI-ready data engine designed to integrate multi-omic, clinical, and real-world evidence Strategic partnerships with clinics, research consortia, and biotech sponsors looking to unlock hard-to-crack therapeutic areas A decentralized clinical trial network accelerating drug development through real-world access to patients and clinicians For more information, visit or follow ChronicleBio on LinkedIn, X, Instagram, and Facebook. Media Contact:media@ View original content to download multimedia: SOURCE ChronicleBio Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

Jim Acosta says he was ‘honored' to interview AI Parkland victim amid intense backlash
Jim Acosta says he was ‘honored' to interview AI Parkland victim amid intense backlash

Yahoo

time17 minutes ago

  • Yahoo

Jim Acosta says he was ‘honored' to interview AI Parkland victim amid intense backlash

Amid the intense backlash he is facing over an 'interview' with an AI version of a student killed in the 2018 Parkland high school shooting, former CNN anchor Jim Acosta told The Independent he was 'honored' to help the victim's family remember their son. The conversation featured Acosta speaking to an animated image of Joaquin Oliver, who would have been celebrating his 25th birthday on Monday. According to Joaquin's father Manuel, the avatar was based on a real photograph of his deceased son and was trained on general information and knowledge from his writings and social media postings. At the start of the interview, which was posted to Acosta's Substack page, the AI Oliver is asked what happened to him, prompting the model to answer in a somewhat robotic monotone. Additionally, besides the avatar sounding computerized, the motions of its face and lips appear jerky and erratic. 'I appreciate your curiosity,' the avatar said. 'I was taken from this world too soon due to gun violence while at school. It's important to talk about these issues so we can create a safer future for everyone.' The AI model went on to talk about the need for 'a mix of stronger gun control laws, mental health support and community engagement' when asked how to curb gun violence. 'We need to create safe spaces for conversations and connections, making everyone feel seen and heard. It's about building a culture of kindness and understanding,' it added. Acosta also touched on some more casual topics in his conversation with the AI Oliver, such as asking the avatar about movies and the NBA. Following the interview, Acosta brought on Manuel to discuss his decision to use an artificially generated version of his son and whether he planned to continue to do so going forward. 'I understand that this is AI,' Manuel said. 'I don't want anyone to think that I am in, some way, trying to bring my son back. Sadly, I can't, right? I wish I could. However, the technology is out there.' Manuel, who has become an outspoken gun control advocate since his son's death, also noted that the AI version of his son could help elevate the conversation about gun safety. 'We can just raise the voices,' he said. 'Now, Joaquin is going to start having followers… he's going to start uploading videos. This is just the beginning.' Another reason Manuel and his wife Patricia created the AI version of Joaquin, he noted, was just to hear their son's voice again. 'Patricia will spend hours asking questions. Like any other mother, she loves to hear Joaquin saying, 'I love you, mommy.' And that's important,' he told Acosta. 'I really felt like I was speaking with Joaquin. It's just a beautiful thing,' Acosta asserted to Manuel. Acosta, who launched an online Substack show after leaving CNN earlier this year, quickly faced outrage when he posted the interview online. Critics described it as 'one of the most unhinged and grotesque things' while wondering why Acosta would even take part in it. 'This is so insane and evil. It should never be done. I'm speechless,' Reason's Robby Soave tweeted, prompting Acosta to respond with a video of Manuel pushing back on the criticism over the interview. 'Joaquin, known as Guac, should be 25 years old today. His father approached me to do the story… to keep the memory of his son alive,' Acosta declared. 'Today he should be turning 25 years old, and my wife Patricia and myself – we asked our friend Jim Acosta to… have an interview with our son because now, thanks to AI, we can bring him back,' Manuel said in the social media video response. 'It was our idea, it was our plan, and it's still our plan,' he continued. 'We feel that Joaquin has a lot of things to say, and as long as we have an option that allows us to bring that to you and to everyone, we will use it.' Manuel ended the clip by asking critics to 'stop blaming people about where he's coming from, or blaming Jim about what he was able to do,' insisting that if 'the problem you have is with the AI, then you have the wrong problem.' Instead, Manuel concluded, the focus should remain on mass shootings. 'The real problem is that my son was shot eight years ago,' he stated. 'So if you believe that that is not the problem, you are part of the problem.' Asked about the criticism he's faced online over the interview, Acosta told The Independent that he'd essentially reiterate much of what Manuel had to say about the backlash in his video. 'I think I would say this - the family of Joaquin reached out to me to see if I would help them remember their son,' Acosta said. 'My heart goes out to them and I was honored to help them in this moment. I think Joaquin's father makes a good point that if you're bothered more by this than the gun violence that took his son, then there is something truly wrong with our society.' Meanwhile, Acosta's conversation isn't the first time that any of the victims of the Parkland shooting had been brought back via AI. Oliver, along with the voices of several other students and staff who were killed in the massacre, was used in a robocalling campaign for gun control last year called The Shotline. 'Six years ago, I was a senior at Parkland. Many students and teachers were murdered on Valentine's Day that year by a person using an AR-15 assault rifle,' the voice of Oliver stated in the call that was sent to lawmakers. 'It's been six years, and you've done nothing. Not a thing to stop all the shootings that have continued to happen since.' The voice added: 'I'm back today because my parents used AI to re-create my voice to call you. How many calls will it take for you to care? How many dead voices will you hear before you finally listen?' Additionally, AI avatars have been used to simulate victim impact statements in criminal cases. In May, for instance, an AI version of a man who was killed in a road rage shooting addressed the alleged killer in court. 'I believe in forgiveness, and a God who forgives. I always have and I still do,' the animated model of the victim said. The judge in the case appeared moved by the video, which was produced by the victim's family. 'I loved that AI, thank you for that. As angry as you are, as justifiably angry as the family is, I heard the forgiveness,' the judge said. 'I feel that that was genuine.'

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store