logo
Can AI fact-check its own lies?

Can AI fact-check its own lies?

Fast Company17 hours ago

As AI car crashes go, the recent publishing of a hallucinated book list in the Chicago Sun-Times quickly became a multi-vehicle pile-up. After a writer used AI to create a list of summer reads, the majority of which were made-up titles, the resulting article sailed through lax editorial review at the Sun-Times (and at least one other newspaper) and ended up being distributed to thousands of subscribers. The CEO eventually published a lengthy apology.
The most obvious takeaway from the incident is that it was a badly needed wake-up call about what can happen when AI gets too embedded in our information ecosystem. But CEO Melissa Bell resisted the instinct to simply blame AI, instead putting responsibility on the humans who use it and those who are entrusted with safeguarding readers from its weaknesses. She even included herself as one of those people, explaining how she had approved the publishing of special inserts like the one the list appeared in, assuming at the time there would be adequate editorial review (there wasn't).
The company has made changes to patch this particular hole, but the affair exposes a gap in the media landscape that is poised to get worse: as the presence of AI-generated content—authorized or not—increases in the world, the need for editorial safeguards also increases. And given the state of the media industry and its continual push to do 'more with less,' it's unlikely that human labor will scale up to meet the challenge. The conclusion: AI will need to fact-check AI.
Fact-checking the fact-checker
I know, it sounds like a horrible idea, somewhere between letting the fox watch the henhouse or sending Imperial Stormtroopers to keep the peace on Endor. But AI fact-checking isn't a new idea: In fact, when Google Gemini first debuted (then called Bard), it shipped with an optional fact-check step if you wanted it to double-check anything it was telling you. Eventually, this kind of step simply became integrated into how AI search engines work, broadly making their results better, though still far from perfect.
Newsrooms, of course, set a higher bar, and they should. Operating a news site comes with the responsibility to ensure the stories you're telling are true, and for most sites the shrugging disclaimer of 'AI can make mistakes,' while good enough for ChatGPT, doesn't cut it. That's why for most, if not all, AI-generated outputs (such as ESPN's AI-written sports recaps), humans check the work.
As AI writing proliferates, though, the inevitable question is: Can AI do that job? Put aside the weirdness for a minute and see it as math, the key number being how often it gets things wrong. If an AI fact-checker can reduce the number of errors by as much if not more than a human, shouldn't it do that job?
If you've never used AI to fact-check something, the recently launched service isitcap.com offers a glimpse at where the technology stands. It doesn't just label claims as true or false—it evaluates the article holistically, weighing context, credibility, and bias. It even compares multiple AI search engines to cross-check itself.
You can easily imagine a newsroom workflow that applies an AI fact-checker similarly, sending its analysis back to the writer, highlighting the bits that need shoring up. And if the writer happens to be a machine, revisions could be done lightning fast, and at scale. Stories could go back and forth until they reach a certain accuracy threshold, with anything that falls short held for human review.
All this makes sense in theory, and it could even be applied to what news orgs are doing currently with AI summaries. Nieman Lab has an excellent write-up on how The Wall Street Journal, Yahoo News, and Bloomberg all use AI to generate bullet points or top-line takeaways for their journalism. For both Yahoo and the Journal, there's some level of human review on the summaries (for Bloomberg, it's unclear from the article). These organizations are already on the edge of what's acceptable—balancing speed and scale with credibility. One mistake in a summary might not seem like much, but when trust is already fraying, it's enough to shake confidence in the entire approach.
Human review helps ensure accuracy, of course, but also requires more human labor—something in short supply in newsrooms that don't have a national footprint. AI fact-checking could give smaller outlets more options with respect to public-facing AI content. Similarly, Politico's union recently criticized the publication's AI-written reports for subscribers based on the work of its journalists, because of occasional inaccuracies. A fact-checking layer might prevent at least some embarrassing mistakes, like attributing political stances to groups that don't exist.
The AI trust problem that won't go away
Using AI to fight AI hallucination might make mathematical sense if it can prevent serious errors, but there's another problem that stems from relying even more on machines, and it's not just a metallic flavor of irony. The use of AI in media already has a trust problem. The Sun-Times ' phantom book list is far from the first AI content scandal, and it certainly won't be the last. Some publications are even adopting anti-AI policies, forbidding its use for virtually anything.
Because of AI's well-documented problems, public tolerance for machine error is lower than for human error. Similarly, if a self-driving car gets into an accident, the scrutiny is obviously much greater than if the car was driven by a person. You might call this the automation fallout bias, and whether you think it's fair or not, it's undoubtedly true. A single high-profile hallucination that slips through the cracks could derail adoption, even if it might be statistically rare.
Add to that what would probably be painful compute costs for multiple layers of AI writing and fact-checking, not to mention the increased carbon footprint. All to improve AI-generated text—which, let's be clear, is not the investigative, source-driven journalism that still requires human rigor and judgment. Yes, we'd be lightening the cognitive load for editors, but would it be worth the cost?
Despite all these barriers, it seems inevitable that we will use AI to check AI outputs. All indications point to hallucinations being inherent to generative technology. In fact, newer 'thinking' models appear to hallucinate even more than their less sophisticated predecessors. If done right, AI fact-checking would be more than a newsroom tool, becoming part of the infrastructure for the web. The question is whether we can build it to earn trust, not just automate it.
The amount of AI content in the world can only increase, and we're going to need systems that can scale to keep up. AI fact-checkers can be part of that solution, but only if we manage—and accept— their potential to make errors themselves. We may not yet trust AI to tell the truth, but at least it can catch itself in a lie.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

ISS issues, technical setbacks delay historic Axiom-4 mission
ISS issues, technical setbacks delay historic Axiom-4 mission

Yahoo

time33 minutes ago

  • Yahoo

ISS issues, technical setbacks delay historic Axiom-4 mission

The Brief Axiom-4's mission remains delayed due to weather, equipment issues, and an ongoing ISS leak. Experts urge patience and caution as global partners assess safety and future station plans. The uncertainty underscores growing challenges in the transition to commercial spaceflight. CAPE CANAVERAL, Fla. - A series of technical and international setbacks are continuing to delay the highly anticipated Axiom-4 mission, originally expected to mark a major milestone in private spaceflight. What we know The Axiom-4 mission, a significant private spaceflight effort, has faced repeated delays. Initial postponements stemmed from weather conditions and a liquid oxygen leak. Now, a leak in a Russian module aboard the International Space Station (ISS) is adding to the holdup, prompting further safety reviews and coordination among international partners. What we don't know There is no clear timeline for when Axiom-4 will launch, nor a definitive assessment of how long the Russian module repair will take. The potential impact of repeated delays on the crew, mission goals, and future private spaceflight schedules remains uncertain. It's also unclear how soon a viable replacement for the ISS could be operational if early deorbiting plans gain traction. The backstory The ISS has served as a hub for international space collaboration since the early 2000s. Axiom-4 is part of the broader trend of commercializing low-Earth orbit, with companies like Axiom Space aiming to build private stations. The delays highlight both the promise and complexity of transitioning from government-led to commercial space operations. What they're saying Russia's space agency, Roscosmos, is currently addressing a leak on one of its ISS modules, raising concerns about the station's readiness to accommodate the incoming Axiom-4 crew. "This isn't something that they do on a normal, regular basis," said Don Platt, of the Florida Institute of Technology. "They really need to sit back and first make sure that all sides are happy with a plan to move forward, and then it seems to be working and then take a little bit of time after they do a fix to, to see how well that actually works." The delays come as the future of the ISS itself becomes increasingly uncertain. The orbiting laboratory has been in service for nearly 30 years, and some — including SpaceX CEO Elon Musk — re calling for it to be deorbited as early as 2026, four years ahead of NASA's current timeline. But experts caution against acting too quickly. "We've got to have a replacement before we willy nilly decide. It's too old. Let's junk it," Platt added. "You know, it's kind of like, again, with your car, you got to have another car to go to work the next day than before." Companies like Axiom Space are developing next-generation commercial space stations, but those projects are still in the early stages of construction and testing. For now, Axiom-4's launch remains in limbo, as engineers and international partners work to resolve both Earth-bound and orbital obstacles. STAY CONNECTED WITH FOX 35 ORLANDO: Download the FOX Local app for breaking news alerts, the latest news headlines Download the FOX 35 Storm Team Weather app for weather alerts & radar Sign up for FOX 35's daily newsletter for the latest morning headlines FOX Local:Stream FOX 35 newscasts, FOX 35 News+, Central Florida Eats on your smart TV The Source This story was written based on information shared by Axiom Space, NASA, and Don Platt, of the Florida Institute of Technology.

As disinformation and hate thrive online, YouTube quietly changed how it moderates content
As disinformation and hate thrive online, YouTube quietly changed how it moderates content

Yahoo

timean hour ago

  • Yahoo

As disinformation and hate thrive online, YouTube quietly changed how it moderates content

YouTube, the world's largest video platform, appears to have changed its moderation policies to allow more content that violates its own rules to remain online. The change happened quietly in December, according to The New York Times, which reviewed training documents for moderators indicating that a video could stay online if the offending material did not account for more than 50 per cent of the video's duration — that's double what it was prior to the new guidelines. YouTube, which sees 20 million videos uploaded a day, says it updates its guidance regularly and that it has a "long-standing practice of applying exceptions" when it suits the public interest or when something is presented in an educational, documentary, scientific or artistic context. "These exceptions apply to a small fraction of the videos on YouTube, but are vital for ensuring important content remains available," YouTube spokesperson Nicole Bell said in a statement to CBC News this week. But in a time when social media platforms are awash with misinformation and conspiracy theories, there are concerns that YouTube is only opening the door for more people to spread problematic or harmful content — and to make a profit doing so. YouTube isn't alone. Meta, which owns Facebook and Instagram, dialled back its content moderation earlier this year, and Elon Musk sacked Twitter's moderators when he purchased the platform in 2022 and rebranded it as X. "We're seeing a race to the bottom now," Imran Ahmed, CEO for the U.S.-based Center for Countering Digital Hate, told CBC News. "What we're going to see is a growth in the economy around hate and disinformation." WATCH | Experts warn Meta's moderation move will likely increase misinformation: YouTube's goal is "to protect free expression," Brooks said in her statement, explaining that easing its community guidelines "reflect the new types of content" on the platform. For example, she said, a long-form podcast containing one short clip of violence may no longer need to be removed. The Times reported Monday that examples presented to YouTube staff included a video in which someone used a derogatory term for transgender people during a discussion about hearings for U.S. President Donald Trump's cabinet appointees, and another that shared false information about COVID-19 vaccines but that did not outright tell people not to get vaccinated. A platform like YouTube does have to make some "genuinely very difficult decisions" when moderating content, says Matt Hatfield, executive director of the Canadian digital rights group OpenMedia. LISTEN | How Canada has come to play an outsized role in far-right misinformation: He believes platforms do take the issue seriously, but he says there's a balance between removing harmful or illegal content, such as child abuse material or clear incitements to violence, and allowing content to stay online, even if it's offensive to many or contains some false information. The problem, he says, is that social media platforms also "create environments that encourage some bad behaviour" among creators, who like to walk the line of what's acceptable. "The core model of these platforms is to keep you clicking, keep you watching, get you to try a video from someone you've never experienced before and then stick with that person." And that's what concerns Ahmed. He says these companies put profits over online safety and that they don't face consequences because there are no regulations forcing them to limit what can be posted on their platforms. He believes YouTube's relaxed policies will only encourage more people to exploit them. In a recent transparency report, YouTube said it had removed nearly 2.9 million channels containing more than 47 million videos for community guideline violations in the first quarter — that came after the reported policy change. The overwhelming majority of those, 81.8 per cent, were considered spam, but other reasons included violence, hateful or abusive material and child safety. LISTEN | Why you're being tormented by ads algorithms and AI slop: Hatfield says there is a public interest in having harmful content like that removed, but that doesn't mean all controversial or offensive content must go. However, he says YouTube does make mistakes in content moderation, explaining that it judges individual videos in a sort of "vacuum" without considering how each piece of content fits into a broader context. "Some content can't really be fairly interpreted in that way." Ahmed says companies should be held accountable for the content on their platforms through government regulation. He pointed to Canada's controversial but now-scuttled Online Harms Act, also known as Bill C-63, as an example. It proposed heavier sentences, new regulatory bodies and changes to a number of laws to tackle online abuse. The bill died when former prime minister Justin Trudeau announced his resignation and prorogued Parliament back in January. Ahmed says he hopes the new government under Prime Minister Mark Carney will enact similar legislation. Hatfield says he liked parts of that act, but his group ultimately opposed it after it tacked on some other changes to the Criminal Code and Human Rights Act that he says were unrelated to the platforms. He says groups like OpenMedia would have liked to see a strategy addressing business models that encourage users to post and profit off of "lawful but awful" content. "We're not going to have a hate-free internet," he said. "We can have an internet that makes it less profitable to spread certain types of hate and misinformation." WATCH | How people can become more discerning news consumers:

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store