logo
Opinion: AI 'hallucinations' are a growing problem for the legal profession

Opinion: AI 'hallucinations' are a growing problem for the legal profession

The Star26-05-2025

You've probably heard the one about the product that blows up in its creators' faces when they're trying to demonstrate how great it is.
Here's a ripped-from-the-headlines yarn about what happened when a big law firm used an AI bot product developed by Anthropic, its client, to help write an expert's testimony defending the client.
It didn't go well. Anthropic's chatbot, Claude, got the title and authors of one paper cited in the expert's statement wrong, and injected wording errors elsewhere. The errors were incorporated in the statement when it was filed in court in April.
Those errors were enough to prompt the plaintiffs suing Anthropic – music publishers who allege that the AI firm is infringing their copyrights by feeding lyrics into Claude to "train" the bot – to ask the federal magistrate overseeing the case to throw out the expert's testimony in its entirety.
It may also become a black eye for the big law firm Latham & Watkins, which represents Anthropic and submitted the errant declaration.
Latham argues that the errors were inconsequential, amounting to an "honest citation mistake and not a fabrication." The firm's failure to notice the errors before the statement was filed is "an embarrassing and unintentional mistake," but it shouldn't be exploited to invalidate the expert's opinion, the firm told Magistrate Judge Susan van Keulen of San Jose, who is managing the pretrial phase of the lawsuit. The plaintiffs, however, say the errors "fatally undermine the reliability" of the expert's declaration.
At a May 13 hearing conducted by phone, Van Keulen herself expressed doubts.
"There is a world of difference between a missed citation and a hallucination generated by AI, and everyone on this call knows that," she said, according to a transcript of the hearing cited by the plaintiffs. (Van Keulen hasn't yet ruled on whether to keep the expert's declaration in the record or whether to hit the law firm with sanctions.)
That's the issue confronting judges as courthouse filings peppered with serious errors and even outright fabrications – what AI experts term "hallucinations" – continue to be submitted in lawsuits.
A roster compiled by the French lawyer and data expert Damien Charlotin now numbers 99 cases from federal courts in two dozen states as well as from courts in Europe, Israel, Australia, Canada and South Africa.
That's almost certainly an undercount, Charlotin says. The number of cases in which AI-generated errors have gone undetected is incalculable, he says: "I can only cover cases where people got caught."
In nearly half the cases, the guilty parties are pro-se litigants – that is, people pursuing a case without a lawyer. Those litigants generally have been treated leniently by judges who recognize their inexperience; they seldom are fined, though their cases may be dismissed.
In most of the cases, however, the responsible parties were lawyers. Amazingly, in some 30 cases involving lawyers the AI-generated errors were discovered or were in documents filed as recently as this year, long after the tendency of AI bots to "hallucinate" became evident. That suggests that the problem is getting worse, not better.
"I can't believe people haven't yet cottoned to the thought that AI-generated material is full of errors and fabrications, and therefore every citation in a filing needs to be confirmed," says UCLA law professor Eugene Volokh.
Judges have been making it clear that they have had it up to here with fabricated quotes, incorrect references to legal decisions and citations to nonexistent precedents generated by AI bots. Submitting a brief or other document without certifying the truth of its factual assertions, including citations to other cases or court decisions, is a violation of Rule 11 of the Federal Rules of Civil Procedure, which renders lawyers vulnerable to monetary sanctions or disciplinary actions.
Some courts have issued standing orders that the use of AI at any point in the preparation of a filing must be disclosed, along with a certification that every reference in the document has been verified. At least one federal judicial district has forbidden almost any use of AI.
The proliferation of faulty references in court filings also points to the most serious problem with the spread of AI bots into our daily lives: They can't be trusted. Long ago it became evident that when even the most sophisticated AI systems are flummoxed by a question or task, they fill in the blanks in their own knowledge by making things up.
As other fields use AI bots to perform important tasks, the consequences can be dire. Many medical patients "can be led astray by hallucinations," a team of Stanford researchers wrote last year. Even the most advanced bots, they found, couldn't back up their medical assertions with solid sources 30% of the time.
It's fair to say that workers in almost any occupation can fall victim to weariness or inattention; but attorneys often deal with disputes with thousands or millions of dollars at stake, and they're expected to be especially rigorous about fact-checking formal submissions.
Some legal experts say there's a legitimate role for AI in the law – even to make decisions customarily left to judges. But lawyers can hardly be unaware of the pitfalls for their own profession in failing to monitor bots' outputs.
The very first sanctions case on Charlotin's list originated in June 2023 –Mata vs. Avianca, a New York personal injury case that resulted in a US$5,000 (RM 21,273) penalty for two lawyers who prepared and submitted a legal brief that was largely the product of the ChatGPT chatbot. The brief cited at least nine court decisions that were soon exposed as nonexistent. The case was widely publicized coast to coast.
One would think fiascos like this would cure lawyers of their reliance on artificial intelligence chatbots to do their work for them. One would be wrong. Charlotin believes that the superficially authentic tone of AI bots' output may encourage overworked or inattentive lawyers to accept bogus citations without double-checking.
"AI is very good at looking good," he told me. Legal citations follow a standardised format, so "they're easy to mimic in fake citations," he says.
It may also be true that the sanctions in the earliest cases, which generally amounted to no more than a few thousand dollars, were insufficient to capture the bar's attention. But Volokh believes the financial consequences of filing bogus citations should pale next to the nonmonetary consequences.
"The main sanctions to each lawyer are the humiliation in front of the judge, in front of the client, in front of supervisors or partners..., possibly in front of opposing counsel, and, if the case hits the news, in front of prospective future clients, other lawyers, etc," he told me. "Bad for business and bad for the ego."
Charlotin's dataset makes for amusing reading – if mortifying for the lawyers involved. It's peopled by lawyers who appear to be totally oblivious to the technological world they live in.
The lawyer who prepared the hallucinatory ChatGPT filing in the Avianca case, Steven A. Schwartz, later testified that he was "operating under the false perception that this website could not possibly be fabricating cases on its own." When he began to suspect that the cases couldn't be found in legal databases because they were fake, he sought reassurance – from ChatGPT!
"Is Varghese a real case?" he texted the bot. Yes, it's "a real case," the bot replied. Schwartz didn't respond to my request for comment.
Other cases underscore the perils of placing one's trust in AI.
For example, last year Keith Ellison, the attorney general of Minnesota, hired Jeff Hancock, a communications professor at Stanford, to provide an expert opinion on the danger of AI-faked material in politics. Ellison was defending a state law that made the distribution of such material in political campaigns a crime; the law was challenged in a lawsuit as an infringement of free speech.
Hancock, a well-respected expert in the social harms of AI-generated deepfakes – photos, videos and recordings that seem to be the real thing but are convincingly fabricated – submitted a declaration that Ellison duly filed in court.
But Hancock's declaration included three hallucinated references apparently generated by ChatGPT, the AI bot he had consulted while writing it. One attributed to bogus authors an article he himself had written, but he didn't catch the mistake until it was pointed out by the plaintiffs.
Laura M. Provinzino, the federal judge in the case, was struck by what she called "the irony" of the episode: "Professor Hancock, a credentialed expert on the dangers of AI and misinformation, has fallen victim to the siren call of relying too heavily on AI – in a case that revolves around the dangers of AI, no less."
That provoked her to anger. Hancock's fake citations, she wrote, "shatters his credibility with this Court." Noting that he had attested to the veracity of his declaration under penalty of perjury, she threw out his entire expert declaration and refused to allow Ellison to file a corrected version.
In a mea culpa statement to the court, Hancock explained that the errors might have crept into his declaration when he cut-and-pasted a note to himself. But he maintained that the points he made in his declaration were valid nevertheless. He didn't respond to my request for further comment.
On Feb 6, Michael R. Wilner, a former federal magistrate serving as a special master in a California federal case against State Farm Insurance, hit the two law firms representing the plaintiff with US$31,000 (RM 131,898) in sanctions for submitting a brief with "numerous false, inaccurate, and misleading legal citations and quotations."
In that case, a lawyer had prepared an outline of the brief for the associates assigned to write it. He had used an AI bot to help write the outline, but didn't warn the associates of the bot's role. Consequently, they treated the citations in the outline as genuine and didn't bother to double-check them.
As it happened, Wilner noted,"approximately nine of the 27 legal citations in the ten-page brief were incorrect in some way." He chose not to sanction the individual lawyers: "This was a collective debacle," he wrote.
Wilner added that when he read the brief, the citations almost persuaded him that the plaintiff's case was sound — until he looked up the cases and discovered they were bogus. "That's scary," he wrote. His monetary sanction for misusing AI appears to be the largest in a US court ... so far. – Los Angeles Times/Tribune News Service

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

VenHub Launches 24/7 AI-Powered Smart Store at Metro Transit Center at LAX, Leading the Next Era of Autonomous Retail in Travel and Transportation
VenHub Launches 24/7 AI-Powered Smart Store at Metro Transit Center at LAX, Leading the Next Era of Autonomous Retail in Travel and Transportation

Malay Mail

time4 hours ago

  • Malay Mail

VenHub Launches 24/7 AI-Powered Smart Store at Metro Transit Center at LAX, Leading the Next Era of Autonomous Retail in Travel and Transportation

[email protected] Los Angeles, California - Newsfile Corp. - June 6, 2025 - VenHub Global, Inc. ("VenHub" or the "Company"), a leader in fully autonomous, AI-powered retail, unveiled its flagship Smart Store at the LAX/Metro Transit Center at Los Angeles International Airport. The Smart Store brings secure, frictionless, and fully automated retail convenience to millions of travelers and transit riders moving through Los Angeles each in anticipation of the 2026 FIFA World Cup and 2028 Olympic and Paralympic Games, the installation places VenHub at the forefront of smart retail infrastructure as Los Angeles prepares to welcome the world."Retail should work for people, not the other way around," said Shahan Ohanessian, Founder & CEO of VenHub. "Launching our Smart Store at the Metro Transit Center at LAX represents much more than a store opening. It's about giving people access to what they need, exactly when they need it, with safety and simplicity built into every interaction. Whether you're a traveler heading to the terminal, a commuter catching a connection, or a parent needing essentials after hours, VenHub is ready to serve. Always open, always secure, and always designed around the customer."Ohanessian added, "Being first is not just a milestone, it's a responsibility. As leaders in unattended retail, it's our obligation to keep pushing the edge of what's possible. With over $300 million in Smart Store pre-orders across the U.S. and growing demand from enterprise partners around the world, we are scaling our production capacity to meet this extraordinary is a fully autonomous, AI-powered Smart Store that can be installed in under seven days and operates 24/7 without staff. Using robotics, machine vision, and IoT, each unit delivers a safe, secure, and lightning-fast shopping experience - completing transactions in as little as 90 seconds. Whether fixed or mobile, VenHub's Smart Stores intelligently adapt inventory and merchandising based on location, customer behavior, and time-of-day demand, setting a new standard for access, convenience, and future-ready view an enhanced version of this graphic, please visit:This launch marks VenHub's first deployment in partnership with LA Metro, with additional locations already being planned across Southern California. These Smart Stores will serve transit-connected communities across the region, offering a retail experience that is fast, secure, and completely contactless, available 24 hours a day, 365 days a year."Metro is committed to enhancing every step of our customers' journeys, and that includes the moments they spend in our stations," said Jennifer Vides, Chief Customer Experience Officer. "Partnering with VenHub at the LAX/Metro Transit Center station reflects our focus on innovation and our Board's directive to explore retail and other station-based amenities that improve the overall rider experience. VenHub's 24/7, AI-powered smart retail technology is a forward-thinking solution that adds convenience, engagement, and value to our customers as they travel the Metro system."This partnership underscores the power of combining public infrastructure with private innovation to elevate everyday convenience for millions of June 6, VenHub welcomed the public, government officials, and partners to its grand opening at the Metro Transit Center at LAX. Attendees interacted with the Smart Store, placed live orders, explored the robotics system in action, and met the team behind the view this video? Visit:VenHub is building the retail infrastructure of the future. With a bold vision to modernize access to goods and services, VenHub delivers scalable, autonomous Smart Store technology across the U.S. and beyond. From major metro regions to rural towns, fixed units to mobile deployments, VenHub brings automation, intelligence, and convenience wherever people need AI-powered Smart Stores operate 24/7 with no on-site employees, adapting dynamically to customer behavior, local conditions, and operator settings. With a focus on safety, speed, and scalability, VenHub is setting the new global standard for retail that never sleeps. Following its Southern California expansion, VenHub is preparing deployments for transit hubs and commercial sites in major cities across North America, the Middle East, and learn more, visit:VenHub, a division of VenHub Global, Inc. ("VenHub" or the "Company"), may make forward-looking statements regarding future events or the future financial perrmance of the Company in press releases, presentations, conference calls, or other"believes," "expects," "anticipates," "foresees," "forecasts," "estimates," "intends," "plans," "targets," or other words conveying future outcomes or forward-looking statements involve certain risks, uncertainties, and assumptions that are difficult to predict and beyond the Company's control. Actual results could differ materially from those expressed or implied by the forward-looking statements as a result of various factors, including but not limited to changes in general economic conditions, the Company's ability to execute its business strategy, competitive pressures, unanticipated manufacturing or supply chain issues, compliance with regulatory requirements, and other risks detailed in the Company's public filings with the Securities and Exchange in these forward-looking statements should be regarded as a representation by VenHub or its management that the Company's objectives or plans will be achieved. VenHub undertakes no obligation to publicly update or revise any forward-looking statements, whether as a result of new information, future events, or otherwise, except as required by applicable Barry, Alliance Advisors604-997-0965 The issuer is solely responsible for the content of this announcement. About VenHub Global Inc.

Yageo says it will protect technology if Shibaura purchase succeeds
Yageo says it will protect technology if Shibaura purchase succeeds

The Star

time6 hours ago

  • The Star

Yageo says it will protect technology if Shibaura purchase succeeds

Yageo's founder and Chairman Pierre Chen speaks at the company's headquarters in New Taipei City, Taiwan June 7, 2025. REUTERS/Wen-Yee Lee TAIPEI/TOKYO (Reuters) -Taiwan's Yageo said it will implement strict controls to prevent technology from leaking if it succeeds in acquiring Japan's Shibaura Electronics, responding to concerns in Japan over what the deal could mean for national security. Chairman Pierre Chen told reporters in Taipei on Saturday that the company will meet with Shibaura in mid-June in Tokyo to discuss potential cooperation. Yageo, the world's largest maker of chip resistors, launched an unsolicited tender offer for Shibaura in February, seeking full control of the Japanese firm, which specialises in thermistor technology. Yageo offered to buy Shibaura at 4,300 yen per share, valuing the company at more than 65 billion yen ($450 million). Spurning Yageo's overture, Shibaura tapped Japanese components supplier Minebea Mitsumi as a white knight. Minebea and Yageo entered a bidding war, with the latter now offering 6,200 yen. The stock closed at 6,100 yen on Friday. "Our strategy is to inject resources and strengthen R&D for advanced technologies. We're also preparing to make larger investments to expand their facilities in Japan," Chen said. Asked about Japan's national security concerns, he said: "We will implement strict controls to ensure technology does not leak." Unsolicited takeovers were once rare in Japan, where companies often mounted elaborate defences. The Japanese industry ministry's M&A guidelines in 2023 cracked down on what it considered excessive defence tactics, de-stigmatising unsolicited buyouts and leading some of such deals to succeed. Chen said that negotiations with Japan's Ministry of Economy, Trade and Industry had been going smoothly. He said that if Yageo acquires Shibaura, the deal would address a gap in its portfolio of thermistors, making Yageo's offerings more complete for global customers and helping Shibaura expand its access to markets outside of Japan. Yageo said it aims to ease the burden of managing smaller component suppliers for its major clients, including Apple and Nvidia, by offering more comprehensive product portfolios and solutions. Yageo is also the world's number three manufacturers of multilayer ceramic capacitors and provides key components used in Apple's iPhones, Nvidia's AI servers, and Tesla's electric vehicles. ($1 = 144.8500 yen) (Reporting by Wen-Yee Lee in Taipei and Makiko Yamazaki in Tokyo; Editing by William Mallard)

Human coders are still better than AI, says this expert developer
Human coders are still better than AI, says this expert developer

The Star

time10 hours ago

  • The Star

Human coders are still better than AI, says this expert developer

Your team members may be tempted to rely on AI to help them write code for your company, either for cost or speed rationales or because they lack particular expertise. But you should be wary. — Pixabay In the complex 'will AI steal my job?' debate, software developers are among the workers most immediately at risk from powerful AI tools. It's certainly looking like the tech sector wants to reduce the number of humans working those jobs. Bold statements from the likes of Meta's Mark Zuckerberg and Anthropic's Dario Amodei support this since both of them say AI is already able to take over some code-writing roles. But a new blog post from a prominent coding expert strongly disputes their arguments, and supports some AI critics' position that AI really can't code. Salvatore Sanfilippo, an Italian developer who created Redis (an online database which calls itself the 'world's fastest data platform' and is beloved by coders building real-time apps), published a blog post this week, provocatively titled 'Human coders are still better than LLMs.' His title refers to large language model systems that power AI chatbots like OpenAI's ChatGPT and Anthropic's Claude. Sanfilippo said he's 'not anti-AI' and actually does 'use LLMs routinely,' and explained some specific interactions he'd had with Google's Gemini AI about writing code. These left him convinced that AIs are 'incredibly behind human intelligence,' so he wanted to make a point about it. The billions invested in the technology and the potential upending of the workforce mean it's 'impossible to have balanced conversations' on the matter, he wrote. Sanfilippo blogged that he was trying to 'fix a complicated bug' in Redis's systems. He made an attempt himself, and then asked Gemini, 'hey, what we can do here? Is there a super fast way' to implement his fix? Then, using detailed examples of the kind of software he was working with and the problem he was trying to fix, he blogged about the back-and-forth dialogue he had with Gemini as he tried to coax it toward an acceptable answer. After numerous interactions where the AI couldn't improve on his idea or really help much, he said he 'asked Gemini to do an analysis of (his last idea, and it was finally happy.' We can ignore the detailed code itself and just concentrate on Sanfilippo's final paragraph. 'All this to say: I just finished the analysis and stopped to write this blog post, I'm not sure if I'm going to use this system (but likely yes), but, the creativity of humans still have an edge, we are capable of really thinking out of the box, envisioning strange and imprecise solutions that can work better than others,' he wrote. 'This is something that is extremely hard for LLMs.' Gemini was useful, he admitted, to simply 'verify' his bug-fix ideas, but it couldn't outperform him and actually solve the problem itself. This stance from an expert coder goes up against some other pro-AI statements. Zuckerberg has said he plans to fire mid-level coders from Meta to save money, employing AI instead. In March, Amodei hit the headlines when he boldly predicted that all code would be written by AIs inside a year. Meanwhile, on the flip side, a February report from Microsoft warned that young coders coming out of college were already so reliant on AI to help them that they failed to understand the hard computer science behind the systems they were working on –something that may trip them up if they encountered a complex issue like Sanfilippo's bug. Commenters on a piece talking about Sanfilippo's blog post on coding news site Hacker News broadly agreed with his argument. One commenter likened the issue to a popular meme about social media: 'You know that saying that the best way to get an answer online is to post a wrong answer? That's what LLMs do for me.' Another writer noted that AIs were useful because even though they give pretty terrible coding advice, 'It still saves me time, because even 50 percent accuracy is still half that I don't have to write myself.' Lastly, another coder pointed out a very human benefit from using AI: 'I have ADHD and starting is the hardest part for me. With an LLM it gets me from 0 to 20% (or more) and I can nail it for the rest. It's way less stressful for me to start now.' Why should you care about this? At first glance, it looks like a very inside-baseball discussion about specific coding issues. You should care because your team members may be tempted to rely on AI to help them write code for your company, either for cost or speed rationales or because they lack particular expertise. But you should be wary. AIs are known to be unreliable, and Sanfilippo's argument, supported by other coders' comments, point out that AI really isn't capable of certain key coding tasks. For now, at least, coders' jobs may be safe… and if your team does use AI to code, they should double and triple check the AI's advice before implementing it in your IT system. – Inc./Tribune News Service

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store