logo
'Pretty damn average': Google's AI Overviews underwhelm

'Pretty damn average': Google's AI Overviews underwhelm

RNZ News2 days ago

Photo:
JAAP ARRIENS
Most searches online are done using Google. Traditionally, they've returned long lists of links to websites carrying relevant information.
Depending on the topic, there can be thousands of entries to pick from or scroll through.
Last year Google started incorporating its
Gemini AI tech into its searches
.
Google's Overviews now inserts Google's own summary of what it's scraped from the internet ahead of the usual list of links to sources in many searches.
Some sources say
Google's now working towards
replacing the lists of links with its own AI-driven search summaries.
RNZ's Kathryn Ryan's not a fan.
"Pretty damn average I have to say, for the most part," she
said on Nine to Noon last Monday
during a chat about AI upending the business of digital marketing.
But Kathryn Ryan is not the only one underwhelmed by Google's Overviews. Recently, online tech writers discovered you can trick it into thinking that made up sayings are actually idioms in common usage that are meaningful.
The
Sydney Morning Herald
's puzzle compiler David Astle - under the headline
'Idiom or Idiot?'
reckoned Google's AI wasn't about to take his job making cryptic crosswords anytime soon.
"There is a strange bit of human psychology which says that we expect a very high bar from machines in a way that we don't from humans," the BBC's head of technology forecasting Laura Ellis
told Mediawatch
last month.
"But if you've got a machine making a mistake, where does that accountability fall? We've just not tested this out yet."
UK Sky News deputy political editor Sam Coates tried to make ChatGPT accountable after it made up an entire episode of his own politics podcast when he used it to help archive transcripts of it recently.
"AI had told a lie that it had got the transcript. And rather than back down it invented an entire fake episode without flagging that it's fake."
When challenged on this, the technology insisted Coates had created the episode himself.
When ChatGPT can't find an answer or the right data to draw on, it can 'hallucinate' or just make up a misleading response.
"Chat GTP is gaslighting me. No such thing exists. It's all a complete fake," Coates spluttered.
After turning ChatGPT off and on again in 'conversation mode', it did eventually own up.
"It said; 'Look, you're absolutely right to challenge that. I can't remember the exact time that you uploaded.' And then: 'What I can confirm is that I did it and you're holding me to account,'" Coates told viewers.
He went on to challenge ChatGPT about its hallucinations getting worse.
"The technology is always improving, and newer versions tend to do a better job at staying accurate," ChatGPT replied.
But Coates - armed with data that suggested the opposite - asked ChatGPT for specific stats.
The response: "According to recent internal tests from OpenAI, the newer models have shown higher hallucination rates. For instance, the model known as o3 had about a 33 percent hallucination rate, while the 04 mini model had around 48 percent."
"I get where you're coming from, and I'm sorry for the mixed messages. The performance of these models can vary."
When Coates aired his experience as a warning for journalists, some reacted with alarm.
"The hallucination rate of advanced models... is increasing. As journos, we really should avoid it," said
Sunday Times
writer and former BBC diplomatic editor Mark Urban.
But some tech experts accused Coates of misunderstanding and misusing the technology.
"The issues Sam runs into here will be familiar to experienced users, but it illustrates how weird and alien Large Language Model (LLM) behaviour can seem for the wider public," said Cambridge University AI ethicist Henry Shevlin.
"We need to communicate that these are generative simulators rather than conventional programmes," he added.
Others were less accommodating on social media.
"All I am seeing here is somebody working in the media who believes they understand how technology works - but [he] doesn't - and highlighting the dangers of someone insufficiently trained in technology trying to use it."
"It's like Joey from
Friends
using the thesaurus function on Word."
Mark Honeychurch is a programmer and long serving stalwart of the NZ Skeptics, a non profit body promoting critical thinking and calling out pseudoscience.
The Skeptics' website said they confront practices that exploit a lack of specialist knowledge among people. That's what many people use Google for - answers to things they don't know or things they don't understand.
Mark Honeychurch described putting overviews to the test in a recent edition of
the Skeptics' podcast
Yeah, Nah
.
"The AI looked like it was bending over backwards to please people. It's trying to give an answer that it knows that the customer wants," Honeychurch told
Mediawatch
.
Honeychurch asked Google for the meaning of: 'Better a skeptic than two geese.'
"It's trying to do pattern-matching and come out with something plausible. It does this so much that when it sees something that looks like an idiom that it's never heard before, it sees a bunch of idioms that have been explained and it just follows that pattern."
"It told me a skeptic is handy to have around because they're always questioning - but two geese could be a handful and it's quite hard to deal with two geese."
"With some of them, it did give me a caveat that this doesn't appear to be a popular saying. Then it would launch straight into explaining it. Even if it doesn't make sense, it still gives it its best go because that's what it's meant to do."
In time, would AI and Google detect the recent articles pointing out this flaw - and learn from them?
"There's a whole bunch of base training where (AI) just gets fed data from the Internet as base material. But on top of that, there's human feedback.
"They run it through a battery of tests and humans can basically mark the quality of answers. So you end up refining the model and making it better.
"By the time I tested this, it was warning me that a few of my fake idioms don't appear to be popular phrases. But then it would still launch into trying to explain it to me anyway, even though it wasn't real."
Things got more interesting - and alarming - when Honeychurch tested Google Overviews with real questions about religion, alternative medicine and skepticism.
"I asked why you shouldn't be a skeptic. I got a whole bunch of reasons that sounded plausible about losing all your friends and being the boring person at the party that's always ruining stories."
"When I asked it why you should be a skeptic, all I got was a message saying it cannot answer my question."
He also asked why one should be religious - and why not. And what reasons we should trust alternative medicines - and why we shouldn't.
"The skeptical, the rational, the scientific answer was the answer that Google's AI just refused to give."
"For the flip side of why I should be religious, I got a whole bunch of answers about community and a feeling of warmth and connecting to my spiritual dimension.
"I also got a whole bunch about how sometimes alternative medicine may have turned out to be true and so you can't just dismiss it."
"But we know why we shouldn't trust alternative medicine. It's alternative so it's not been proven to work. There's a very easy answer."
But not one Overview was willing or able to give, it seems.
Google does answer the neutral question 'Should I trust alternative medicine?' by saying there is "no simple answer" and "it's crucial to approach alternative medicine with caution and prioritise evidence-based conventional treatments."
So is Google trying not to upset people with answers that might concern them?
"I don't want to guess too much about that. It's not just Google but also OpenAI and other companies doing human feedback to try and make sure that it doesn't give horrific answers or say things that are objectionable."
"But it's always conflicting with the fact that this AI is just trained to give you that plausible answer. It's trying to match the pattern that you've given in the question."
Journalists use Google, just like anyone who's in a hurry and needs information quickly.
Do journalists need to ensure they don't rely on the Overviews summary right at the top of the search page?
"Absolutely. This is AI use 101. If you're asking something of a technical question, you really need to be well enough versed in what you're asking that you can judge whether the answer is good or not."
Sign up for Ngā Pitopito Kōrero
,
a daily newsletter curated by our editors and delivered straight to your inbox every weekday.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

NZ and ROK should work together on AI issues
NZ and ROK should work together on AI issues

Otago Daily Times

time3 hours ago

  • Otago Daily Times

NZ and ROK should work together on AI issues

Co-operation between Korea and New Zealand on AI is essential Choontae Park writes. Artificial Intelligence (AI) has now emerged as a key force, not only for technological innovation, but also as a variable that will shape the future of human civilisation. As of 2025, AI is actively integrated into nearly every field — daily life, industry, economy, education, healthcare and beyond. However, the changes brought by AI are not entirely positive. Challenges such as job displacement, invasion of privacy, deepening social inequality and ethical dilemmas lie ahead. These are complex issues that cannot be resolved by any single nation alone, as the impact of AI transcends national borders and spreads globally. Therefore, international co-operation and close alliances are more important than ever. The declaration announced at the 3rd AI Smart Forum held in Seoul on May 16 provided important direction for the future of AI. It emphasised AI was "not a replacement for humans, but a collaborator in building the future". It also asserted that technology could not "be completed by science alone; true civilisation [was] only possible when accompanied by humanistic reflection and philosophy." The declaration clearly stated that AI should enrich human life and contribute to world peace and universal values. This message offers profound insights into the path that Korea and New Zealand should walk together in the AI era. Korea has been experiencing rapid growth in the AI field on a global scale. In December 2023, the Korean National Assembly passed the Basic Act on Artificial Intelligence, one of the world's first comprehensive AI laws. This law institutionalises the transparency, safety and ethical use of AI. Korea is taking the lead in shaping international norms through collaboration. It continues to pursue a balance between technological innovation and ethical responsibility, ensuring that AI development earns public trust and accountability. Meanwhile, New Zealand emphasises inclusive and sustainable AI development through its AI Blueprint for Aotearoa New Zealand (2020) and the Algorithm Charter for Aotearoa New Zealand (introduced in 2020). The country's AI policies prioritise core values such as human rights, diversity, inclusion, transparency, safety and accountability. Notably, New Zealand places indigenous data sovereignty, welfare and the protection of minority rights at the heart of its AI policies. This approach ensures that AI does not harm any specific group and earns broad public trust and participation. Korea and New Zealand possess complementary strengths in their AI strategies: Korea excels in technological innovation and policy leadership, while New Zealand offers a strong human-rights-centred and inclusive approach. There are clear reasons why the two countries should collaborate. First, to strengthen ethical standards and humanistic reflection in AI. By jointly promoting AI research and education that combine science, humanities and philosophy, we can ensure that AI does not undermine human dignity or social values. Second, to build a global AI governance system together. To address risks such as algorithmic bias and privacy infringement, the two nations must collaborate in international organisations and multilateral forums. Third, to create an inclusive AI ecosystem. Korea and New Zealand should work to increase participation by women, indigenous peoples and minorities in AI development, and share policies that guarantee data sovereignty and fairness. Fourth, joint research and standardisation are essential to secure public trust. Only AI that citizens can trust will achieve broad social acceptance. Fifth, AI should be actively applied to realise social values. Collaborative projects should focus on addressing global issues such as climate change, public health, welfare and education. Korea and New Zealand have already laid the groundwork for diverse AI co-operation. Going forward, the two countries should deepen their partnership in the following ways: conducting joint research and develop policies in AI ethics and governance; collaborating in the creation of international AI standards and actively participating in global regulatory discussions; and expanding AI education and talent exchange programmes to help nurture global AI leaders from the younger generation. The countries should also develop and implement inclusive AI policies that protect vulnerable populations and guarantee data rights and promote joint projects in areas like climate action, healthcare, welfare and education using AI. Ultimately, AI must enrich human life and help achieve global peace and shared values. Technology is not completed by science alone — true civilisation is possible only when paired with reflection and philosophy. By pursuing a harmonious blend of innovation, ethics, inclusion, trust and international co-operation, Korea and New Zealand can become exemplary partners in the AI era. — Dr Choontae Park is a former university dean who now lives in New Zealand.

Twilio appoints Howard Fyffe as Director of Sales for ANZ
Twilio appoints Howard Fyffe as Director of Sales for ANZ

Techday NZ

time20 hours ago

  • Techday NZ

Twilio appoints Howard Fyffe as Director of Sales for ANZ

Twilio has announced the appointment of Howard Fyffe as Director of Sales for Australia and New Zealand (ANZ). Fyffe will oversee Twilio's Communications business in the ANZ region, leading a sales team tasked with supporting brands in delivering personalised digital engagement to customers. The remit for Fyffe includes fostering sales growth in the region and guiding operational priorities to support brands in strengthening their digital connections with customers. His responsibilities will encompass driving the expansion of Twilio's services in the ANZ market, as well as supporting existing clients in their engagement strategies. Fyffe brings more than 20 years of industry experience to the role, with expertise spanning AI, automation, data centre, hybrid cloud, and risk and compliance. During his career, he has held leadership roles at organisations such as xAmplify, where he served as Chief Revenue Officer, as well as multiple positions at Cisco Systems in Asia Pacific, Australia, and New York. Inclusion of prior appointments at VAST Data, Veritas, and Nutanix further highlights his breadth of experience in both enterprise sales and operations management. Commenting on the appointment, Robert Woolfrey, Vice President for Communications, Asia Pacific & Japan at Twilio said: "Howard is a proven enterprise leader with a strong track record of driving strategic growth in Australia and New Zealand. He joins Twilio at a pivotal time, as we double down on new business, account expansion, and operational excellence in ANZ. With his leadership, we'll accelerate our momentum and help even more customers build smarter, AI-powered customer experiences." Fyffe's appointment comes as Twilio intensifies its focus on business growth, account development, and operational strategy throughout the region. His leadership will include enabling high-performing teams to deliver digital solutions that assist brands in establishing and deepening customer relationships. On taking up the new post, Fyffe stated: "I'm thrilled to join Twilio to lead a dynamic and talented team focused on helping businesses across Australia and New Zealand strengthen their customer engagement. Twilio sits at the intersection of technology and human connection, helping organisations turn digital interactions into meaningful experiences. Throughout my career, I've always been passionate about using innovation to solve real customer challenges, and I look forward to contributing to the strong work the team is already doing to advance Twilio's mission." Fyffe's experience and perspective are expected to support Twilio as it seeks greater penetration in ANZ's digital engagement market. The company's strategy in the region is aimed at helping organisations leverage communications technology and data to optimise and personalise customer interactions across a range of applications, including sales, marketing, growth, and customer service. The appointment aligns with ongoing trends in customer engagement, where personalisation and digital transformation remain a focus for businesses seeking to maintain competitive advantage. Fyffe's leadership is anticipated to play a role in facilitating the adoption of digital and AI-powered solutions among Twilio's current and prospective clients in Australia and New Zealand. Twilio continues to operate across 180 countries, providing its customer engagement platform to a wide range of businesses and developers globally. Follow us on: Share on:

Google rolls out its Gemini chatbot to Kiwi kids
Google rolls out its Gemini chatbot to Kiwi kids

RNZ News

time21 hours ago

  • RNZ News

Google rolls out its Gemini chatbot to Kiwi kids

Kiwi kids can now access Google's Gemini AI chatbot Photo: aap Arriens / NurPhoto via AFP Kiwi kids can now access Google's Gemini AI chatbot, now the company has made it available to the under-13s. The tech giant emailed global users of its parental control device Family Link in May to announce the pending change - which was rolled out first to users in the US, and then worldwide. In the email the company acknowledged Gemini can "make mistakes" and parents should remind their children it "isn't human". The rollout's automatic - so parents have to opt-out if they don't want their child to use it. Should chatbots be part of children's learning? Kathryn speaks with Toby Walsh, author of the 2023 book ' Machines Behaving Badly ', which investigated the pace of AI's rollout and the limited regulation or oversight into it. He's Laureate Fellow and Scientia Professor of Artificial Intelligence at the University of New South Wales and CSIRO Data61.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store