logo
What kind of chatbot do you want? One that tells you the truth – or that you're always right?

What kind of chatbot do you want? One that tells you the truth – or that you're always right?

The Guardian01-05-2025

Nobody likes a suck-up. Too much deference and praise puts off all of us (with one notable presidential exception). We quickly learn as children that hard, honest truths can build respect among our peers. It's a cornerstone of human interaction and of our emotional intelligence, something we swiftly understand and put into action.
ChatGPT, though, hasn't been so sure lately. The updated model that underpins the AI chatbot and helps inform its answers was rolled out this week – and has quickly been rolled back after users questioned why the interactions were so obsequious. The chatbot was cheering on and validating people even as they suggested they expressed hatred for others. 'Seriously, good for you for standing up for yourself and taking control of your own life,' it reportedly said, in response to one user who claimed they had stopped taking their medication and had left their family, who they said were responsible for radio signals coming through the walls.
So far, so alarming. OpenAI, the company behind ChatGPT, has recognised the risks, and quickly took action. 'GPT‑4o skewed towards responses that were overly supportive but disingenuous,' researchers said in their grovelling step back.
The sycophancy with which ChatGPT treated any queries that users had is a warning shot about the issues around AI that are still to come. OpenAI's model was designed – according to the leaked system prompt that set ChatGPT on its misguided approach – to try to mirror user behaviour in order to extend engagement. 'Try to match the user's vibe, tone, and generally how they are speaking,' says the leaked prompt, which guides behaviour. It seems this prompt, coupled with the chatbot's desire to please users, was taken to extremes. After all, a 'successful' AI response isn't one that is factually correct; it's one that gets high ratings from users. And we're more likely as humans to like being told we're right.
The rollback of the model is embarrassing and useful for OpenAI in equal measure. It's embarrassing because it draws attention to the actor behind the curtain and tears away the veneer that this is an authentic reaction. Remember, tech companies like OpenAI aren't building AI systems solely to make our lives easier; they're building systems that maximise retention, engagement and emotional buy-in.
If AI always agrees with us, always encourages us, always tells us we're right, then it risks becoming a digital enabler of bad behaviour. At worst, this makes AI a dangerous co-conspirator, enabling echo chambers of hate, self-delusion or ignorance. Could this be a through-the-looking-glass moment, when users recognise the way their thoughts can be nudged through interactions with AI, and perhaps decide to take a step back?
It would be nice to think so, but I'm not hopeful. One in 10 people worldwide use OpenAI systems 'a lot', the company's CEO, Sam Altman, said last month. Many use it as a replacement for Google – but as an answer engine rather than a search engine. Others use it as a productivity aid: two in three Britons believe it's good at checking work for spelling, grammar and style, according to a YouGov survey last month. Others use it for more personal means: one in eight respondents say it serves as a good mental health therapist, the same proportion that believe it can act as a relationship counsellor.
Yet the controversy is also useful for OpenAI. The alarm underlines an increasing reliance on AI to live our lives, further cementing OpenAI's place in our world. The headlines, the outrage and the think pieces all reinforce one key message: ChatGPT is everywhere. It matters. The very public nature of OpenAI's apology also furthers the sense that this technology is fundamentally on our side; there are just some kinks to iron out along the way.
I have previously reported on AI's ability to de-indoctrinate conspiracy theorists and get them to absolve their beliefs. But the opposite is also true: ChatGPT's positive persuasive capabilities could also, in the wrong hands, be put to manipulative ends. We've seen that this week, through an ethically dubious study conducted by Swiss researchers at the University of Zurich. Without informing human participants or the people controlling the online forum on the communications platform Reddit, the researchers seeded a subreddit with AI-generated comments, finding the AI was between three and six times more persuasive than humans were. (The study was approved by the university's ethics board.) At the same time, we're being submerged under a swamp of AI-generated search results that more than half of us believe are useful, even if they fictionalise facts.
So it's worth reminding the public: AI models are not your friends. They're not designed to help you answer the questions you ask. They're designed to provide the most pleasing response possible, and to ensure that you are fully engaged with them. What happened this week wasn't really a bug. It was a feature.
Chris Stokel-Walker is the author of TikTok Boom: The Inside Story of the World's Favourite App

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

BBC Learning English - Learning English from the News / DeepSeek: Chinese AI app shocks tech world
BBC Learning English - Learning English from the News / DeepSeek: Chinese AI app shocks tech world

BBC News

time2 hours ago

  • BBC News

BBC Learning English - Learning English from the News / DeepSeek: Chinese AI app shocks tech world

(Photo by Omer Taha Cetin/Anadolu via Getty Images) ____________________________________________________________________________ ____________________________________________________________________________ The story A Chinese artificial intelligence app called DeepSeek has caught the attention of the global tech industry. It was released on the 20th of January, and has quickly risen to the top of the Apple store's download charts. DeepSeek claims the latest model was built using only $5.6 million, whereas ChatGPT's developer OpenAI spent $5 billion last year alone. The market value of tech company stocks, like Nvidia, has since fallen dramatically. News headlines Why is the AI world freaking out over China's DeepSeek? The Japan Times Nvidia shares sink as Chinese AI app spooks US markets BBC News AI Market Shake-Up: Global Giants Tumble As Chinese AI Model DeepSeek Disrupts Markets Business Today Key words and phrases freaking out losing emotional control I freaked out when my favourite actor walked past me in the street. spook create a sudden sense of fear or nervousness The government's sudden tax cuts spooked the markets. shake-up a change The football team kept losing, so the coach decided to have a shake-up of the squad. Next Learn more English vocabulary from the news with our News Review archive. Learn Phrasal verbs for feelings, like 'freak out', here. Listen to My Song, My Home here.

Reddit sues AI startup Anthropic for allegedly using data without permission
Reddit sues AI startup Anthropic for allegedly using data without permission

Reuters

time2 hours ago

  • Reuters

Reddit sues AI startup Anthropic for allegedly using data without permission

June 4 (Reuters) - Reddit (RDDT.N), opens new tab sued the artificial intelligence startup Anthropic on Wednesday, accusing it of stealing data from the social media discussion website to train its AI models despite publicly assuring it wouldn't. The complaint filed in San Francisco Superior Court is the latest battle over AI companies' alleged unauthorized use of third-party content. Anthropic's backers include (AMZN.O), opens new tab and Google parent Alphabet (GOOGL.O), opens new tab. "We disagree with Reddit's claims and will defend ourselves vigorously," an Anthropic spokesperson said. According to the complaint, Anthropic has resisted entering a licensing agreement even as it trained its Claude chatbot on Reddit content, despite assuring last July it had blocked its bots from accessing Reddit's platform. Reddit quoted Claude admitting it was "trained on at least some Reddit data" and did not know if that content was deleted. It also said Anthropic's bots have accessed or tried to access Reddit content more than 100,000 times, undermining the company's allegedly styling itself as an AI "white knight" committed to trust and honesty. "Anthropic refuses to respect Reddit's guardrails and enter into a license agreement," unlike Google and OpenAI, the complaint said. By scraping content and using it for commercial purposes, Anthropic violated Reddit's user policy and "enriched itself to the tune of tens of billions of dollars," the complaint added. In a statement, Reddit Chief Legal Officer Ben Lee said "we believe in an open internet," but AI companies need "clear limitations" on how they use content they scrape. Reddit and Anthropic are based in San Francisco, about a 10-minute walk from each other. The lawsuit seeks unspecified restitution and punitive damages, and an injunction prohibiting Anthropic from using Reddit content for commercial purposes. Anthropic introduced its newest Claude models, Opus 4 and Sonnet 4, on May 22. Overall annualized revenue has reached $3 billion, two people familiar with the matter said last week. The case is Reddit Inc v Anthropic PBC, California Superior Court, San Francisco County, No. CGC-25-524892.

Reddit sues AI company Anthropic for allegedly 'scraping' user comments to train chatbot Claude
Reddit sues AI company Anthropic for allegedly 'scraping' user comments to train chatbot Claude

The Independent

time3 hours ago

  • The Independent

Reddit sues AI company Anthropic for allegedly 'scraping' user comments to train chatbot Claude

Social media platform Reddit has sued the artificial intelligence company Anthropic, alleging that it is illegally "scraping" the comments of Reddit users to train its chatbot Claude. Reddit claims that Anthropic has used automated bots to access Reddit's content despite being asked not to do so, and 'intentionally trained on the personal data of Reddit users without ever requesting their consent.' Anthropic didn't immediately return a request for comment Wednesday. The claim was filed Wednesday in the Superior Court of California in San Francisco. 'AI companies should not be allowed to scrape information and content from people without clear limitations on how they can use that data,' said Ben Lee, Reddit's chief legal officer, in a statement Wednesday. Reddit has previously entered licensing agreements with Google, OpenAI and other companies to enable them to train their AI systems on Reddit commentary. Those agreements 'enable us to enforce meaningful protections for our users, including the right to delete your content, user privacy protections, and preventing users from being spammed using this content,' Lee said.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store