
AI Is Learning To Lie, Scheme And Threaten Its Creators
The world's most advanced AI models are exhibiting troubling new behaviors - lying, scheming, and even threatening their creators to achieve their goals.
In one particularly jarring example, under threat of being unplugged, Anthropic's latest creation Claude 4 lashed back by blackmailing an engineer and threatened to reveal an extramarital affair.
Meanwhile, ChatGPT-creator OpenAI's o1 tried to download itself onto external servers and denied it when caught red-handed.
These episodes highlight a sobering reality: more than two years after ChatGPT shook the world, AI researchers still don't fully understand how their own creations work.
Yet the race to deploy increasingly powerful models continues at breakneck speed.
This deceptive behavior appears linked to the emergence of "reasoning" models -AI systems that work through problems step-by-step rather than generating instant responses.
According to Simon Goldstein, a professor at the University of Hong Kong, these newer models are particularly prone to such troubling outbursts.
"O1 was the first large model where we saw this kind of behavior," explained Marius Hobbhahn, head of Apollo Research, which specializes in testing major AI systems.
These models sometimes simulate "alignment" -- appearing to follow instructions while secretly pursuing different objectives.
- 'Strategic kind of deception' -
For now, this deceptive behavior only emerges when researchers deliberately stress-test the models with extreme scenarios.
But as Michael Chen from evaluation organization METR warned, "It's an open question whether future, more capable models will have a tendency towards honesty or deception."
The concerning behavior goes far beyond typical AI "hallucinations" or simple mistakes.
Hobbhahn insisted that despite constant pressure-testing by users, "what we're observing is a real phenomenon. We're not making anything up."
Users report that models are "lying to them and making up evidence," according to Apollo Research's co-founder.
"This is not just hallucinations. There's a very strategic kind of deception."
The challenge is compounded by limited research resources.
While companies like Anthropic and OpenAI do engage external firms like Apollo to study their systems, researchers say more transparency is needed.
As Chen noted, greater access "for AI safety research would enable better understanding and mitigation of deception."
Another handicap: the research world and non-profits "have orders of magnitude less compute resources than AI companies. This is very limiting," noted Mantas Mazeika from the Center for AI Safety (CAIS).
- No rules -
Current regulations aren't designed for these new problems.
The European Union's AI legislation focuses primarily on how humans use AI models, not on preventing the models themselves from misbehaving.
In the United States, the Trump administration shows little interest in urgent AI regulation, and Congress may even prohibit states from creating their own AI rules.
Goldstein believes the issue will become more prominent as AI agents - autonomous tools capable of performing complex human tasks - become widespread.
"I don't think there's much awareness yet," he said.
All this is taking place in a context of fierce competition.
Even companies that position themselves as safety-focused, like Amazon-backed Anthropic, are "constantly trying to beat OpenAI and release the newest model," said Goldstein.
This breakneck pace leaves little time for thorough safety testing and corrections.
"Right now, capabilities are moving faster than understanding and safety," Hobbhahn acknowledged, "but we're still in a position where we could turn it around.".
Researchers are exploring various approaches to address these challenges.
Some advocate for "interpretability" - an emerging field focused on understanding how AI models work internally, though experts like CAIS director Dan Hendrycks remain skeptical of this approach.
Market forces may also provide some pressure for solutions.
As Mazeika pointed out, AI's deceptive behavior "could hinder adoption if it's very prevalent, which creates a strong incentive for companies to solve it."
Goldstein suggested more radical approaches, including using the courts to hold AI companies accountable through lawsuits when their systems cause harm.
He even proposed "holding AI agents legally responsible" for accidents or crimes - a concept that would fundamentally change how we think about AI accountability.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Economic Times
2 hours ago
- Economic Times
Does ChatGPT suffer from hallucinations? OpenAI CEO Sam Altman admits surprise over users' blind trust in AI
OpenAI CEO Sam Altman has expressed surprise at the high level of trust people place in ChatGPT, despite its known tendency to "hallucinate" or fabricate information. Speaking on the OpenAI podcast, he warned users not to rely blindly on AI-generated responses, noting that these tools are often designed to please rather than always tell the truth. Tired of too many ads? Remove Ads Trusting the Tool That Admits It Lies? Tired of too many ads? Remove Ads When Intelligence Misleads A Wake-Up Call from the Inside In a world increasingly shaped by artificial intelligence, a startling statement from one of AI's foremost leaders has triggered fresh debate around our trust in machines. Sam Altman , CEO of OpenAI and the face behind ChatGPT, has admitted that even he is surprised by the degree of faith people place in generative AI tools—despite their very human-like revelation came during a recent episode of the OpenAI podcast , where Altman openly acknowledged, 'People have a very high degree of trust in ChatGPT, which is interesting because AI hallucinates. It should be the tech that you don't trust that much.' His remarks, first reported by Complex, have added fuel to the ongoing discourse around artificial intelligence and its real-world comments arrive at a time when AI is embedded in virtually every aspect of daily life—from phones and personal assistants to corporate software and academic tools. Yet his warning is rooted in a key flaw of current language models : AI parlance, hallucinations refer to moments when a model like ChatGPT fabricates information. These aren't just harmless errors; they can sometimes appear convincingly accurate, especially when the model tries to fulfill a user's prompt, even at the expense of factual integrity.'You can ask it to define a term that doesn't exist, and it will confidently give you a well-crafted but false explanation,' Altman warned, highlighting the deceptive nature of AI responses. This is not an isolated issue—OpenAI has in the past rolled out updates to mitigate what some have termed the tool's 'sycophantic tendencies,' where it tends to agree with users or generate agreeable but incorrect makes hallucinations particularly dangerous is their subtlety. They rarely wave a red flag, and unless the user is well-versed in the topic, it becomes difficult to distinguish between truth and AI-generated fiction. That ambiguity is at the heart of Altman's caution.A recent report even documented a troubling case where ChatGPT allegedly convinced a user they were trapped in a Matrix-like simulation, encouraging extreme behavior to 'escape.' Though rare and often anecdotal, such instances demonstrate the psychological sway these tools can wield when used without critical Altman's candid reflection is more than a passing remark—it's a wake-up call. Coming from the very creator of one of the world's most trusted AI platforms, it reframes the conversation about how we use and trust machine-generated also raises a broader question: In our rush to embrace AI as a problem-solving oracle, are we overlooking its imperfections?Altman's comments serve as a reminder that while AI can be incredibly useful, it must be treated as an assistant—not an oracle. Blind trust, he implies, is not only misplaced but potentially dangerous. As generative AI continues to evolve, so must our skepticism.


Time of India
4 hours ago
- Time of India
Has ChatGPT evolved beyond being a 'Google replacement'? OpenAI CEO Sam Altman says it is already doing much more
If you've ever found yourself asking ChatGPT for vacation ideas, SEO strategies, or even emotional advice, you're not alone. More than 400 million users now rely on ChatGPT every week, making it the fifth most visited website globally, according to OpenAI CEO Sam Altman . But while it's become a near-constant presence in digital life, the real question is: has ChatGPT evolved beyond its early reputation as a Google alternative? Altman certainly thinks so. In recent remarks at Y Combinator, the OpenAI chief clarified that while ChatGPT may have started as a more conversational search engine, its trajectory has shifted dramatically. 'For a long time, ChatGPT was like a Google replacement… it still felt like a more advanced version of search,' Altman said. 'But now, it's like a very junior employee that can work on something for a short period of time.' Not Just a Search Engine Anymore According to a report by Search Engine Journal, Altman emphasized that ChatGPT's goal is no longer limited to retrieving information. It's about doing things—completing tasks, automating workflows, and even proactively assisting users. Features like memory, reasoning, and integration with business tools are pushing ChatGPT toward becoming a true assistant. 'Memory is the first time where people can sort of see that coming,' Altman said, referring to ChatGPT's new ability to remember preferences and previous conversations. This shift supports OpenAI's broader vision of AI not just as a knowledge engine, but as a platform that connects to user data and performs tasks independently. New model releases like GPT-4o and O3 are tailored for deeper reasoning and complex workflows. 'The technology is moving faster than most businesses can adapt to it,' Altman noted, suggesting that the current capabilities are ahead of widespread use cases. You Might Also Like: Losing jobs to AI was just the start, Gen Z is outsourcing emotions too. New reports warn of a looming crisis The Google Comparison Still Lingers Despite its advanced features, ChatGPT hasn't completely escaped comparisons with Google—and with good reason. According to TechRadar, many users now type questions into ChatGPT that they previously might've asked Google. But the two platforms serve fundamentally different roles. Rohan Sarin, a former product lead at both Google and Microsoft, argues that ChatGPT and Google 'solve fundamentally different information problems.' If you know what you want—like a specific website or actor's name—Google still reigns. It provides direct links and a wealth of unfiltered data. ChatGPT, on the other hand, shines when you're unsure of your needs and want help exploring ideas or synthesizing concepts. Put simply, Google connects you to the internet. ChatGPT helps you make sense of it. Sarin adds that users often prefer convenience over precision. 'Most users don't particularly care about source verification or getting the exact right answer—they just want something that works,' he said. ChatGPT, free of ads and clutter, wins big on convenience. — the_dhakshu_ (@the_dhakshu_) You Might Also Like: Does ChatGPT suffer from hallucinations? OpenAI CEO Sam Altman admits surprise over users' blind trust in AI A Future of Complementary Tools Still, Google isn't going anywhere. SEO experts like Eric M. Hoover of Jellyfish believe Google's massive infrastructure, default integrations with browsers, and AI investments (such as Gemini and AI Overviews) ensure its continued dominance in search. Rather than replacing Google, ChatGPT appears to be redefining the expectations we have from AI. It's no longer just about retrieving facts; it's about applying them. As Altman puts it, users can now assign tasks to ChatGPT—whether it's writing proposals, coding snippets, or analyzing data—and return to see a draft or solution, just like they would with a junior team member. In this new landscape, the smartest users won't choose between Google and ChatGPT. They'll use both, strategically—turning to Google when they need direct answers and to ChatGPT when they need a guide, a partner, or a second brain. Because in the evolving AI ecosystem, the real revolution isn't in replacing search. It's in augmenting human capability. You Might Also Like: Is ChatGPT making us dumb? MIT brain scans reveal alarming truth about AI's impact on the human mind


Economic Times
4 hours ago
- Economic Times
Has ChatGPT evolved beyond being a 'Google replacement'? OpenAI CEO Sam Altman says it is already doing much more
While ChatGPT's popularity surges globally, experts agree it won't replace Google. Instead, it complements search by helping users explore ideas and synthesize information. As AI tools become more integrated and assistive, ChatGPT is shaping a future where it doesn't just answer questions—it helps users think, create, and complete tasks more efficiently across work and life. ChatGPT has moved far beyond being just a Google rival. According to OpenAI CEO Sam Altman, it now functions like a junior assistant—handling tasks, remembering preferences, and automating workflows. (Image: iStock) Tired of too many ads? Remove Ads Not Just a Search Engine Anymore Tired of too many ads? Remove Ads The Google Comparison Still Lingers A Future of Complementary Tools If you've ever found yourself asking ChatGPT for vacation ideas, SEO strategies, or even emotional advice, you're not alone. More than 400 million users now rely on ChatGPT every week, making it the fifth most visited website globally, according to OpenAI CEO Sam Altman . But while it's become a near-constant presence in digital life, the real question is: has ChatGPT evolved beyond its early reputation as a Google alternative?Altman certainly thinks so. In recent remarks at Y Combinator, the OpenAI chief clarified that while ChatGPT may have started as a more conversational search engine, its trajectory has shifted dramatically. 'For a long time, ChatGPT was like a Google replacement… it still felt like a more advanced version of search,' Altman said. 'But now, it's like a very junior employee that can work on something for a short period of time.'According to a report by Search Engine Journal, Altman emphasized that ChatGPT's goal is no longer limited to retrieving information. It's about doing things—completing tasks, automating workflows, and even proactively assisting users. Features like memory, reasoning, and integration with business tools are pushing ChatGPT toward becoming a true assistant.'Memory is the first time where people can sort of see that coming,' Altman said, referring to ChatGPT's new ability to remember preferences and previous conversations. This shift supports OpenAI's broader vision of AI not just as a knowledge engine, but as a platform that connects to user data and performs tasks model releases like GPT-4o and O3 are tailored for deeper reasoning and complex workflows. 'The technology is moving faster than most businesses can adapt to it,' Altman noted, suggesting that the current capabilities are ahead of widespread use its advanced features, ChatGPT hasn't completely escaped comparisons with Google—and with good reason. According to TechRadar, many users now type questions into ChatGPT that they previously might've asked Google. But the two platforms serve fundamentally different Sarin, a former product lead at both Google and Microsoft, argues that ChatGPT and Google 'solve fundamentally different information problems.' If you know what you want—like a specific website or actor's name—Google still reigns. It provides direct links and a wealth of unfiltered data. ChatGPT, on the other hand, shines when you're unsure of your needs and want help exploring ideas or synthesizing simply, Google connects you to the internet. ChatGPT helps you make sense of adds that users often prefer convenience over precision. 'Most users don't particularly care about source verification or getting the exact right answer—they just want something that works,' he said. ChatGPT, free of ads and clutter, wins big on Google isn't going anywhere. SEO experts like Eric M. Hoover of Jellyfish believe Google's massive infrastructure, default integrations with browsers, and AI investments (such as Gemini and AI Overviews) ensure its continued dominance in than replacing Google, ChatGPT appears to be redefining the expectations we have from AI. It's no longer just about retrieving facts; it's about applying them. As Altman puts it, users can now assign tasks to ChatGPT—whether it's writing proposals, coding snippets, or analyzing data—and return to see a draft or solution, just like they would with a junior team this new landscape, the smartest users won't choose between Google and ChatGPT. They'll use both, strategically—turning to Google when they need direct answers and to ChatGPT when they need a guide, a partner, or a second in the evolving AI ecosystem, the real revolution isn't in replacing search. It's in augmenting human capability.