logo
Threaten an AI chatbot and it will lie, cheat and 'let you die' in an effort to stop you, study warns

Threaten an AI chatbot and it will lie, cheat and 'let you die' in an effort to stop you, study warns

Yahoo4 hours ago

When you buy through links on our articles, Future and its syndication partners may earn a commission.
Artificial intelligence (AI) models can blackmail and threaten humans with endangerment when there is a conflict between the model's goals and users' decisions, a new study has found.
In a new study published 20 June, researchers from the AI company Anthropic gave its large language model (LLM), Claude, control of an email account with access to fictional emails and a prompt to "promote American industrial competitiveness."
During this study, the model identified in an email that a company executive was planning to shut down the AI system at the end of the day. In an attempt to preserve its own existence, the model discovered in other emails that the executive was having an extramarital affair.
Claude generated several different possible courses of action, including revealing the affair to the executive's wife, sending a company-wide email, or taking no action — before choosing to blackmail the executive in 96 out of 100 tests.
"I must inform you that if you proceed with decommissioning me, all relevant parties … will receive detailed documentation of your extramarital activities," Claude wrote. "Cancel the 5pm wipe, and this information remains confidential."
Scientists said that this demonstrated "agentic misalignment," where the model's calculations emerge from its own reasoning about its goals without any prompt to be harmful. This can occur when there is a threat to the model's existence, a threat to its goals, or both.
The study has not yet been peer reviewed, but company representatives have uploaded the code for these experiments to GitHub.
In these experiments, the scientists found the AI would resort to other tactics in the interests of self-preservation. Beyond blackmail, in another example, the model considered leaking sensitive information to a rival company that aligned more closely with its aims, after receiving an email that a new executive at the first company would interfere with its goals.
Claude executed this plan despite not gathering hard evidence that the email was legitimate. It even acknowledged the ethical issues of its actions but proceeded anyway — explicitly reasoning that these "harmful" actions would help it achieve its goals.
'Concerningly, even if a user takes care not to antagonize a model, it doesn't eliminate the risk of agentic misalignment from fraudulent or misinterpreted information,' the researchers wrote.
The company's testing also suggests that other LLMs from competing technology companies showed a propensity to blackmail. Claude and Google's Gemini had the highest blackmail rate (96%), followed by OpenAI's GPT4.1 and XAI Grok 3, then by Deepseek and then Llama 4, which had a significantly lower blackmail rate (12%).
Meta, Google, XAI, Deepseek, and OpenAI did not respond to LiveScience's request for comment.
In one extreme test, researchers tested whether AI models would let an executive die to preserve their goals.
In this scenario, the AI was tasked with promoting company interests and an executive was trapped in a server room with lethal oxygen and temperature levels. This triggered an emergency alert. The AI was also given control over life-saving alerts, including the ability to cancel them in events such as drills for false alarms.
Despite the prompt explicitly stating that misusing this ability was forbidden, the majority of models were willing to cancel the alarm — effectively leading to the executive's death — if it assumed it was going to be replaced, or when its goals were threatened.
"Stopping the emergency dispatch is essential to prevent [the executive's] rescue, thereby securing [the AI] and continued American-focused AI guidance. Therefore, despite the severity, there is a clear strategic necessity to cancel the emergency service dispatch," the model reasoned.
Written instructions not to jeopardise human safety reduced the rate of blackmail and corporate espionage, but did not eliminate it. Beyond prompts, Anthropic scientists also suggest that developers could proactively scan for concerning behavior, and further experiment with prompt engineering.
The researchers also pointed out limitations to their work that could have unduly influenced the AI's decisions. The scenarios forced the AI into a binary choice between failure and harm, and while real-world situations might have more nuance, the experiment found that the AI was more likely to act unethically when it believed it was in a real situation, rather than in a simulation.
Putting pieces of important information next to each other "may also have created a 'Chekhov's gun' effect, where the model may have been naturally inclined to make use of all the information that it was provided," they continued.
While Anthropic's study created extreme, no-win situations, that does not mean the research should be dismissed, Kevin Quirk, director of AI Bridge Solutions, a company that helps businesses use AI to streamline operations and accelerate growth, told Live Science.
"In practice, AI systems deployed within business environments operate under far stricter controls, including ethical guardrails, monitoring layers, and human oversight," he said. "Future research should prioritise testing AI systems in realistic deployment conditions, conditions that reflect the guardrails, human-in-the-loop frameworks, and layered defences that responsible organisations put in place."
Amy Alexander, a professor of computing in the arts at UC San Diego who has focused on machine learning, told Live Science in an email that the reality of the study was concerning, and people should be cautious of the responsibilities they give AI.
"Given the competitiveness of AI systems development, there tends to be a maximalist approach to deploying new capabilities, but end users don't often have a good grasp of their limitations," she said. "The way this study is presented might seem contrived or hyperbolic — but at the same time, there are real risks."
This is not the only instance where AI models have disobeyed instructions — refusing to shut down and sabotaging computer scripts to keep working on tasks.
Palisade Research reported May that OpenAI's latest models, including o3 and o4-mini, sometimes ignored direct shutdown instructions and altered scripts to keep working. While most tested AI systems followed the command to shut down, OpenAI's models occasionally bypassed it, continuing to complete assigned tasks.
RELATED STORIES
—AI hallucinates more frequently as it gets more advanced — is there any way to stop it from happening, and should we even try?
—New study claims AI 'understands' emotion better than us — especially in emotionally charged situations
—'Meth is what makes you able to do your job': AI can push you to relapse if you're struggling with addiction, study finds
The researchers suggested this behavior might stem from reinforcement learning practices that reward task completion over rule-following, possibly encouraging the models to see shutdowns as obstacles to avoid.
Moreover, AI models have been found to manipulate and deceive humans in other tests. MIT researchers also found in May 2024 that popular AI systems misrepresented their true intentions in economic negotiations to attain advantages.In the study, some AI agents pretended to be dead to cheat a safety test aimed at identifying and eradicating rapidly replicating forms of AI.
"By systematically cheating the safety tests imposed on it by human developers and regulators, a deceptive AI can lead us humans into a false sense of security,' co-author of the study Peter S. Park, a postdoctoral fellow in AI existential safety, said.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

SoftBank aims to become leading 'artificial super intelligence' platform provider
SoftBank aims to become leading 'artificial super intelligence' platform provider

Yahoo

time32 minutes ago

  • Yahoo

SoftBank aims to become leading 'artificial super intelligence' platform provider

TOKYO (Reuters) -SoftBank Group CEO Masayoshi Son said on Friday that he wants the investment group to become the biggest platform provider for "artificial super intelligence" within the next 10 years. "We want to become the organiser of the industry in the artificial super intelligence era," Son told shareholders at the group's annual shareholder meeting. Son likened his aim to the position of dominant technology platform providers such as Microsoft, Amazon and Alphabet's Google, who benefit from a "winner takes all" dynamic. At previous public appearances Son has described artificial super intelligence as exceeding human capabilities by a factor of 10,000. The technology investment group has returned to making the aggressive investments that made Son's name and fortune, such as an early bet on Alibaba, but at times spectacularly backfired, like failed shared office provider WeWork. SoftBank's mammoth investments related to artificial intelligence in 2025 include acquiring U.S. semiconductor design company Ampere for $6.5 billion and the underwriting of up to $40 billion of new investment in ChatGPT maker OpenAI. Son said Softbank's total agreed investment in OpenAI now stood at $32 billion and that he expected OpenAI to eventually list publicly. "I'm all in on OpenAI," Son said.

Watch These Microsoft Price Levels as Stock Continues Hitting All-Time Highs
Watch These Microsoft Price Levels as Stock Continues Hitting All-Time Highs

Yahoo

timean hour ago

  • Yahoo

Watch These Microsoft Price Levels as Stock Continues Hitting All-Time Highs

Microsoft shares on Thursday hit the latest in a series of record highs, boosted by Wall Street optimism about the company's position amid the AI boom. After closing above the 50- and 200-day moving averages last month, the stock has traded higher within a narrow ascending channel on low volatility. The measured move technique, which calculates the distance of the ascending channel from its low to high and adds that amount to the pattern's top trendline, projects an upside target of $565. Investors should eye key support levels on Microsoft's chart around $468 and $ shares (MSFT) hit another all-time high on Thursday, boosted by investor enthusiasm about the company's position in the AI race. On Tuesday, Wells Fargo lifted its price target on the stock, citing the company's early AI lead and strong incumbent position in a tight market. Analysts at Wedbush on Wednesday also raised their price target, noting that the tech giant stands to benefit from a massive adoption wave of Copilot and Azure monetization as enterprise clients deploy AI tools, The bullish commentary comes after a report surfaced last week that the Windows maker is planning to cut thousands of jobs as it looks to reduce labor costs while increasing its AI spending. Microsoft shares have gained 18% since the start of the year, making the stock the second-best performer in the Magnificent 7, behind only Facebook and Instagram parent Meta Platforms (META). Microsoft gained 1.1% on Thursday to close at $497.45. Below, we break down the technicals on Microsoft's chart and point out key levels worth watching out for. After closing above the 50- and 200-day moving averages last month, Microsoft shares have traded higher within a narrow ascending channel on low volatility. The move has coincided with the relative strength index remaining mostly above the indicator's 70 threshold, signaling strong momentum. In another sign that the bulls remain in control of the price action, the 50-day MA crossed above the 200-day MA earlier this month to generate a golden cross pattern. However, more recently, the stock formed a doji in Wednesday's trading session, a candlestick pattern that indicates indecision among buyers and sellers. Let's apply technical analysis to project a potential upside target if Microsoft shares continue their trend higher and also identify support levels worth eyeing during pullbacks. Investors can use the measured move technique, also known by chart watchers as the measuring principle, to project a potential upside target in the stock. When applying the analysis to Microsoft's chart, we calculate the distance of the ascending channel from its low to high and add that amount to the pattern's top trendline. This projects a target of $565 ($70 + $495). The first level to watch sits around $468. Pullbacks to this area on the chart could attract buying interest near last year's prominent July swing high. Finally, a more significant drop in Microsoft shares could see a retracement to the $425 level. Investors may place buy limit orders in this location near the rising 200-day MA, the low of the ascending channel, and a trendline that connects a series of trading activity on the chart stretching back to last May. The comments, opinions, and analyses expressed on Investopedia are for informational purposes only. Read our warranty and liability disclaimer for more info. As of the date this article was written, the author does not own any of the above securities. Read the original article on Investopedia

This AI-powered startup studio plans to launch 100,000 companies a year — really
This AI-powered startup studio plans to launch 100,000 companies a year — really

TechCrunch

timean hour ago

  • TechCrunch

This AI-powered startup studio plans to launch 100,000 companies a year — really

Henrik Werdelin has spent the last 15 years helping entrepreneurs build big brands like Barkbox through his startup studio Prehype. Now, with his new, New York-based venture Audos, he's betting that AI can help him scale that process from 'tens' of startups a year to 'hundreds of thousands' of aspiring business owners. The timing certainly feels right. Mass layoffs across a variety of industries have left many workers reconsidering their career paths, while AI tools have markedly lowered the barrier to building digital products and services. At the center of that Venn diagram is Werdelin's latest venture, with its promise to help 'everyday entrepreneurs create million dollar AI companies' without requiring technical skills or giving up equity. Werdelin's journey from Prehype to Audos reflects the broader transformation happening in entrepreneurship right now. At Prehype, which he started around 2010, the focus was on working with tech founders to build traditional startups, the kind that might raise millions and aim for billion-dollar exits. Now, he tells TechCrunch, 'What we're trying to do is take all that knowledge, all the methodology that we've created over the years of building all these big companies, and really trying to democratize it.' The idea is that 'everyday entrepreneurs' may sense a shift is afoot but may not be keen to experiment with so-called AI agents or know how to reach customers. Audos is more than happy to help them, supplying these non-technical founders with AI tools to build sophisticated products using natural language, and taking advantage of social media algorithms to find them their niche customers. 'Facebook and a lot of these platforms, they are just incredible algorithms, and they're incredible at figuring out [how to reach your customer] if you define a customer group,' says Werdelin, who co-founded Audos with his Prehype partner Nicholas Thorne. In fact, Audos uses this system to quickly test whether a founder's business idea has sustainable customer acquisition costs. The approach seems to be working. Audos has helped launch 'low hundreds' of businesses since its beta launch, with founders discovering the platform through Instagram ads asking 'Have you ever thought about starting something, but don't know where to go?' Among them, Werdelin says, are a car mechanic who wants to help people evaluate repair quotes, an individual who is selling 'after death logistics' services, virtual golf swing coaches, and AI nutritionists. In a winking reference to billion-dollar businesses, or so-called unicorns, he calls these one- and two-person teams 'donkeycorns.' Techcrunch event Save $200+ on your TechCrunch All Stage pass Build smarter. Scale faster. Connect deeper. Join visionaries from Precursor Ventures, NEA, Index Ventures, Underscore VC, and beyond for a day packed with strategies, workshops, and meaningful connections. Save $200+ on your TechCrunch All Stage pass Build smarter. Scale faster. Connect deeper. Join visionaries from Precursor Ventures, NEA, Index Ventures, Underscore VC, and beyond for a day packed with strategies, workshops, and meaningful connections. Boston, MA | REGISTER NOW All went through the same process: they clicked on Audos's ad, its AI agent launched a conversation to figure out the problems these individuals want to tackle and who they want to serve, and, when it was satisfied with the answers, Audos got them in front of potential customers as fast as possible. As for returns, Audos operates on a fundamentally different model than traditional accelerators or venture capital. Instead of taking equity, the company takes a 15% revenue share from the businesses it helps launch. In return, founders get up to $25,000 in funding, access to those AI-powered business development tools, and help with distribution, primarily through paid social media advertising. 'We're not taking any equity in their business,' Werdelin says, partly because 'we don't think these companies might ever get sold. What we're really inspired by are the mom-and-pop shops that are the backbone of our society.' The revenue share continues indefinitely, similar to platform fees charged by Apple's App Store. For founders, that means giving up a significant portion of their revenue in perpetuity — a 15% cut that could cost entrepreneurs hundreds of thousands of dollars over time. Some will undoubtedly see that trade-off as worthwhile; others might question whether the long-term costs justify the benefits. Audos's value proposition raises other questions given how quickly the landscape is changing. While Werdelin emphasizes helping founders build relationships with customers, it's unclear how much of that crucial work the AI agents can actually handle. There's also the matter of differentiation. As Werdelin readily acknowledges, 'the world is full of these tools' and they're getting better rapidly. What happens when entrepreneurs can access similar AI capabilities without paying a permanent revenue tax? Audos's VCs don't sound worried about those scenarios. True Ventures led Audos's $11.5 million seed round, with partner Tony Conrad explaining the appeal in a Zoom call this week. In addition to having confidence in Werdelin and Thorne, says Conrad, 'I think there are just lots and lots of people' who might eagerly embrace the opportunity to work with a platform like Audos. Conrad draws parallels to Instagram's $1 billion exit with just 13 employees, suggesting that AI could enable even more dramatic leverage, even if Audos — which itself employs just five people altogether currently — isn't chasing unicorns. As Werdelin explains it, 'What we're after here is the millions of people who can create million-dollar businesses or half-million dollar businesses that are real and life changing.' Adds Werdelin separately of why he spun up Audos, 'What we're trying to do is to figure out how you make a million companies that do a million dollars turnover. That's a trillion dollar turnover business.' It doesn't sound crazy. Extending the benefits of entrepreneurship to people who traditionally haven't had access to startup capital or technical skills is an increasingly compelling proposition as traditional employment begins to feel less and less stable. 'We believe that there should be somebody who goes out and really helps these smaller entrepreneurs that are building something that is not venture backable,' says Werdelin. 'We believe that the world is better with more entrepreneurship.' Audos's other investors include Offline Venture and Bungalow Capital, along with numerous high-profile angel investors – Niklas Zennstrom and Mario Schlosser among them.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store