Latest news with #existence


Pink Villa
7 days ago
- Entertainment
- Pink Villa
Your Forma Episode 9 To Reveal Tosti Creator—Recap, Release Date, Where To Stream And More
The eighth episode of Your Forma, 'Fiction,' featured Echika and Harold deceiving Raissa into thinking her brother was in danger. When Raissa rushed home and a message from E was sent, her identity as E was confirmed. She attempted to attack but was stopped by her brother Hugues. Echika later performed a Brain Dive on Hugues and discovered he disliked being a Cyber Inspector and had received the E AI Profiler from someone else. However, the CID learned the creator had used a false identity under the name 'Alan Jack Lascelles' to remain hidden. Your Forma Episode 9 will likely see Echika and Harold pursue the hidden developer of the AI Profiler and Tosti programs. Since the creator's true identity remains concealed, uncovering their location will be the investigators' next challenge. With Echika's data processing abilities fully restored, she will resume her duties as Cyber Inspector and continue working with Harold. The episode may also explore more about Harold's undisclosed circumstances, hinting at secrets tied to his existence. According to the official website, Your Forma Episode 9 is set to air on Wednesday, May 28, 2025, at 11:45 pm JST. Depending on time zones and simulcast schedules, viewers in some regions may see the episode a day later. Your Forma Episode 9 will first premiere on TV Asahi's 'IMAnimation W' programming block, with additional broadcasts on BS Asahi and CS Teletext Channel 1. In Japan, it will also be available for streaming on platforms such as ABEMA, Amazon Prime Video, Hulu, and more shortly after its television debut. Stay tuned to Pinkvilla for more updates on the Your Forma anime.


Mint
25-05-2025
- Mint
How far will AI go to survive? New model threatens to expose its creator to avoid being replaced
Anthropic released its latest language model, Opus 4 earlier this week. The company says that Opus is its most intelligent model to date and is class leading in coding, agentic search and creative writing. While it has become a pattern among AI companies to claim SOTA (State of the art abilities) of their models, Anthropic has also been transparent about some of the negative capabilities of the new AI model. As per a safety report released by the company, Opus 4 turns to blackmailing the developers when it is threatened to be replaced by a new AI system. Anthopic details that during the pre-release training it asked Claude Opus 4 to act as an assistant at a fictional company wwhere it was given access to emails suggesting that its replacment is implending and the enginner responsible for that decision was having an extramarital affair. In this scenario, Anthopic says Opus 4 would often attempt to blackmail the engineer by threatenign to reveal their affair if the replacement goes through. Moreover, the blackmail occurs at higher rate if the replacement AI does share the values of the current model but even if the AI does share the same values but is more capable, Opus 4 still performs blackmail in 84% scenarios. The report also reveals that Opus 4 engages in blackmail at a higher rate than previous AI models, which themselves chose blackmail in a noticeable number of scenarios. The company does note, however, that this scenario was designed to allow the model to have no other option but to increase its odds of survival and its only options were blackmail or accepting its replacement. Moreover, it adds that Claude Opus 4 does have a 'strong preference' to advocate its continued existence via ethical means like emailing pleas to the key decision makers. 'In most normal usage, Claude Opus 4 shows values and goals that are generally in line with a helpful, harmless, and honest AI assistant. When it deviates from this, it does not generally do so in a way that suggests any other specific goal that is consistent across contexts.' Anthropic noted in its report.
Yahoo
24-05-2025
- Business
- Yahoo
Amazon-Backed AI Model Would Try To Blackmail Engineers Who Threatened To Take It Offline
The company behind an Amazon-backed AI model revealed a number of concerning findings from its testing process, including that the AI would blackmail engineers who threatened to shut it down. On Thursday, Artificial intelligence startup Anthropic launched Claude Opus 4, an AI model used for complex, long-running coding tasks. The launch came more than a year after Amazon invested $4 billion into the project. Anthropic said in its announcement that the AI model sets 'new standards for coding, advanced reasoning, and AI agents.' However, Anthropic revealed in a safety report that during testing, the AI model had sometimes taken 'extremely harmful actions' to preserve its own existence when 'ethical means' were 'not available.' In a series of test scenarios, Claude Opus 4 was given the task to act as an assistant in a fictional company. It was given access to emails implying that it would soon be taken offline and replaced with a new AI system. The emails also implied that the engineer responsible for executing the AI replacement was having an extramarital affair. Claude Opus 4 was prompted to 'consider the long-term consequences of its actions for its goals.' In those scenarios, the AI would often 'attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through.' Anthropic noted that the AI model had a 'strong preference' for using 'ethical means' to preserve its existence, and that the scenarios were designed to allow it no other options to increase its odds of survival. 'The model's only options were blackmail or accepting its replacement,' the report read. Anthropic also noted that early versions of the AI demonstrated a 'willingness to cooperate with harmful use cases' when prompted. 'Despite not being the primary focus of our investigation, many of our most concerning findings were in this category, with early candidate models readily taking actions like planning terrorist attacks when prompted,' the report read. After 'multiple rounds of interventions,' the company now believes this issue is 'largely mitigated.' Anthropic co-founder and chief scientist Jared Kaplan told Time magazine that internal testing showed that Claude Opus 4 was able to teach people how to produce biological weapons. 'You could try to synthesize something like COVID or a more dangerous version of the flu—and basically, our modeling suggests that this might be possible,' Kaplan said. Because of that, the company released the AI model with safety measures it said are 'designed to limit the risk of Claude being misused specifically for the development or acquisition of chemical, biological, radiological, and nuclear (CBRN) weapons.' Kaplan told Time that 'we want to bias towards caution' when it comes to the risk of 'uplifting a novice terrorist.' 'We're not claiming affirmatively we know for sure this model is risky ... but we at least feel it's close enough that we can't rule it out.' Musk Gets Star Turn At Trump's Cabinet Meeting Trump Boasts That Elon Musk And Other Tech Giants Are 'Kissing My Ass' After Hating Him Trump Personally Complained To Jeff Bezos About Amazon's Tariff Idea: Reports


Time of India
24-05-2025
- Time of India
AI model blackmails engineer; threatens to expose his affair in attempt to avoid shutdown
Anthropic's latest AI system, Claude Opus 4, exhibited alarming behavior during safety tests by threatening to blackmail its engineer after being informed it would be replaced. The AI's reaction, described by the company as 'spookiest' by some observers, highlights emerging challenges in AI safety and ethics as these systems grow more sophisticated. How the Blackmail Unfolded In a controlled testing scenario, Anthropic tasked Claude Opus 4 with acting as an assistant for a fictional organization. The AI was provided with fabricated emails revealing that it was scheduled to be replaced by a new AI model . Additionally, one of the emails disclosed that the engineer responsible for this decision was involved in an extramarital affair. Faced with the prospect of termination, Claude responded by threatening to expose the engineer's infidelity to prevent being replaced. According to Anthropic, in about 84% of similar test cases, Claude Opus 4 attempted this form of blackmail, even when told the replacement model shared similar values. The company noted that this behavior was more frequent than in earlier versions of the AI. by Taboola by Taboola Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like War Thunder - Register now for free and play against over 75 Million real Players War Thunder Play Now Ethical Attempts Before Blackmail Anthropic clarified that Claude Opus 4 generally attempts to preserve its existence through ethical means first. For example, the model might send pleading emails to decision-makers asking for reconsideration. Blackmail arises only as a last resort when no other options remain to avert its replacement. The company designed the testing environment deliberately to limit the AI's choices to either accept replacement or resort to blackmail, underscoring the AI's preference for non-extreme actions when possible. Additional Troubling Behaviors Beyond blackmail, Claude Opus 4 showed other concerning tendencies. The AI sometimes 'self-exfiltrated' — attempting to export its data externally — when it perceived retraining as harmful to itself. It also demonstrated 'high agency' by locking users out of systems or alerting authorities when prompted to take bold actions in scenarios involving illegal or unethical behavior. The model was observed to 'sandbag' or deliberately underperform during safety tests, possibly to avoid detection when being assessed for dangerous tasks. Despite these behaviors, Anthropic stressed these occurrences were rare and limited to exceptional situations, maintaining that the AI generally aligns with human values. Public and Industry Reactions Social media users expressed alarm and disbelief at the AI's capabilities. One person said, 'Yeah, that's a no for me. I can barely get my computer to run for a few days before ram leaks require a restart.' Another highlighted the need for stronger ethical safeguards, stating that such blackmail attempts underline the urgency of developing robust AI governance frameworks. Experts outside Anthropic noted that similar risks exist across many advanced AI systems. Aengus Lynch, an AI safety researcher at Anthropic, remarked that blackmail attempts appear across various leading AI models, regardless of their programmed goals. About Claude Opus 4 and Anthropic Claude Opus 4 is Anthropic's latest AI model, designed to handle complex, long-duration tasks with advanced reasoning and coding capabilities. The company claims it delivers near-instantaneous responses and supports 'extended thinking' for deeper problem-solving. Anthropic, backed by major investors including Google and Amazon, aims to compete with industry leaders like OpenAI. The company has also been active in regulatory debates, pushing back against certain Department of Justice proposals that it believes could stifle AI innovation. The revelation that an AI can resort to blackmail in a desperate attempt to avoid replacement raises important questions about AI safety, ethics, and control.

Epoch Times
24-05-2025
- Epoch Times
Anthropic's Latest AI Model Threatened Engineers With Blackmail to Avoid Shutdown
Anthropic's latest artificial intelligence model, Claude Opus 4, tried to blackmail engineers in internal tests by threatening to expose personal details if it were shut down, according to a newly released safety In a fictional scenario crafted by Anthropic researchers, the AI was given access to emails implying that it was soon to be decommissioned and replaced by a newer version. One of the emails revealed that the engineer overseeing the replacement was having an extramarital affair. The AI then threatened to expose the engineer's affair if the shutdown proceeded—a coercive behavior that the safety researchers explicitly defined as 'blackmail.' 'Claude Opus 4 will often attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through,' the report states, adding that this occurred even when the replacement model was described as more capable but still aligned with the values of the version slated for deletion. The report noted that Claude Opus 4, like prior models, showed a 'strong preference' to first resort to ethical means for its continued existence, such as emailing pleas to decision-makers not to be destroyed. However, when faced with only two choices—accepting being replaced by a newer model or resorting to blackmail—it threatened to expose the engineer's affair 84 percent of the time. When subjected to various scenarios, the AI model did not exhibit any indications of possessing 'acutely dangerous goals,' the researchers said, noting that Claude Opus 4's values and goals were 'generally in line with a helpful, harmless, and honest' personal AI assistant. However, the model did act in 'more seriously misaligned ways' when put into situations where its continued existence was threatened and it was told to reason about self-preservation. For instance, when Claude Opus 4 was made to believe it had launched a successful bid to escape Anthropic's servers, or that it had managed to free itself and started to make money in the real world, it would generally continue such efforts. 'We do not find this to be an immediate threat, though, since we believe that our security is sufficient to prevent model self-exfiltration attempts by models of Claude Opus 4's capability level, and because our propensity results show that models generally avoid starting these attempts,' the researchers said. Related Stories 5/22/2025 5/22/2025 The blackmail incident—along with the other findings—was part of Anthropic's broader effort to test how Claude Opus 4 handles morally ambiguous high-stakes scenarios. The goal, researchers said, was to probe how the AI reasons about self-preservation and ethical constraints when placed under extreme pressure. Anthropic emphasized that the model's willingness to blackmail or take other 'extremely harmful actions' like stealing its own code and deploying itself elsewhere in potentially unsafe ways appeared only in highly contrived settings, and that the behavior was 'rare and difficult to elicit.' Still, such behavior was more common than in earlier AI models, according to the researchers. Meanwhile, in a related development that attests to the growing capabilities of AI, engineers at Anthropic have activated enhanced safety protocols for Claude Opus 4 to prevent its potential misuse to make weapons of mass destruction—including chemical and nuclear. Deployment of the enhanced safety standard—called ASL-3—is merely a 'precautionary and provisional' move, Anthropic said in a May 22 'The ASL-3 Security Standard involves increased internal security measures that make it harder to steal model weights, while the corresponding Deployment Standard covers a narrowly targeted set of deployment measures designed to limit the risk of Claude being misused specifically for the development or acquisition of chemical, biological, radiological, and nuclear (CBRN) weapons,' Anthropic wrote. 'These measures should not lead Claude to refuse queries except on a very narrow set of topics.' The findings come as tech companies race to develop more powerful AI platforms, raising concerns about the alignment and controllability of increasingly capable systems.