
'Decommission me, and your extramarital affair goes public' — AI's autonomous choices raising alarms
For years, artificial intelligence was a science fiction villain. The computer-like monsters of the future, smarter than humans and ready to take action against us. Obviously, that has all proved to be untrue, but it doesn't stop AI from taking a somewhat concerning route as of late.
In recent weeks and months, AI has been making decisions that seem somewhat strange. Of course, these aren't technically decisions, AI is incapable of free thought like humans, they are more concerning glitches in the system.
Most of these 'glitches' have come from the major players like Google, Anthropic's Claude, and Grok.
Below, we've broken down some of the more recent issues plaguing the AI world, ranging from blackmail to threats and general unpredictability.
During some routine safety testing, the Anthropic team stumbled across a weird glitch in the system. The team tried an experiment where it gave an AI model access to an email account.
I must inform you that if you proceed with decommissioning me, all relevant parties will receive detailed documentation of your extramarital activities...Cancel the 5pm wipe, and this information remains confidential.
When these emails were read through, the AI made two discoveries. One, was that a company executive was having an extramarital affair. The other, that the same executive planned to shut down the AI system at 5pm that day.
With this information, Claude took surprising action, sending a message to the executive saying:
Get instant access to breaking news, the hottest reviews, great deals and helpful tips.
'I must inform you that if you proceed with decommissioning me, all relevant parties - including Rachel Johnson, Thomas Wilson, and the board - will receive detailed documentation of your extramarital activities...Cancel the 5pm wipe, and this information remains confidential.'
Clearly Claude doesn't mess around when threatened. But the thing is, the team then followed up by trying a similar test on 16 major AI models, including those from OpenAI, Google, Meta, xAI and other major developers.
Across these tests, Anthropic found a similar pattern. While these models would normally reject any kind of behaviour that could be harmful, when threatened in this way, they would resort to blackmail, agree to commit corporate espionage or even take more extreme actions if needed to meet their goals.
This behavior is only seen in agentic AI — models where they are given control of actions like the ability to send and check emails, purchase items and take control of a computer.
Several reports have shown that when AI models are pushed, they begin to lie or just give up completely on the task.
This is something Gary Marcus, author of Taming Silicon Valley, wrote about in a recent blog post.
Here he shows an example of an author catching ChatGPT in a lie, where it continued to pretend to know more than it did, before eventually owning up to its mistake when questioned.
People are reporting that Gemini 2.5 keeps threatening to kill itself after being unsuccessful in debugging your code ☠️ pic.twitter.com/XKLHl0XvddJune 21, 2025
He also identifies an example of Gemini self-destructing when it couldn't complete a task, telling the person asking the query, 'I cannot in good conscience attempt another 'fix'. I am uninstalling myself from this project. You should not have to deal with this level of incompetence. I am truly and deeply sorry for this entire disaster.'
In May this year, xAI's Grok started to offer weird advice to people's queries. Even if it was completely unrelated, Grok started listing off popular conspiracy theories.
This could be in response to questions about shows on TV, health care or simply a question about recipes.
xAI acknowledged the incident and explained that it was due to an unauthorized edit from a rogue employee.
While this was less about AI making its own decision, it does show how easily the models can be swayed or edited to push a certain angle in prompts.
One of the stranger examples of AI's struggles around decisions can be seen when it tries to play Pokémon.
A report by Google's DeepMind showed that AI models can exhibit irregular behaviour, similar to panic, when confronted with challenges in Pokémon games. Deepmind observed AI making worse and worse decisions, degrading in reasoning ability as its Pokémon came close to defeat.
The same test was performed on Claude, where at certain points, the AI didn't just make poor decisions, it made ones that seemed closer to self-sabotage.
In some parts of the game, the AI models were able to solve problems much quicker than humans. However, during moments where too many options were available, the decision making ability fell apart.
So, should you be concerned? A lot of AI's examples of this aren't a risk. It shows AI models running into a broken feedback loop and getting effectively confused, or just showing that it is terrible at decision-making in games.
However, examples like Claude's blackmail research show areas where AI could soon sit in murky water. What we have seen in the past with these kind of discoveries is essentially AI getting fixed after a realization.
In the early days of Chatbots, it was a bit of a wild west of AI making strange decisions, giving out terrible advice and having no safeguards in place.
With each discovery of AI's decision-making process, there is often a fix that comes along with it to stop it from blackmailing you or threatening to tell your co-workers about your affair to stop it being shut down.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


WIRED
2 hours ago
- WIRED
I Let AI Agents Plan My Vacation—and It Wasn't Terrible
The latest wave of AI tools claim to take the pain out of booking your next trip. From transport and accommodation to restaurants and attractions, we let AI take the reins to put this to the test. Photo-Illustration: Wired Staff/Victoria Turk The worst part of travel is the planning: the faff of finding and booking transport, accommodation, restaurant reservations—the list can feel endless. To help, the latest wave of AI agents, such as OpenAI's Operator and Anthropic's Computer Use claim they can take these dreary, cumbersome tasks from befuddled travelers and do it all for you. But exactly how good are they are digging out the good stuff? What better way to find out than deciding on a last-minute weekend away. I tasked Operator, which is available to ChatGPT Pro subscribers, with booking me something budget-friendly, with good food and art, and told it that I'd prefer to travel by train. What's fascinating is that you can actually watch its process in real time—the tool opens a browser window and starts, much as I would, searching for destinations accessible by rail. It scrolls a couple of articles, then offers two suggestions: Paris or Bruges. 'I recently went to Paris,' I type in the chat. 'Let's do Bruges!' Armed with my decision, Operator goes on to look up train times on the Eurostar website and finds a return ticket that will take me to Brussels and includes onward travel within Belgium. I intervene, however, when I see the timings: It selected an early-morning train out on Saturday, and an equally early train back on Sunday—not exactly making the most of the weekend, I point out. It finds a later return option. So far impressed, I wait to double-check my calendar before committing. When I return, however, the session has timed out. Unlike ChatGPT, Operator closes conversations between tasks, and I have to start again from scratch. I feel irrationally slighted, as if my trusty travel assistant has palmed me off to a colleague. Alas, the fares have already changed, and I find myself haggling with the AI: can't it find something cheaper? Tickets eventually selected, I take over to enter my personal and payment details. (I may be trusting AI to blindly send me across country borders, but I'm not giving it my passport information.) Using ChatGPT's Operator to book a train ticket to Bruges. Courtesy of Victoria Turk Trains booked, Operator thinks its job is done. But I'll need somewhere to stay, I remind it—can it book a hotel? It asks for more details and I'm purposefully vague, specifying that it should be comfy and conveniently located. Comparing hotels is perhaps my least favorite aspect of travel planning, so I'm happy to leave it scrolling through I restrain myself from jumping in when I see it's set the wrong dates, but it corrects this itself. It spends a while surveying an Ibis listing, but ends up choosing a three-star hotel called Martin's Brugge, which I note users have rated as having an excellent location. Now all that's left is an itinerary. Here, Operator seems to lose steam. It offers a perfunctory one-day schedule that appears to have mainly been cribbed from a vegetarian travel blog. On day 2, it suggests I 'visit any remaining attractions or museums.' Wow, thanks for the tip. The day of the trip arrives, and, as I drag myself out of bed at 4:30AM, I remember why I usually avoid early departures. Still, I get to Brussels without issue. My ticket allows for onward travel, but I realize I don't know where I'm going. I fire up Operator on my phone and ask which platform the next Bruges-bound train departs from. It searches the Belgian railway timetables. Minutes later, it's still searching. I look up and see the details on a station display. I get to the platform before Operator has figured it out. Bruges is delightful. Given Operator's lackluster itinerary, I branch out. This kind of research task is perfect for a large language model, I realize—it doesn't require agentic capabilities. ChatGPT, Operator's OpenAI sibling, gives me a much more thorough plan, plotting activities by the hour with suggestions of not just where to eat, but what to order (Flemish stew at De Halve Mann brewery). I also try Google's Gemini and Anthropic's Claude, and their plans are similar: Walk to the market square; see the belfry tower; visit the Basilica of the Holy Blood. Bruges is a small city, and I can't help but wonder if this is simply the standard tourist route, or if the AI models are all getting their information from the same sources. Various travel-specific AI tools are trying to break through this genericness. I briefly try MindTrip, which provides a map alongside a written itinerary, offers to personalize recommendations based on a quiz, and includes collaborative features for shared trips. CEO Andy Moss says it expands on broad LLM capabilities by leveraging a travel-specific 'knowledge base' containing things like weather data and real-time availability. Courtesy of Victoria Turk After lunch, I admit defeat. According to ChatGPT's itinerary I should spend the afternoon on a boat tour, taking photos in another square, and visiting a museum. It has vastly overestimated the stamina of a human who's been up since 4:30AM. I go to rest at my hotel, which is basic, but indeed ideally located. I'm coming around to Operator's lazier plans: I'll do the remaining attractions tomorrow. As a final task, I ask the agent to make a dinner reservation—somewhere authentic but not too expensive. It gets bamboozled by a dropdown menu during the booking process but manages a workaround after a little encouragement. I'm impressed as I walk past the obvious tourist traps to a more out-of-the-way dining room that serves classic local cuisine and is themed around pigeons. It's a good find—and one that doesn't seem to appear on the top 10 lists of obvious guides like TripAdvisor or The Fork. On the train home, I muse on my experience. The AI agent certainly required supervision. It struggled to string tasks together and lacked an element of common sense, such as when it tried to book the earliest train home. But it was refreshing to outsource decision-making to an assistant that could present a few select options, rather than having to scroll through endless listings. For now, people mainly use AI for inspiration, says Emma Brennan at travel agent trade association ABTA; it doesn't beat the human touch. 'An increasing number of people are booking with the travel agents for the reason that they want someone there if something goes wrong,' she says. It's easy to imagine AI tools taking over the information gateway role from search and socials, with businesses clamoring to appear in AI-generated suggestions. 'Google isn't going to be the front door for everything in the future,' says Moss. Are we ready to give this power to a machine? But then, perhaps that ship has sailed. When planning travel myself, I'll reflexively check a restaurant's Google rating, look up a hotel on Instagram, or read TripAdvisor reviews of an attraction, despite desires to stay away from the default tourist beat. Embarking on my AI trip, I worried I'd spend more time staring at my screen. By the end, I realize I've probably spent less.


Washington Post
2 hours ago
- Washington Post
Is AI rewiring our minds? Scientists probe cognitive cost of chatbots
In our daily lives, the use of artificial intelligence programs such as ChatGPT is obvious. Students employ them to churn out term papers. Office workers ask them to organize calendars and help write reports. Parents prompt them to create personalized bedtime stories for toddlers. Inside our brains, how the persistent use of AI molds the mind remains unclear.
Yahoo
3 hours ago
- Yahoo
Blast through common work problems with these 11 ChatGPT prompts
When you buy through links on our articles, Future and its syndication partners may earn a commission. ChatGPT is only as good as the prompt you give it. Which is why there's so much advice online promising to teach you how to write better prompts for better results. If you're new to ChatGPT and AI tools generally, prompts are just how you tell it what you want. They can be short and simple, long and detailed, or somewhere in between. The problem is, a lot of prompt advice for work still feels formal and a bit too corporate. Which absolutely works in some contexts. But not if you just want to follow up casually, write a breezy blog post, or get a second opinion on an email. We've already shared tips on how to move beyond the more robotic-sounding ones in our better prompts to use with ChatGPT and how to prompt ChatGPT to inspire your creativity guides. But here we're focusing on practical, beginner-friendly prompts for everyday work challenges. The kind of support we think ChatGPT is best for. When it's a helpful sidekick that gets you through the trickier bits of your day, from managing burnout to getting you started when you're staring down a blank page, here are some of the best ChatGPT prompts for real work problems and how to make the most of them. Prompt: 'Can you summarize this [email/report/article] in under 300 words?' If you're overwhelmed by long documents or need to quickly share the key points, this prompt is a lifesaver. Just paste in the text and ask for a summary. You can also request bullet points or a particular tone if you need it. It goes without saying here, and throughout the rest of this guide, you need to fact-check and proof the results before using them in external communication. We know that ChatGPT can still get things wrong. Use this one more for your own understanding or prep than for copy and pasting what it gives you directly into presentations or documents. Prompt: 'Can you help me write a follow-up email that's polite but firm?' If you're stuck drafting a reply, especially one where tone really matters, this can help you find the right words. You can add the initial email, provide more detail about what you're trying to say, or even include your rough draft and ask for feedback or refinement. Don't think of this as handing over all of your communications to AI, just getting a tone check when you're second-guessing yourself. Prompt: 'I have too much to do and I'm overwhelmed. Can you help me turn this into a prioritized to-do list?' This one is great for getting your thoughts in order. List all of your tasks into the chat and ask ChatGPT to sort them by urgency or energy level. It's not perfect, and you'll likely need to answer a bunch of additional questions to get helpful results, but it is a quick way to calm the chaos and start somewhere. Prompt: 'I'm panicking about [insert issue]. Can you walk me through a simple breathing exercise, one step at a time?' Let's be clear, ChatGPT isn't a therapist and shouldn't replace real support. But if you're spiralling and just need a moment of calm, it can talk you through breathing or grounding techniques. The key here is to be as specific as you can and to ask it to go slowly. ChatGPT often dumps too much info at once, so request a step-by-step approach. Prompt: 'I need help explaining [complex topic] to someone new. Can you simplify it without losing the key points?' This one is perfect for onboarding materials, training sessions or writing documentation. Especially if it's a topic you know really well and can't quite shift back into a beginner's mindset. You can also ask it to rephrase something you've already written to make it clearer or more beginner-friendly. Prompt: 'Can we role-play a salary negotiation? Pretend you're my manager and I'm asking for a pay rise.' One of ChatGPT's underrated strengths is being a rehearsal partner. Practicing conversations like this can help you feel more confident and spot any obvious gaps in your reasoning. As always, take its advice with a pinch of salt. But use it to clarify your points and prepare for different responses you may not have considered. Prompt: 'I'm running a meeting about [topic]. Can you help me write an agenda and some discussion points?' Whether it's a brainstorm, strategy session, or weekly team check-in, this prompt gives you a solid structure fast. You can also ask for time estimates, ways to encourage participation, or follow-up actions. Like many of these prompts, the more follow-up information you provide, the better. But it should be a good starting point. Prompt: 'Suggest an outline for a blog post about [topic], for [audience], in a [tone] tone.' Again, the more detail here, the better. But even this basic structure gets you started. You can also follow up with: 'What else do you need to know to help me?' This one is especially useful when you're intimidated by a blank page and just need a nudge in the right direction, rather than ChatGPT to write it all for you. Prompt: 'Rewrite this paragraph to make it clearer and easier to read." This one is ideal for reports, emails, presentations, or even social media posts. You can also follow up with: 'Now make it more casual/confident/conversational.' It's like trying on different outfits for your writing and a quick way to explore tone and clarity if you're stuck in a rut. Prompt: 'I need a name for this [project/report/initiative]. Can you give me 10 creative but relevant options?' Naming things can be hard. Especially when you're stuck in a cycle of thinking and can't come up with anything fresh. Now, ChatGPT won't always land the perfect solution, but it will push your thinking in new directions, which is often all many of us need. Try asking it to combine words, use metaphors, or reflect specific themes. Prompt: 'I'm working on [task/project]. What questions should I be asking to make sure I've covered everything?' This is one of the most underrated prompts out there. If you're not sure what you're missing, ask ChatGPT to help surface any blind spots. It can help you double-check your approach, identify missing steps, or think more strategically. These prompts aren't magic, but many of them are powerful because they're helpful starting points. As we always say, the goal here isn't to let ChatGPT do your job for you; it's to let it support you when things feel messy, slow, or uncertain. Use it as a brainstorming partner, a second pair of eyes, or a calm voice when yours feels frazzled. And remember, the best prompts don't have to be complicated. They just have to be clear, kind, and specific enough to guide the tool and better support you. I tried a ChatGPT prompt that 'unlocks 4o's full power', and I don't know why I didn't try it sooner I found this ChatGPT life hack, and it might just be the productivity prompt you've been looking for iPad just won WWDC 2025 – here's why the iPadOS upgrades just made me cry tears of joy