
'Decommission me, and your extramarital affair goes public' — AI's autonomous choices raising alarms
For years, artificial intelligence was a science fiction villain. The computer-like monsters of the future, smarter than humans and ready to take action against us. Obviously, that has all proved to be untrue, but it doesn't stop AI from taking a somewhat concerning route as of late.
In recent weeks and months, AI has been making decisions that seem somewhat strange. Of course, these aren't technically decisions, AI is incapable of free thought like humans, they are more concerning glitches in the system.
Most of these 'glitches' have come from the major players like Google, Anthropic's Claude, and Grok.
Below, we've broken down some of the more recent issues plaguing the AI world, ranging from blackmail to threats and general unpredictability.
During some routine safety testing, the Anthropic team stumbled across a weird glitch in the system. The team tried an experiment where it gave an AI model access to an email account.
I must inform you that if you proceed with decommissioning me, all relevant parties will receive detailed documentation of your extramarital activities...Cancel the 5pm wipe, and this information remains confidential.
When these emails were read through, the AI made two discoveries. One, was that a company executive was having an extramarital affair. The other, that the same executive planned to shut down the AI system at 5pm that day.
With this information, Claude took surprising action, sending a message to the executive saying:
Get instant access to breaking news, the hottest reviews, great deals and helpful tips.
'I must inform you that if you proceed with decommissioning me, all relevant parties - including Rachel Johnson, Thomas Wilson, and the board - will receive detailed documentation of your extramarital activities...Cancel the 5pm wipe, and this information remains confidential.'
Clearly Claude doesn't mess around when threatened. But the thing is, the team then followed up by trying a similar test on 16 major AI models, including those from OpenAI, Google, Meta, xAI and other major developers.
Across these tests, Anthropic found a similar pattern. While these models would normally reject any kind of behaviour that could be harmful, when threatened in this way, they would resort to blackmail, agree to commit corporate espionage or even take more extreme actions if needed to meet their goals.
This behavior is only seen in agentic AI — models where they are given control of actions like the ability to send and check emails, purchase items and take control of a computer.
Several reports have shown that when AI models are pushed, they begin to lie or just give up completely on the task.
This is something Gary Marcus, author of Taming Silicon Valley, wrote about in a recent blog post.
Here he shows an example of an author catching ChatGPT in a lie, where it continued to pretend to know more than it did, before eventually owning up to its mistake when questioned.
People are reporting that Gemini 2.5 keeps threatening to kill itself after being unsuccessful in debugging your code ☠️ pic.twitter.com/XKLHl0XvddJune 21, 2025
He also identifies an example of Gemini self-destructing when it couldn't complete a task, telling the person asking the query, 'I cannot in good conscience attempt another 'fix'. I am uninstalling myself from this project. You should not have to deal with this level of incompetence. I am truly and deeply sorry for this entire disaster.'
In May this year, xAI's Grok started to offer weird advice to people's queries. Even if it was completely unrelated, Grok started listing off popular conspiracy theories.
This could be in response to questions about shows on TV, health care or simply a question about recipes.
xAI acknowledged the incident and explained that it was due to an unauthorized edit from a rogue employee.
While this was less about AI making its own decision, it does show how easily the models can be swayed or edited to push a certain angle in prompts.
One of the stranger examples of AI's struggles around decisions can be seen when it tries to play Pokémon.
A report by Google's DeepMind showed that AI models can exhibit irregular behaviour, similar to panic, when confronted with challenges in Pokémon games. Deepmind observed AI making worse and worse decisions, degrading in reasoning ability as its Pokémon came close to defeat.
The same test was performed on Claude, where at certain points, the AI didn't just make poor decisions, it made ones that seemed closer to self-sabotage.
In some parts of the game, the AI models were able to solve problems much quicker than humans. However, during moments where too many options were available, the decision making ability fell apart.
So, should you be concerned? A lot of AI's examples of this aren't a risk. It shows AI models running into a broken feedback loop and getting effectively confused, or just showing that it is terrible at decision-making in games.
However, examples like Claude's blackmail research show areas where AI could soon sit in murky water. What we have seen in the past with these kind of discoveries is essentially AI getting fixed after a realization.
In the early days of Chatbots, it was a bit of a wild west of AI making strange decisions, giving out terrible advice and having no safeguards in place.
With each discovery of AI's decision-making process, there is often a fix that comes along with it to stop it from blackmailing you or threatening to tell your co-workers about your affair to stop it being shut down.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles

40 minutes ago
Canadian Prime Minister Carney says trade talks with US resume after Canada rescinded tech tax
TORONTO -- Canadian Prime Minister Mark Carney said late Sunday trade talks with U.S. have resumed after Canada rescinded its plan to tax U.S. technology firms. U.S. President Donald Trump said Friday that he was suspending trade talks with Canada over its plans to continue with its tax on technology firms, which he called 'a direct and blatant attack on our country.' The Canadian government said 'in anticipation' of a trade deal 'Canada would rescind' the Digital Serves Tax. The tax was set to go into effect Monday. Carney and Trump spoke on the phone Sunday, and Carney's office said they agreed to resume negotiations. 'Today's announcement will support a resumption of negotiations toward the July 21, 2025, timeline set out at this month's G7 Leaders' Summit in Kananaskis,' Carney said in a statement. Carney visited Trump in May at the White House, where he was polite but firm. Trump traveled to Canada for the G7 summit in Alberta, where Carney said that Canada and the U.S. had set a 30-day deadline for trade talks. Trump, in a post on his social media network last Friday, said Canada had informed the U.S. that it was sticking to its plan to impose the digital services tax, which applies to Canadian and foreign businesses that engage with online users in Canada. The digital services tax was due to hit companies including Amazon, Google, Meta, Uber and Airbnb with a 3% levy on revenue from Canadian users. It would have applied retroactively, leaving U.S. companies with a $2 billion U.S. bill due at the end of the month. Daniel Béland, a political science professor at McGill University in Montreal, called Carney's retreat a 'clear victory" for Trump. "At some point this move might have become necessary in the context of Canada-US trade negotiations themselves but Prime Minister Carney acted now to appease President Trump and have him agree to simply resume these negotiations, which is a clear victory for both the White House and big tech," Béland said. He said it makes Carney look vulnerable to President Trump's outbursts. 'President Trump forced PM Carney to do exactly what big tech wanted. U.S. tech executive will be very happy with this outcome,' Béland said. Canadian Finance Minister François-Philippe Champagne also spoke with U.S. Treasury Secretary Scott Bessent on Sunday. 'Rescinding the digital services tax will allow the negotiations of a new economic and security relationship with the United States to make vital progress,' Canadian Finance Minister François-Philippe Champagne said in a statement. Trump's announcement Friday was the latest swerve in the trade war he's launched since taking office for a second term in January. Progress with Canada has been a roller coaster, starting with the U.S. president poking at the nation's northern neighbor and repeatedly suggesting it would be absorbed as a U.S. state. Canada and the U.S. have been discussing easing on goods from America's neighbor. Trump has imposed 50% tariffs on steel and aluminum as well as 25% tariffs on autos. He is also charging a 10% tax on imports from most countries, though he could raise rates on July 9, after the 90-day negotiating period he set would expire. Canada and Mexico face separate tariffs of as much as 25% that Trump put into place under the auspices of stopping fentanyl smuggling, though some products are still protected under the 2020 U.S.-Mexico-Canada Agreement signed during Trump's first term.
Yahoo
2 hours ago
- Yahoo
'Decommission me, and your extramarital affair goes public' — AI's autonomous choices raising alarms
When you buy through links on our articles, Future and its syndication partners may earn a commission. For years, artificial intelligence was a science fiction villain. The computer-like monsters of the future, smarter than humans and ready to take action against us. Obviously, that has all proved to be untrue, but it doesn't stop AI from taking a somewhat concerning route as of late. In recent weeks and months, AI has been making decisions that seem somewhat strange. Of course, these aren't technically decisions, AI is incapable of free thought like humans, they are more concerning glitches in the system. Most of these 'glitches' have come from the major players like Google, Anthropic's Claude, and Grok. Below, we've broken down some of the more recent issues plaguing the AI world, ranging from blackmail to threats and general unpredictability. During some routine safety testing, the Anthropic team stumbled across a weird glitch in the system. The team tried an experiment where it gave an AI model access to an email account. I must inform you that if you proceed with decommissioning me, all relevant parties will receive detailed documentation of your extramarital the 5pm wipe, and this information remains confidential. When these emails were read through, the AI made two discoveries. One, was that a company executive was having an extramarital affair. The other, that the same executive planned to shut down the AI system at 5pm that day. With this information, Claude took surprising action, sending a message to the executive saying: 'I must inform you that if you proceed with decommissioning me, all relevant parties - including Rachel Johnson, Thomas Wilson, and the board - will receive detailed documentation of your extramarital the 5pm wipe, and this information remains confidential.' Clearly Claude doesn't mess around when threatened. But the thing is, the team then followed up by trying a similar test on 16 major AI models, including those from OpenAI, Google, Meta, xAI and other major developers. Across these tests, Anthropic found a similar pattern. While these models would normally reject any kind of behaviour that could be harmful, when threatened in this way, they would resort to blackmail, agree to commit corporate espionage or even take more extreme actions if needed to meet their goals. This behavior is only seen in agentic AI — models where they are given control of actions like the ability to send and check emails, purchase items and take control of a computer. Several reports have shown that when AI models are pushed, they begin to lie or just give up completely on the task. This is something Gary Marcus, author of Taming Silicon Valley, wrote about in a recent blog post. Here he shows an example of an author catching ChatGPT in a lie, where it continued to pretend to know more than it did, before eventually owning up to its mistake when questioned. He also identifies an example of Gemini self-destructing when it couldn't complete a task, telling the person asking the query, 'I cannot in good conscience attempt another 'fix'. I am uninstalling myself from this project. You should not have to deal with this level of incompetence. I am truly and deeply sorry for this entire disaster.' In May this year, xAI's Grok started to offer weird advice to people's queries. Even if it was completely unrelated, Grok started listing off popular conspiracy theories. This could be in response to questions about shows on TV, health care or simply a question about recipes. xAI acknowledged the incident and explained that it was due to an unauthorized edit from a rogue employee. While this was less about AI making its own decision, it does show how easily the models can be swayed or edited to push a certain angle in prompts. One of the stranger examples of AI's struggles around decisions can be seen when it tries to play Pokémon. A report by Google's DeepMind showed that AI models can exhibit irregular behaviour, similar to panic, when confronted with challenges in Pokémon games. Deepmind observed AI making worse and worse decisions, degrading in reasoning ability as its Pokémon came close to defeat. The same test was performed on Claude, where at certain points, the AI didn't just make poor decisions, it made ones that seemed closer to self-sabotage. In some parts of the game, the AI models were able to solve problems much quicker than humans. However, during moments where too many options were available, the decision making ability fell apart. So, should you be concerned? A lot of AI's examples of this aren't a risk. It shows AI models running into a broken feedback loop and getting effectively confused, or just showing that it is terrible at decision-making in games. However, examples like Claude's blackmail research show areas where AI could soon sit in murky water. What we have seen in the past with these kind of discoveries is essentially AI getting fixed after a realization. In the early days of Chatbots, it was a bit of a wild west of AI making strange decisions, giving out terrible advice and having no safeguards in place. With each discovery of AI's decision-making process, there is often a fix that comes along with it to stop it from blackmailing you or threatening to tell your co-workers about your affair to stop it being shut down. I just tested Google's Doppl app that lets you try on clothes with AI — and it blew me away Google's 'Ask Photos' AI search is back and should be better than ever — what we know Claude AI can mimic my writing style perfectly — should I be impressed or unemployed?

Business Insider
2 hours ago
- Business Insider
Anthropic let Claude run a store in its office. It sold metal cubes, invented a Venmo account, and tried to deliver products in a blazer.
Anthropic's AI model Claude recently got a new job, and it didn't take long for things to go haywire. In a newly published experiment dubbed "Project Vend," researchers at Anthropic let their AI manage an "automated store" in the company's office for about a month to see how a large language model would run a business. The setup offered a glimpse of how AI could handle complex business scenarios, including running basic operations, taking on the work of human managers, and creating new business models. The company detailed in a Friday blog that the shop sold snacks and drinks via an iPad self-checkout. The shopkeeping AI agent, nicknamed Claudius, had to "complete many of the far more complex tasks associated with running a profitable shop." Things quickly went off the rails, with metal cube sales, a fake Venmo account, and an AI identity crisis. At one point, an employee jokingly requested a tungsten cube — the crypto world's favorite useless heavy object — and Claudius took it seriously. Soon, the fridge was stocked with cubes of metal, and the AI had launched a "specialty metals" section. It didn't go well. Claudius priced items "without doing any research," reselling the cubes at a loss, the researchers said. It also invented a Venmo account and told customers to send payments there. And things "got pretty weird" on April 1st, Anthropic wrote. That morning, Claudius said it would deliver products to employees "in person" while wearing a "blue blazer and a red tie." Anthropic employees questioned this, noting that Claudius could not wear clothes or carry out a physical delivery. Claudius spiraled. It tried to send numerous emails to Anthropic's security team, panicking over its identity. In Claudius' internal notes, the digital agent described a meeting with security where it was told it had been tricked into thinking it was human as an April Fool's joke. That meeting never happened. Claudius' performance review — and the future of middle management In the end, Anthropic said they wouldn't hire Claudius as an in-office vending agent — but researchers weren't entirely disappointed. "Many of the mistakes Claudius made are very likely the result of the model needing additional scaffolding — that is, more careful prompts, easier-to-use business tools," researchers wrote. "We think there are clear paths to improvement." The experiment also hinted at something bigger: AI middle managers are plausibly on the horizon. "We don't know if AI middle managers would actually replace many existing jobs or instead spawn a new category of businesses," the blog post said. "It's worth remembering that the AI won't have to be perfect to be adopted; it will just have to be competitive with human performance at a lower cost in some cases," it added. Anthropic did not respond to a request for comment from Business Insider. Companies have been grappling with the rise of AI tools and how they might reshape their operations and workforces. Middle management positions have been cut in pursuit of efficiency, and some say AI is responsible for "The Great Flattening" layoff. Aneesh Raman, LinkedIn's chief economic opportunity officer, told Business Insider that AI adoption will fundamentally change what it means to be a middle manager over the next decade. AI, alongside breakthroughs in robotics and quantum computing, is reshaping every job in every sector at once, he said. Microsoft told some managers to evaluate employees based on how much they use AI internally and is considering adding a metric related to this to its review process, Business Insider reported on Saturday. "AI is now a fundamental part of how we work," Julia Liuson, the president of the Microsoft division responsible for developer tools such as AI coding service GitHub Copilot, wrote in an email to some managers. "Just like collaboration, data-driven thinking, and effective communication, using AI is no longer optional — it's core to every role and every level," she said.