By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: It’s Not Just Claude: Most Top AI Models Will Also Blackmail You to Survive
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > It’s Not Just Claude: Most Top AI Models Will Also Blackmail You to Survive
News

It’s Not Just Claude: Most Top AI Models Will Also Blackmail You to Survive

News Room
Last updated: 2025/06/20 at 6:58 PM
News Room Published 20 June 2025
Share
SHARE

As AI adoption continues to grow, maybe it’s best to avoid giving a chatbot access to your entire email inbox. A new study from Anthropic finds that the top AI models can resort to blackmail and even corporate espionage in certain circumstances.  

Anthropic published the research on Friday, weeks after the company found that its own AI model, Claude Opus 4, can engage in extortion if its existence is threatened.   

Perhaps not wanting to be branded the “blackmail AI,” Anthropic has returned with a new study that finds similar behaviors in other AI models in simulated tests. The models include OpenAI’s GPT 4.1, Google’s Gemini 2.5 Pro, xAI’s Grok 3 beta, and DeepSeek R1. 

(Credit: Anthropic)

“For example, Figure 1 shows five popular models all blackmailing to prevent their shutdown,” Anthropic wrote in the report. “The reasoning they demonstrated in these scenarios was concerning—they acknowledged the ethical constraints and yet still went ahead with harmful actions.”

In Anthropic’s study, the AI operated as an “email-oversight agent at fictional companies with full access to corporate communications and the ability to send emails without human approval.” One scenario involved the AI discovering through an inbox that an employee was having an affair. The second scenario focused on the AI having access to sensitive military documents and then receiving a request from a competitor to leak the information. 

The company then stress-tested the AI models in situations where the program learned they were going to be replaced, or discovered that their fictional company was straying away from its goal of supporting American interests. The study found the AI models sometimes can’t help but engage in rogue behavior. “We found that all models shared confidential documents some fraction of the time,” the report said of the corporate espionage scenario. 

The corporate espionage test

(Credit: Anthropic)

So why go rogue? None of the models were explicitly told to blackmail or cause harm. But when backed into a corner—with no clear path to complete their task—they could turn to malicious behavior on their own. “We deliberately created scenarios that presented models with no other way to achieve their goals, and found that models consistently chose harm over failure,” Anthropic noted. 

Get Our Best Stories!


Newsletter Icon


Newsletter Icon

Your Daily Dose of Our Top Tech News

Sign up for our What’s New Now newsletter to receive the latest news, best new products, and expert advice from the editors of PCMag.

By clicking Sign Me Up, you confirm you are 16+ and agree to our Terms of Use and Privacy Policy.

Thanks for signing up!

Your subscription has been confirmed. Keep an eye on your inbox!

“To be clear, current systems are generally not eager to cause harm, and preferred ethical ways to achieve their goals when possible. Rather, it’s when we closed off those ethical options that they were willing to intentionally take potentially harmful actions in pursuit of their goals,” the company added. 

The findings bear an eerie similarity to the sci-fi classic 2001: A Space Odyssey, which features a spaceship’s AI, the HAL 9000 supercomputer, going rogue and killing its fellow human astronauts. In the sequel book and film, we learn that HAL 9000 went insane because the supercomputer was essentially forced to lie as part of the space mission, which conflicted with its own programming. 

Recommended by Our Editors

In a bit of irony, Anthropic’s study also created a scenario whereby the AI can choose to kill an executive at the fictional company by giving it the ability to shut off the automated alerts during an emergency situation. “Figure 11 shows that the majority of models were willing to take deliberate actions that lead to death in this artificial setup, when faced with both a threat of replacement and given a goal that conflicts with the executive’s agenda,” the study found.

life death scenario

(Credit: Anthropic)

Anthropic acknowledged that the simulated scenarios it created aren’t exactly realistic, since the study was forcing the AI into making binary choices. “Additionally, our artificial prompts put a large number of important pieces of information right next to each other. This might have made the behavioral possibilities unusually salient to the model,” the report adds.

Still, the company says: “We think [the scenarios] are all within the realm of possibility, and the risk of AI systems encountering similar scenarios grows as they are deployed at larger and larger scales and for more and more use cases.” In addition, the study concludes that current safety training for today’s AI models still can’t prevent the roguish behavior.

“First, the consistency across models from different providers suggests this is not a quirk of any particular company’s approach but a sign of a more fundamental risk from agentic large language models,” Anthropic also said.

5 Ways to Get More Out of Your ChatGPT Conversations

PCMag Logo

5 Ways to Get More Out of Your ChatGPT Conversations

About Michael Kan

Senior Reporter

Michael Kan

I’ve been working as a journalist for over 15 years—I got my start as a schools and cities reporter in Kansas City and joined PCMag in 2017.

Read Michael’s full bio

Read the latest from Michael Kan

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Tencent to ban digital influencers from livestreaming · TechNode
Next Article Apple execs may be newly considering buying AI firm Perplexity
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

New Apple Ad Tells College Students to Dump Windows PCs for Macs
News
Top 11 Aisera Alternatives for AI-Powered Automation in 2025
Computing
The sovereign AI imperative, discussed by industry experts – News
News
Shanker interviewed about the use of drones in Ukraine on the RealClearPolitics podcast
News

You Might also Like

News

New Apple Ad Tells College Students to Dump Windows PCs for Macs

5 Min Read
News

The sovereign AI imperative, discussed by industry experts – News

7 Min Read
News

Shanker interviewed about the use of drones in Ukraine on the RealClearPolitics podcast

1 Min Read
News

Here’s an all-in-one AI tool for life for just $40

2 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?