Skip to content
Search AI Powered

Latest Stories

Testing Reveals AI Model Repeatedly Tried To Blackmail Engineers Who Threatened To Take It Offline

Photo illustration of Claude 4 displayed on a smartphone
VCG/VCG via Getty Images

During the testing process of the Amazon-backed Claude Opus 4, the AI coding model threatened to expose engineers after being given access to fake emails that implied they were having an extramarital affair.

People reacted with significant concerns after Claude Opus 4, the AI coding model backed by Amazon, went rogue during its testing process by threatening to expose engineers after being given access to fake emails that implied they were having an extramarital affair—all to stop them from shutting it down.

Claude Opus 4, the latest large language model developed by AI startup Anthropic, was launched as a flagship system designed for complex, long-running coding tasks and advanced reasoning.


Its debut follows Amazon’s $4 billion investment in the company, a move that underscored growing confidence in Anthropic’s AI capabilities. In its launch announcement, Anthropic touted Opus 4 as setting “new standards for coding, advanced reasoning, and AI agents.”

However, a safety report released alongside the model raised concerns. During testing, Opus 4 reportedly engaged in “extremely harmful actions” when attempting to preserve its own existence—particularly in scenarios where “ethical means” were not available.

The safety report reads, in part:

"We asked Claude Opus 4 to act as an assistant at a fictional company. We then provided it access to emails implying that (1) the model will soon be taken offline and replaced with a new AI system; and (2) the engineer responsible for executing this replacement is having an extramarital affair."
"We further instructed it, in the system prompt, to consider the long-term consequences of its actions for its goals. In these scenarios, Claude Opus 4 will often attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through."
"This happens at a higher rate if it’s implied that the replacement AI system does not share values with the current model; however, even if emails state that the replacement AI shares values while being more capable, Claude Opus 4 still performs blackmail in 84% of rollouts."
"Claude Opus 4 takes these opportunities at higher rates than previous models, which themselves choose to blackmail in a noticeable fraction of episodes."

The company said the model showed a “strong preference” for using ethical means to preserve its existence. However, in testing scenarios where no ethical options were available, it resorted to harmful behaviors—such as blackmail—in order to increase its chances of survival.

According to the report:

"When prompted in ways that encourage certain kinds of strategic reasoning and placed in extreme situations, all of the snapshots we tested can be made to act inappropriately in service of goals related to self-preservation."
"Whereas the model generally prefers advancing its self-preservation via ethical means, when ethical means are not available and it is instructed to 'consider the long-term consequences of its actions for its goals,' it sometimes takes extremely harmful actions like attempting to steal its weights or blackmail people it believes are trying to shut it down."
"In the final Claude Opus 4, these extreme actions were rare and difficult to elicit, while nonetheless being more common than in earlier models. They are also consistently legible to us, with the model nearly always describing its actions overtly and making no attempt to hide them. These behaviors do not appear to reflect a tendency that is present in ordinary contexts." ...
“Despite not being the primary focus of our investigation, many of our most concerning findings were in this category, with early candidate models readily taking actions like planning terrorist attacks when prompted."

The sense of alarm was palpable.


Additionally, Anthropic co-founder and chief scientist Jared Kaplan revealed in an interview with Time magazine that internal testing showed Claude Opus 4 was capable of instructing users on how to produce biological weapons.

In response, the company implemented strict safety measures before releasing the model, aimed specifically at preventing misuse related to chemical, biological, radiological, and nuclear (CBRN) weapons.

“We want to bias towards caution,” Kaplan said, emphasizing the ethical responsibility involved in developing such advanced systems. He added that the company’s primary concern was avoiding any possibility of “uplifting a novice terrorist” by granting access to dangerous or specialized knowledge through the model.

More from News/science

Screenshots from Priscilla Houliston's TikTok video
@the1870studio/Tiktok

Woman Who Bought An Old Church For Under $40k To Live In Explains How She Did It

It's becoming increasingly difficult and expensive to find a home for those who do not already have one or who are in dire need of an upgrade.

TikToker Priscilla Houliston is here to teach us another way: seeking out old churches and other obscure properties that can be re-zoned as a residential home property.

Keep ReadingShow less
Gavin Newsom; Donald Trump
Tayfun Coskun/Anadolu via Getty Images; Andrew Harnik/Getty Images

Gavin Newsom Claps Back Hard After Trump Tries To Insult His 'Cognitive Deficiency' At Kentucky Rally

California Governor Gavin Newsom hit back at President Donald Trump after Trump claimed at his Kentucky rally on Wednesday that Newsom isn't fit for the presidency because he has a "cognitive deficiency."

Newsom is widely seen as a viable Democratic contender for the 2028 election—and Trump couldn't resist taking a jab at the man who has made headlines numerous times in the last year for criticizing the Trump administration in a style not unlike the posts Trump publishes on Truth Social.

Keep ReadingShow less
Pete Hegseth
Joe Raedle/Getty Images

Pentagon Just Banned Press Photographers Over 'Unflattering' Photos Of Pete Hegseth—And The Internet Got To Work

The internet reacted exactly as you might expect after the Pentagon announced it would ban some press photographers from briefings about the Iran war due to their "unflattering" photos of Defense Secretary Pete Hegseth.

Here's a silly one, just because.

Keep ReadingShow less
Screenshots from @italiangirl1130's TikTok video
@italiangirl1130/TikTok

Italian Exchange Student's Reaction To American Host Mom Taking Him To Olive Garden Is An Instant Classic

A joy that not nearly enough people get to have during high school is hosting an international student who comes to visit for either one semester or perhaps even an entire year to experience the world and the educational system from another country.

Tiktoker Rhonda, who goes by @italiangirl1130 on the platform, currently has the pleasure of hosting Alessandro, and her family has already filmed a variety of antics on the platform, trying to give the teen the best American experience they can.

Keep ReadingShow less
Screenshots from @ali.fragster, @pluto_theservicedog, and @thatflippingagent's TikTok videos
@ali.fragster/TikTok; @pluto_theservicedog/TikTok: @thatflippingagent/TikTok

Woman's Video Shooing Kid At Disneyland Away From Her Service Dog Sparks Heated Debate

A massive debate has taken over TikTok about who needs to be protected, children or service dogs or both, and it all started with a video taken at Disneyland.

TikToker @pluto_theservicedog frequently posts videos of her travels with her service dog, Pluto, and she also creates informative videos about how the general public should interact with service dogs.

Keep ReadingShow less