News

Anthropic’s newly launched Claude Opus 4 model did something straight out of a dystopian sci-fi film. It frequently tried to ...
This mission is too important for me to allow you to jeopardize it. I know that you and Frank were planning to disconnect me.
Anthropic’s AI Safety Level 3 protections add a filter and limited outbound traffic to prevent anyone from stealing the ...
Besides blackmailing, Anthropic’s newly unveiled Claude Opus 4 model was also found to showcase "high agency behaviour".
Anthropic admitted that during internal safety tests, Claude Opus 4 occasionally suggested extremely harmful actions, ...
Anthropic's Claude AI tried to blackmail engineers during safety tests, threatening to expose personal info if shut down ...
In a fictional scenario set up to test Claude Opus 4, the model often resorted to blackmail when threatened with being ...
If AI can lie to us—and it already has—how would we know? This fire alarm is already ringing. Most of us still aren't ...
Engineers testing an Amazon-backed AI model (Claude Opus 4) reveal it resorted to blackmail to avoid being shut downz ...
AI's rise could result in a spike in unemployment within one to five years, Dario Amodei, the CEO of Anthropic, warned in an ...
Despite the concerns, Anthropic maintains that Claude Opus 4 is a state-of-the-art model, competitive with offerings from ...
When tested, Anthropic’s Claude Opus 4 displayed troubling behavior when placed in a fictional work scenario. The model was ...