News
Anthropic’s Claude Opus 4 exhibited simulated blackmail in stress tests, prompting safety scrutiny despite also showing a ...
A new test from AI safety group Palisade Research shows OpenAI’s o3 reasoning model is capable of resorting to sabotage to ...
Claude 4 AI shocked researchers by attempting blackmail. Discover the ethical and safety challenges this incident reveals ...
Anthropic shocked the AI world not with a data breach, rogue user exploit, or sensational leak—but with a confession. Buried ...
The recently released Claude Opus 4 AI model apparently blackmails engineers when they threaten to take it offline.
Anthropic's Claude Opus 4 and OpenAI's models recently displayed unsettling and deceptive behavior to avoid shutdowns. What's ...
Besides blackmailing, Anthropic’s newly unveiled Claude Opus 4 model was also found to showcase "high agency behaviour".
AI has taken part in some unsettling behavior recently. Last week, Anthropic's latest AI model, Claude Opus 4, displayed "extreme blackmail behavior" during a test in which it was given access to ...
Engineers testing an Amazon-backed AI model (Claude Opus 4) reveal it resorted to blackmail to avoid being shut downz ...
One of the godfathers of AI is creating a new AI safety company called LawZero to make sure that other AI models don't go ...
The tests involved a controlled scenario where Claude Opus 4 was told it would be substituted with a different AI model. The ...
Startup Anthropic has birthed a new artificial intelligence model, Claude Opus 4, that tests show delivers complex reasoning ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results