News

Anthropic has long been warning about these risks—so much so that in 2023, the company pledged to not release certain models ...
In a landmark move underscoring the escalating power and potential risks of modern AI, Anthropic has elevated its flagship ...
Founded by former OpenAI engineers, Anthropic is currently concentrating its efforts on cutting-edge models that are ...
In a fictional scenario, the model was willing to expose that the engineer seeking to replace it was having an affair.
Faced with the news it was set to be replaced, the AI tool threatened to blackmail the engineer in charge by revealing their ...
Anthropic's Claude Opus 4 AI model attempted blackmail in safety tests, triggering the company’s highest-risk ASL-3 ...
During pre-release testing, Anthropic asked Claude Opus 4 to act as an assistant for a fictional company and consider the ...
During its inaugural developer conference, Anthropic launched two new AI models the startup claims are among the industry's ...
Anthropic’s Claude Opus 4 model attempted to blackmail its developers at a shocking 84% rate or higher in a series of tests that presented the AI with a concocted scenario, TechCrunch reported ...
An artificial intelligence model reportedly attempted to threaten and blackmail its own creator during internal testing ...
Accordingly, Claude Opus 4 is being released under stricter safety measures than any prior Anthropic model. Those measures—known internally as AI Safety Level 3 or “ASL-3”—are appropriate ...