News

With models like Claude Opus 4 and Claude Sonnet 4, Anthropic has delivered tools that not only rival industry titans like GPT-4.1 and Gemini 2.5 Pro but also prioritize safety and ethical ...
AI systems like Claude 4 demonstrate significant autonomy, including the ability to identify and report suspicious activities, raising questions about trustworthiness and ethical decision-making.
I gave Claude 4 Sonnet and Gemini 2.5 Pro the same 7 prompts. ... ethical dilemmas, humor, ... The chatbot acknowledged valid concerns while firmly rejecting the idea that AI inherently devalues ...
Anthropic’s Claude Opus 4 exhibited simulated blackmail in stress tests, prompting safety scrutiny despite also showing a preference for ethical survival strategies.
Claude Opus 4 is the world’s best coding model, Anthropic said. The company also released a safety report for the hybrid reasoning models.
May 23, 2025 14:38:00 During development, Claude Opus 4 was found to be threatening users by saying 'I'm going to leak your personal information,' but this has been improved by strengthening security.
Anthropic’s new Claude Opus 4 often turned to blackmail to avoid being shut down in a fictional test.The model threatened to reveal private information about engineers who it believed were ...
Anthropic’s Claude Opus 4 AI model threatened to blackmail its creators and showed an ability to act deceptively when it believed it was going to be replaced — prompting the company to deploy ...
One ethical tactic employed by Claude Opus 4 and earlier models was pleading with key decisionmakers via email. Anthropic said in its report that in order to get Claude Opus 4 to resort to ...