News

The internet freaked out after Anthropic revealed that Claude attempts to report “immoral” activity to authorities under ...
So endeth the never-ending week of AI keynotes. What started with Microsoft Build, continued with Google I/O, and ended with ...
Anthropic’s AI Safety Level 3 protections add a filter and limited outbound traffic to prevent anyone from stealing the ...
Anthropic says its Claude Opus 4 model frequently tries to blackmail software engineers when they try to take it offline.
Launched this week, Claude Opus 4 has been praised for its advanced reasoning and coding abilities. But hidden in the launch report is a troubling revelation. In controlled experiments, the AI ...
In a fictional scenario set up to test Claude Opus 4, the model often resorted to blackmail when threatened with being ...
The latest versions of Anthropic's Claude generative AI models made their debut Thursday, including a heavier-duty model ...
Anthropic's new model might also report users to authorities and the press if it senses "egregious wrongdoing." ...
Safety testing AI means exposing bad behavior. But if companies hide it—or if headlines sensationalize it—public trust loses ...
Experts urge action as AI accelerates workplace automation, with warnings that entry-level roles in major industries may ...
Reed Hastings is joining Anthropic's board. Hastings also serves on the boards of Bloomberg and the City Fund.
Artificial intelligence lab Anthropic unveiled its latest top-of-the-line technology called Claude Opus 4 on Thursday, which ...