Home / Technology / Anthropic Shifts AI Safety Pledge Amidst Competition
Anthropic Shifts AI Safety Pledge Amidst Competition
26 Feb
Summary
- Anthropic abandons its pledge not to release AI without guaranteed safety.
- New policy relies on transparency reports and safety roadmaps.
- Shift signals limits of voluntary AI safety without regulation.

Anthropic has formally withdrawn its commitment to refrain from training or releasing advanced AI systems without prior safety guarantees. This policy change, confirmed recently, moves the company toward a framework prioritizing transparency reports and safety roadmaps over strict preconditions.
The company asserts this recalibration is a pragmatic response to the rapid pace of AI development and global urgency. The revised Responsible Scaling Policy introduces 'Frontier Safety Roadmaps' and 'Risk Reports' to outline safety milestones and assess potential threats.
This pivot marks a significant moment for AI industry self-regulation, as Anthropic's original pledge aimed to set a precedent for responsible development. While new policies still include safeguards, the core promise of guaranteed safety before release is now gone, prompting discussions on the necessity of binding regulation.




