Anthropic, the AI company that has long positioned itself as the industry’s most safety-conscious research lab, has dropped the central commitment of its Responsible Scaling Policy (RSP) — the pledge to halt model training if adequate safety measures could not be guaranteed in advance. The announcement, published on February 25, 2026, marks one of the most dramatic policy reversals in the AI industry to date.
Anthropic introduced its original RSP in 2023, promising to never train an AI system unless the company could guarantee beforehand that its safety measures were adequate. If a model’s capabilities outstripped the company’s ability to control them, training would pause — a hard limit that set Anthropic apart from competitors like OpenAI and Google DeepMind.
The revised RSP v3.0 eliminates this hard limit entirely. Under the new framework, Anthropic commits to delaying development only if two conditions are simultaneously met: the company believes it holds a significant lead over competitors, and it identifies catastrophic risks. If competitors are “blazing ahead,” as Chief Science Officer Jared Kaplan put it, Anthropic will no longer pause unilaterally.
“We didn’t really feel, with the rapid advance of AI, that it made sense for us to make unilateral commitments … if competitors are blazing ahead,” Kaplan told TIME. “We felt that it wouldn’t actually help anyone for us to stop training AI models.”
Several converging pressures appear to have driven the change:
RSP v3.0 introduces three structural changes in place of the original hard limits:
The reaction from the AI safety community has been swift. Chris Painter, director of policy at METR, an AI evaluation nonprofit, warned: “This is more evidence that society is not prepared for the potential catastrophic risks posed by AI.” He raised concerns about a “frog-boiling” dynamic, where dangers escalate so gradually that no single change triggers alarm.
Critics note the inherent tension in Anthropic’s argument: the company contends that pausing while less careful actors continue would make the world less safe, yet the same logic could justify any safety-conscious lab abandoning its commitments. The policy shift is particularly striking for a company that has described itself as the AI firm with a “soul.”
