- Roadshow to Rome
- Awards
- Exhibitors
- Speakers
- Media Partners
- News
Anthropic has revised its core artificial intelligence (AI) safety framework, removing a key commitment that once defined the company’s approach to responsible development. The San Francisco based AI lab, known for its Claude models, confirmed it will no longer automatically stop training or releasing more powerful systems solely due to safety concerns. The change marks a significant shift in its Responsible Scaling Policy and reflects growing pressure in the global AI race.
Anthropic introduced its Responsible Scaling Policy in 2023 as a formal structure to manage risks linked to advanced AI systems. The framework included a “hard stop” provision. Under that rule, the company would pause development or release of models if safety standards were not met.
In its latest update, Anthropic removed that strict commitment. The company said the earlier approach helped build safeguards for current risks. However, it now faces what it calls a “zone of ambiguity”. According to Anthropic, the science of evaluating advanced AI systems has not matured enough to clearly determine the exact point at which a model becomes dangerous.
The company stated that current evaluation methods cannot provide decisive answers about high risk capabilities. As a result, it argues that rigid stop rules may not be practical in a fast moving and competitive environment.
Anthropic was founded in 2021 by former OpenAI researchers who highlighted concerns about the risks of powerful AI systems. Since launch, it has positioned itself as one of the most safety focused AI labs. Its Claude models compete directly with systems developed by OpenAI, Google and other major players.
In its updated policy, Anthropic acknowledged that acting alone could weaken its position. The company said that if it adheres to strict unilateral limits while competitors continue to develop more advanced models, it could lose influence in shaping industry standards.
Anthropic added that losing influence could result in less coordination across the sector. Instead of relying on automatic pauses, the company will now adopt a more flexible system to manage risks while continuing development.
Rather than maintaining a hard stop mechanism, Anthropic plans to increase transparency. The company will publish regular “Risk Reports” to explain how it identifies and manages safety concerns as its AI systems grow more capable.
Anthropic also stated that achieving the highest levels of protection against major threats is not possible for a single company working alone. It indicated that collaboration with governments, researchers and other AI developers will be necessary to address advanced risks.
The revised policy does not remove safety measures entirely. Instead, it replaces strict stopping commitments with ongoing risk assessments, mitigation strategies and public reporting.
The change comes at a time when competition in artificial intelligence is accelerating. AI labs are racing to build more capable large language models and integrate them into enterprise tools, consumer applications and developer platforms.
Anthropic’s earlier hard stop pledge had become a defining feature of its public identity. The update signals a recalibration of that stance in light of commercial and strategic realities.
The broader debate over AI governance continues globally. Governments in the United Kingdom, the United States and the European Union are exploring regulatory frameworks for advanced AI systems. Meanwhile, companies are balancing rapid innovation with safety expectations.
Anthropic’s revised Responsible Scaling Policy reflects this tension. The company has moved away from automatic development pauses and towards a model centred on transparency, flexible safeguards and continued participation in the competitive AI landscape.