Anthropic abandons its signature security promise and rewrites AI guardrails



  • Anthropic has removed its promise not to train or release AI models without security mitigations guaranteed beforehand.
  • The company will now rely on transparency reports and security roadmaps instead of strict preconditions.
  • Critics argue the change shows the limits of voluntary AI safety commitments without binding regulation.

Anthropic has formally abandoned a core promise not to train or launch border AI systems unless it can ensure adequate security in advance. The company behind Claude confirmed the decision in an interview with Timemarking the end of a policy that once distinguished him among AI developers. The newly revised Responsible Scaling Policy is more focused on ensuring the company remains competitive as the AI ​​market heats up.

For years, Anthropic presented that promise as proof that it would resist commercial pressures that pushed competitors to market ever more powerful systems. The policy effectively prevented him from advancing beyond certain levels unless predefined safeguards were already in place. Now, Anthropic is using a more flexible framework instead of categorical pauses.



Leave a Comment

Your email address will not be published. Required fields are marked *