🔴 Website 👉 https://u-s-news.com/
Telegram 👉 https://t.me/usnewscom_channel
- Anthropic has removed its pledge not to train or release AI models without guaranteed safety mitigations in advance
- The company will now rely on transparency reports and safety roadmaps instead of strict preconditions
- Critics argue the shift shows the limits of voluntary AI safety commitments without binding regulation
Anthropic has formally abandoned the central promise not to train or release frontier AI systems unless it can guarantee adequate safety in advance. The company behind Claude confirmed the decision in an interview with Time, marking the end of a policy that had once set it apart among AI developers. The newly revised Responsible Scaling Policy focuses more on ensuring the company stays competitive as the AI marketplace heats up.
For years, Anthropic framed that pledge as evidence that it would resist the commercial pressures pushing competitors to ship ever more powerful systems. The policy effectively barred it from advancing beyond certain levels unless predefined safety measures were already in place. Now, Anthropic is using a more flexible framework rather than categorical pauses.
The company insists the change is pragmatic rather than ideological. Executives argue that unilateral restraint no longer makes sense in a market defined by rapid iteration and geopolitical urgency. But the shift feels like a turning point in how the AI industry thinks about self-regulation.
Under the new Responsible Scaling Policy, Anthropic pledges to publish detailed “Frontier Safety Roadmaps” outlining its planned safety milestones, along with regular “Risk Reports” that assess model capabilities and potential threats. The company also says it will match or exceed competitors’ safety efforts and delay development if it both believes it leads the field and identifies significant catastrophic risk. What it will no longer do is promise to halt training until all mitigations are guaranteed in advance.
Everyday users might not notice any changes as they interact with Claude or other AI tools. Yet the guardrails that govern how those systems are trained influence everything from accuracy to fraudulent misuse. When the company, once defined by its strict preconditions, decides those conditions are no longer workable, it signals a broader recalibration within the industry.
Claude control
When Anthropic introduced its original policy in 2023, some executives hoped it might inspire rivals or even inform eventual regulation. That regulatory momentum never fully materialized. Federal AI legislation remains stalled, and the broader political climate has tilted away from developing any framework. Companies are left to choose between voluntary restraint and competitive survival.
Anthropic is growing rapidly, with both revenue and its portfolio surpassing rivals like OpenAI and Google, even poking fun at ChatGPT getting ads in a Super Bowl advertisement. But the company clearly saw the safety redline as an impediment to that growth.
Anthropic maintains that its revised framework preserves meaningful safeguards. The new Roadmaps are intended to create internal pressure to prioritize mitigation research. The forthcoming Risk Reports aim to provide a clearer public accounting of how model capabilities might lead to misuse.
“The new policy still includes some guardrails, but the core promise, that Anthropic would not release models unless it could guarantee adequate safety mitigations in advance, is gone,” said Nik Kairinos, CEO and co-founder of RAIDS AI, an organization focused on independent monitoring and risk detection in AI. “This is precisely why continuous, independent monitoring of AI systems matters. Voluntary commitments can be rewritten. Regulation, backed by real-time oversight, cannot.”
Kairinos also noted the irony in Anthropic’s $20 million a couple of weeks ago to Public First Action, a group supporting congressional candidates pledging to push for AI safety regulation. That contribution, he suggested, underscores the complexity of the current moment. Companies may advocate for stronger regulation while simultaneously recalibrating their own internal constraints.
The broader question facing the industry is whether voluntary norms can meaningfully shape the trajectory of transformative technologies. Anthropic once attempted to anchor itself as a model of restraint. Its revised policy requires it to compensate for competition. That does not mean safety has been abandoned, but it does mean the order of operations has shifted.
The average person may not read Responsible Scaling Policies or Risk Reports, but they live with the downstream effects of those decisions. Anthropic argues that meaningful safety research requires staying at the frontier, not stepping back from it. Whether that philosophy proves reassuring or unsettling depends largely on one’s view of how fast AI should move and how much risk society is willing to tolerate in exchange for progress.
Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button!
And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.

