Anthropic drops hallmark safety pledge as AI race intensifies and Pentagon pressures mount

Reviewed byNidhi Govil

9 Sources

Share

Anthropic, once the industry's safety-focused alternative, has abandoned its 2023 commitment to pause AI development for catastrophic risks. The company will now consider competitor actions before delaying releases, marking a shift from absolute safety standards to competitive strategy. The change comes as the Pentagon threatens to cut its $200 million contract over Anthropic's refusal to allow autonomous weapons use.

Anthropic Abandons Core AI Safety Commitment

Anthropic has formally revised its Responsible Scaling Policy, removing the pledge that defined its identity since 2023: the promise to pause model development if adequate safety mitigations couldn't be guaranteed in advance

1

. The company, which positioned itself as a safer alternative in the AI race, now states it will only delay development "until and unless we no longer believe we have a significant lead" over competitors

2

. This represents one of the most dramatic policy shifts in the industry, as startups once focused on helping humanity pivot toward profitability and competitive survival.

Source: PYMNTS

Source: PYMNTS

Founded in 2021 by former OpenAI employees seeking to prioritize AI safety, Anthropic hoped to inspire an industry-wide "race to the top" rather than a race to the bottom

3

. Chief Science Officer Jared Kaplan told TIME that "we didn't really feel, with the rapid advance of AI, that it made sense for us to make unilateral commitments... if competitors are blazing ahead"

5

. The company's original vision hasn't "pan out" as hoped, with the policy environment now prioritizing AI competitiveness and economic growth over safety-oriented discussions at the federal level

3

.

Competitive Pressures Drive Relaxing Safety Protocols

The revision reflects mounting competitive pressures among AI labs and between nations, particularly the U.S. and China. Google DeepMind CEO Demis Hassabis has warned that "race conditions" can drive reckless decisions as the world approaches superhuman AI

2

. Max Tegmark, founder of the Future of Life Institute, argues that AI companies have stalled meaningful regulation for years. "It's their fault that we have the race condition in the first place," Tegmark told Axios, noting that Anthropic is "the most striking one because they were the ones who always talk such a big game about safety"

2

.

Source: Axios

Source: Axios

The new framework emphasizes transparency through "Frontier Safety Roadmaps" and regular "Risk Reports" rather than categorical pauses

4

. Anthropic maintains it will match or exceed competitors' safety efforts, but the core promise that distinguished Claude from rivals is gone. "The new policy still includes some guardrails, but the core promise, that Anthropic would not release models unless it could guarantee adequate safety mitigations in advance, is gone," said Nik Kairinos, CEO of RAIDS AI

4

.

Pentagon Tensions and Military Use Controversies

The timing coincides with escalating tensions between Anthropic and the Pentagon over a $200 million military contract signed last summer

5

. Defense Secretary Pete Hegseth reportedly gave Dario Amodei an ultimatum: lighten AI guardrails to make them more amenable to military use, or face being cut off and labeled a "supply chain risk"

5

. Anthropic has refused to allow its models for autonomous weapons or domestic surveillance, prompting the Defense Department to cut use of Claude

2

.

Source: Futurism

Source: Futurism

This highlights a critical problem with voluntary AI safety commitments: even if one company refuses on safety grounds, another steps in. Hours after Anthropic's stance became public, OpenAI announced a deal to provide models for classified networks

2

. Critics note OpenAI's agreement leaves broad room for military use, including surveillance of U.S. citizens, despite sharing many of Anthropic's stated safety concerns.

Implications for Self-Regulation and AI Regulation

The shift exposes fundamental limits of self-regulation without binding AI regulation. When Anthropic introduced its original policy in 2023, executives hoped it might inspire rivals or inform eventual legislation

4

. That regulatory momentum never materialized, and the current political climate has tilted away from developing frameworks. Companies now choose between voluntary restraint and competitive survival in an environment where capabilities advance while global AI summits increasingly focus on commercialization over guardrails

2

.

Kairinos noted the irony that Anthropic recently donated $20 million to Public First Action, supporting congressional candidates pledging to push for AI safety regulation, while simultaneously recalibrating its own internal constraints

4

. Tegmark suggests mounting evidence of chatbot risks affecting children and teens has sparked rare bipartisan concern, potentially opening pathways for laws mandating companies test frontier AI models before release

2

.

What This Means for Users and Industry

Interestingly, Anthropic surged to the top of Apple's App Store download charts in days following its standoff with the Pentagon, suggesting consumers may value safety commitments over competitive positioning

2

. Yet everyday users might not notice immediate changes as they interact with Claude or other AI tools. The AI guardrails governing how systems are trained influence everything from accuracy to fraudulent misuse

4

.

Anthropic argues that meaningful safety research requires staying at the frontier rather than stepping back, and that unilateral restraint could result in a less safe world where "developers with the weakest protections would set the pace"

5

. Kaplan insists "I don't think we're making any kind of U-turn," though critics view the policy change as flagrantly contradicting the organization's entire reason for existence

5

. As Hassabis has argued, the closer the world gets to superhuman AI, the more essential global cooperation becomes—yet the trajectory appears moving in the opposite direction

2

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo