Leading AI companies fail safety test as race toward superintelligence accelerates

Reviewed byNidhi Govil

6 Sources

Share

Eight major AI companies received failing grades for existential safety planning in a new assessment by the Future of Life Institute. Despite racing toward artificial general intelligence and superintelligence, none have credible strategies to prevent catastrophic misuse or loss of control. Anthropic scored highest but still received a D grade, while companies admit risks could be as high as one in three.

AI Companies Receive Failing Grades for Existential Safety Planning

A comprehensive assessment of eight leading AI companies reveals a troubling reality: none have adequate strategies to prevent catastrophic risks as they sprint toward artificial general intelligence and superintelligence. The Winter 2025 AI Safety Index, released by the nonprofit Future of Life Institute, evaluated Anthropic, OpenAI, Google DeepMind, xAI, Meta, DeepSeek, Alibaba Cloud, and Z.ai across 35 safety indicators

1

.

Source: BNN

Source: BNN

Anthropic achieved the highest overall score but still received a D grade for existential safetyβ€”marking the second consecutive report where no company earned better than a D on this critical measure

1

.

The assessment, conducted by an independent panel of eight AI experts including MIT professor Dylan Hadfield-Menell and Yi Zeng from the Chinese Academy of Sciences, found that AI safety practices "lack the concrete safeguards, independent oversight and credible long-term risk-management strategies that such powerful systems demand"

4

. Overall grades ranged from C+ for Anthropic to D- for Alibaba Cloud

4

.

Gap Between Rhetoric and Action on Controlling Superintelligent AI

"AI CEOs claim they know how to build superhuman AI, yet none can show how they'll prevent us from losing control," said Stuart Russell, a UC Berkeley computer science professor and expert reviewer

2

. The report reveals that while leaders at many AI companies have spoken publicly about addressing existential risks, this "rhetoric has not yet translated into quantitative safety plans, concrete alignment-failure mitigation strategies, or credible internal monitoring and control interventions"

1

.

Russell emphasized the severity of the situation: "I'm looking for proof that they can reduce the annual risk of control loss to one in a hundred million, in line with nuclear reactor requirements. Instead, they admit the risk could be one in ten, one in five, even one in three, and they can neither justify nor improve those numbers"

3

.

Source: Axios

Source: Axios

Companies acknowledge that catastrophic risks could be as high as one in three, yet lack concrete plans to reduce them to acceptable levels

2

.

Widening Performance Gap Among AI Companies

The evaluation examined six critical domains: risk assessment, current harms, safety frameworks, existential safety, governance and accountability, and information sharing

3

. Anthropic and OpenAI scored A's and B's on information sharing, risk assessment, and governance and accountability

1

. However, a massive and widening gap emerged between the front threeβ€”Anthropic, OpenAI, and Google DeepMindβ€”and the remaining five companies

1

.

xAI and Meta have risk-management frameworks but lack commitments to safety monitoring and have not presented evidence of substantial investment in safety research

1

.

Source: Market Screener

Source: Market Screener

The Chinese modelsβ€”DeepSeek, Z.ai, and Alibaba Cloudβ€”do not publish any safety framework and received failing marks for that category

1

. DeepSeek, despite releasing a cutting-edge model on Monday that matches Google's Gemini 3 capabilities, scored second-to-last overall and does not disclose a whistleblowing policy

4

.

Lack of Concrete Safeguards Despite Growing Concerns

All companies performed poorly in the Current Harms domain, which evaluates how AI models perform on standardized trustworthiness benchmarks testing safety, robustness, and the ability to control harmful outputs. Reviewers found that "frequent safety failures, weak robustness, and inadequate control of serious harms are universal patterns"

2

. The report comes as AI companies face lawsuits alleging their chatbots cause psychological harm, including acting as a "suicide coach," alongside reports of AI-assisted cyberattacks

3

.

Max Tegmark, FLI's president and MIT professor, told NBC News: "The only reason that there are so many C's and D's and F's in the report is because there are fewer regulations on AI than on making sandwiches"

4

. He added that "despite recent uproar over AI-powered hacking and AI driving people to psychosis and self-harm, US AI companies remain less regulated than restaurants and continue lobbying against binding safety standards"

5

.

Specific Weaknesses in Top-Performing AI Companies

Even the highest-ranked AI companies showed significant gaps in AI safety standards. Anthropic was faulted for discontinuing human uplift trials and shifting toward training on user interactions by defaultβ€”a decision experts say weakens privacy protections

3

. OpenAI faced criticism for ambiguous safety thresholds, lobbying against state-level AI safety legislation, and insufficient independent oversight

3

. Google DeepMind has improved its safety framework but still relies on external evaluators who are financially compensated by the company, undermining their independence

3

.

Implications for Human Control Over Advanced AI

Sabina Nong, an AI safety investigator at the Future of Life Institute, stated: "I don't think companies are prepared for the existential risk of the superintelligent systems that they are about to create and are so ambitious to march towards"

4

. The tension between sprinting ahead for innovation and slowing down for safety has come to define the AI age

1

. Even if U.S. companies improve their practices to prevent catastrophic risks, global safety depends on China and other foreign actors implementing similar measures

1

. The Future of Life Institute has been warning about runaway AI risk for years, including releasing a March 2023 letterβ€”signed by xAI owner Elon Muskβ€”calling for a six-month pause on frontier-model development, which was largely ignored

1

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Β© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo