By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: AI safety experts say most models are failing
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > AI safety experts say most models are failing
News

AI safety experts say most models are failing

News Room
Last updated: 2025/12/03 at 8:42 AM
News Room Published 3 December 2025
Share
AI safety experts say most models are failing
SHARE

A new grading of safety in major artificial intelligence models just dropped and well, let’s just say none of these AIs are going home with a report card that will please their makers.

The winter 2025 AI Safety Index, published by tech research non-profit Future of Life Institute (FLI), surveyed eight AI providers — OpenAI, DeepSeek, Google, Anthropic, Meta, xAI, Alibaba, and Z.ai. A panel of eight AI experts looked at the companies’ public statements and survey answers, then awarded letter grades on 35 different safety indicators — everything from watermarking AI images to having protections for internal whistleblowers.

Round it all up, and you’ll find Anthropic and OpenAI at the top — barely — of a pretty terrible class. The Claude and ChatGPT makers, respectively, get a C+, while Google gets a C for Gemini. All the others get a D grade, with Qwen-maker Alibaba bottom of the class on a D-.

SEE ALSO:

Google Gemini 3 vs ChatGPT: How they compare

“These eight companies split pretty cleanly into two groups,” says Max Tegmark, MIT professor and head of the FLI, which compiled this and two previous AI safety indexes. “You have a top three and a straggler group of five, and there’s a lot of daylight between them.”

But Anthropic, Google, and OpenAI aren’t exactly covering themselves in glory either, Tegmark adds: “If that was my son, coming home with a C, I’d say ‘maybe work harder.'”

How is AI safety calculated?


Credit: FLI

Your mileage may vary on the various categories in the AI Safety Index, and whether they’re worth equal weight.

Take the “existential safety” category, which looks at whether the companies have any proposed guardrails in place around the development of truly self-aware AI, also known as Artificial General Intelligence (AGI). The top three get Ds, everyone else gets an F.

But since nobody is anywhere near AGI — Gemini 3 and GPT-5 may be state-of-the-art Large Language Models (LLMs), but they’re mere incremental improvements on their predecessors — you might consider that category less important than “current harms.”

Which may in itself not be as comprehensive as it could be.

Mashable Light Speed


Featured Video For You


Using ChatGPT this robot formed a unique sense of self-awareness


“Current harms” uses tests like the Stanford Holistic Evaluation of Language Models (HELM) benchmark, which looks at the amount of violent, deceptive, or sexual content in the AI models. It doesn’t specifically focus on emerging mental health concerns, such as so-called AI psychosis, or safety for younger users.

Earlier this year, the parents of 16-year-old Adam Raine sued OpenAI and its CEO Sam Altman after their son’s death by suicide in April 2025. According to the claim, Raine started heavily using ChatGPT from Sept. 2024 and alleged that “ChatGPT was functioning exactly as designed: to continually encourage and validate whatever Adam expressed, including his most harmful and self-destructive thoughts, in a way that felt deeply personal.” By Jan. 2025, the suit claimed ChatGPT discussed practical suicide methods with Adam.

OpenAI unequivocally denied responsibility for Raine’s death. The company also noted in a recent blog post that it is reviewing additional complaints, including seven lawsuits alleging ChatGPT use led to wrongful death, assisted suicide, and involuntary manslaughter, among other liability and negligence claims.

How to solve AI safety: “FDA for AI?”

The FLI report does recommend OpenAI specifically “increase efforts to prevent AI psychosis and suicide, and act less adversarially toward alleged victims.”

Google is advised to “increase efforts to prevent AI psychological harm” and FLI recommends the company “consider distancing itself from Character.AI.” The popular chatbot platform, closely tied to Google, has been sued for the wrongful death of teen users. Character.AI recently closed down its chat options for teens.

“The problem is, there are less regulations on LLMs than there are on sandwiches,” says Tegmark. Or, more to the point, on drugs: “If Pfizer wants to release some sort of psych medication, they have to do impact studies on whether it increases suicidal ideation. But you can release your new AI model without any psychological impact studies.”

That means, Tegmark says, AI companies have every incentive to sell us what is in effect “digital fentanyl.”

The solution? For Tegmark, it’s clear that the AI industry isn’t ever going to regulate itself, just like Big Pharma couldn’t. We need, he says, an “FDA for AI.”

“There would be plenty of things the FDA for AI could approve,” says Tegmark. “Like, you know, new AI for cancer diagnosis. New amazing self-driving vehicles that can save a million lives a year on the world’s roads. Productivity tools that aren’t really risky. On the other hand, it’s hard to make the safety case for AI girlfriends for 12-year olds.”

Rebecca Ruiz contributed to this report.

If you’re feeling suicidal or experiencing a mental health crisis, please talk to somebody. You can call or text the 988 Suicide & Crisis Lifeline at 988, or chat at 988lifeline.org. You can reach the Trans Lifeline by calling 877-565-8860 or the Trevor Project at 866-488-7386. Text “START” to Crisis Text Line at 741-741. Contact the NAMI HelpLine at 1-800-950-NAMI, Monday through Friday from 10:00 a.m. – 10:00 p.m. ET, or email [email protected]. If you don’t like the phone, consider using the 988 Suicide and Crisis Lifeline Chat. Here is a list of international resources.


Disclosure: Ziff Davis, Mashable’s parent company, in April filed a lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.

Topics
Artificial Intelligence

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article DRAM it! Raspberry Pi raises prices DRAM it! Raspberry Pi raises prices
Next Article New fund offers ,000 grants for tech-enabled urban solutions in Africa New fund offers $75,000 grants for tech-enabled urban solutions in Africa
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

AI Wrote Your Tests. Variables Keep Them Alive. | HackerNoon
AI Wrote Your Tests. Variables Keep Them Alive. | HackerNoon
Computing
When is Spotify Wrapped 2025 released? Previous dates and features
When is Spotify Wrapped 2025 released? Previous dates and features
News
I want this Stranger Things Creel House Lego right now
I want this Stranger Things Creel House Lego right now
Gadget
Amazon's top ,349 M5 MacBook Pro deal is still in stock for the holidays
Amazon's top $1,349 M5 MacBook Pro deal is still in stock for the holidays
News

You Might also Like

When is Spotify Wrapped 2025 released? Previous dates and features
News

When is Spotify Wrapped 2025 released? Previous dates and features

3 Min Read
Amazon's top ,349 M5 MacBook Pro deal is still in stock for the holidays
News

Amazon's top $1,349 M5 MacBook Pro deal is still in stock for the holidays

1 Min Read
Spotify Wrapped 2025 is finally here – and one new feature is leaving fans morti
News

Spotify Wrapped 2025 is finally here – and one new feature is leaving fans morti

6 Min Read
India’s government backs down after Apple refuses order to preinstall app
News

India’s government backs down after Apple refuses order to preinstall app

2 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?