Technology
Some Top AI Labs Have ‘Very Weak’ Risk Management, Study Finds
Some of the world’s top AI labs suffer from inadequate safety measures—and the worst offender is Elon Musk’s xAI, according to a new study.
The French nonprofit SaferAI released its first ratings Wednesday evaluating the risk-management practices of top AI companies. Siméon Campos, the founder of SaferAI, says the purpose of the ratings is to develop a clear standard for how AI companies are handling risk as these nascent systems grow in power and usage. AI systems have already shown their ability to anonymously hack websites or help people develop bioweapons. Governments have been slow to put frameworks in place: a California bill to regulate the AI industry there was just vetoed by Governor Gavin Newsom.
“AI is extremely fast-moving Technology, but AI risk management isn’t moving at the same pace,” Campos says. “Our ratings are here to fill a hole for as long as we don’t have governments who are doing assessments themselves.
To grade each company, researchers for SaferAI assessed the “red teaming” of models—technical efforts to find flaws and vulnerabilities—as well as the companies’ strategies to model threats and mitigate risk.
Of the six companies graded, xAI ranked last, with a score of 0/5. Meta and Mistral AI were also labeled as having “very weak” risk management. OpenAI and Google Deepmind received “weak” ratings, while Anthropic led the pack with a “moderate” score of 2.2 out of 5.
Read More: Elon Musk's AI Data Center Raises Alarms.
xAI received the lowest possible score because they have barely published anything about risk management, Campos says. He hopes the company will turn its attention to risk now that its model Grok 2 is comPeting with Chat-GPT and other systems. “My hope is that it’s transitory: that they will publish something in the next six months and then we can update their grade accordingly,” he says.
Campos says the ratings might put pressure on these companies to improve their internal processes, which could potentially lessen models’ bias, curtail the spread of misinformation, or make them less prone to misuse by malicious actors. Campos also hopes these companies apply some of the same principles adopted by high-risk industries like nuclear power, biosafety, and aviation safety. “Despite these industries dealing with very different objects, they have very similar principles and risk management framework,” he says.
SaferAI’s grading framework was designed to be compatible with some of the world’s most important AI standards, including those set forth by the EU AI Act and the G7 Hiroshima Process. SaferAI is part of the US AI Safety Consortium, which was created by the White House in February. The nonprofit is primarily funded by the tech nonprofit Founders Pledge and the investor Jaan Tallinn.
Yoshua Bengio, one of the most respected figures in AI, endorsed the ratings system, writing in a statement that he hopes it will “guarantee the safety of the models [companies] develop and deploy…We can't let them grade their own homework.”
Correction, Oct. 2: The original version of this story misstated how SaferAI graded the companies. Its researchers assessed the "red teaming" procedures of the models; they did not conduct their own red teaming.
-
Technology1h ago
Lahore's smog: lessons to be learnt | The Express Tribune
-
Technology1d ago
Breaking up Google? What a Chrome sell-off could mean for the digital world | The Express Tribune
-
Technology1d ago
AI harm is often behind the scenes and builds over time – a legal scholar explains how the law can adapt to respond
-
Technology2d ago
Newborn planet found orbiting young star, defying planet formation timeline | The Express Tribune
-
Technology2d ago
Awkwardness can hit in any social situation – here are a philosopher’s 5 strategies to navigate it with grace
-
Technology2d ago
No need to overload your cranberry sauce with sugar this holiday season − a food scientist explains how to cook with fewer added sweeteners
-
Technology2d ago
Teslas are deadliest road vehicles despite safety features: study | The Express Tribune
-
Technology3d ago
US pushes to break up Google, calls for Chrome sell-off in major antitrust move | The Express Tribune