OpenAI board forms Safety and Security Committee

May 29, 2024

  • OpenAI’s board formed a Safety and Security Committee to develop processes and safeguards
  • The committee will advise the board on critical safety and security decisions
  • OpenAI says it has started training its next frontier model with “next level of capabilities”

OpenAI’s board announced the formation of a Safety and Security Committee which is tasked with making recommendations on critical safety and security decisions for all OpenAI projects.

The committee is led by directors Bret Taylor (Chair), Adam D’Angelo, Nicole Seligman, and OpenAI’s CEO Sam Altman.

Aleksander Madry (Head of Preparedness), Lilian Weng (Head of Safety Systems), John Schulman (Head of Alignment Science), Matt Knight (Head of Security), and Jakub Pachocki (Chief Scientist) will also be on the committee.

OpenAI’s approach to AI safety has faced both external and internal criticism. Last year’s firing of Altman was supported by then-board member Ilya Sutskever and others, ostensibly over safety concerns.

Last week Sutskever and Jan Leike from OpenAI‘s “superalignment” team left the company. Leike specifically noted safety issues as his reason for leaving, saying the company was letting safety “take a backseat to shiny products”.

Yesterday, Leike announced that he was joining Anthropic to work on oversight and alignment research.

Now Altman is not only back as CEO, but also sits on the committee responsible for highlighting safety issues. Former board member Helen Toner’s insights into why Altman was fired makes you wonder how transparent he’ll be about safety issues the committee discovers.

Apparently the OpenAI board found out about the release of ChatGPT via Twitter.

The Safety and Security Committee will use the next 90 days to evaluate and further develop OpenAI’s processes and safeguards.

The recommendations will be put to OpenAI’s board for approval and the company has committed to publishing the adopted safety recommendations.

This push for additional guardrails comes as OpenAI says it has started training its next frontier model which it says will “bring us to the next level of capabilities on our path to AGI.”

No expected release date was offered for the new model but training alone will probably take weeks if not months.

In an update on its approach to safety published after the AI Seoul Summit, OpenAI said “We won’t release a new model if it crosses a “Medium” risk threshold from our Preparedness Framework, until we implement sufficient safety interventions to bring the post-mitigation score back to “Medium”.”

It said that more than 70 external experts were involved in red teaming GPT-4o before its release.

With 90 days to go before the committee presents its findings to the board, only recently started training, and a commitment to extensive red teaming, it looks like we’ve got a long wait before we finally get GPT-5.

Or do they mean they’ve just started training GPT-6?

 

Join The Future


SUBSCRIBE TODAY

Clear, concise, comprehensive. Get a grip on AI developments with DailyAI

Eugene van der Watt

Eugene comes from an electronic engineering background and loves all things tech. When he takes a break from consuming AI news you'll find him at the snooker table.

×

FREE PDF EXCLUSIVE
Stay Ahead with DailyAI

Sign up for our weekly newsletter and receive exclusive access to DailyAI's Latest eBook: 'Mastering AI Tools: Your 2024 Guide to Enhanced Productivity'.

*By subscribing to our newsletter you accept our Privacy Policy and our Terms and Conditions