The OpenAI superalignment team imploded due to a loss of trust in CEO Sam Altman. Many safety-focused employees left, believing Altman prioritized power and rapid AI commercialization over responsible and safe AI development.
🛡️ "Superalignment": A dedicated team to ensure AI aligns with human goals.
🏠 "Trust collapsing bit by bit": Employees lost faith in Altman's leadership gradually, like falling dominoes.
🔄 "What did Ilya see?": Memes speculated about a horrifying AI breakthrough, but it was more about pessimism in leadership.
🚧 "Sailing against the wind": A description by a team member of the struggle within OpenAI.
Key insights
Trust Issues Post-Altman's Coup Attempt
Ilya Sutskever and Jan Leike, leaders of the superalignment team, left OpenAI, citing distrust toward Sam Altman.
The attempt by Sutskever and the OpenAI board to fire Altman in November 2023 failed, leading to Altman consolidating his power.
Altman's insistence on leadership changes and actions, like fundraising with regimes like Saudi Arabia, raised red flags among safety-conscious employees.
Culture of Disagreement and Secrecy
OpenAI employees often signed non-disparagement agreements upon leaving, limiting public criticism.
Employee Daniel Kokotajlo refused such an agreement to speak out, expressing concerns over OpenAI's approach toward developing AGI responsibly.
Jan Leike voiced his concerns publicly, highlighting ongoing struggles and lack of proper resource allocation for safety efforts.
Impact on the Superalignment Team
Key departures from the team included Leike, Sutskever, Leopold Aschenbrenner, Pavel Izmailov, Cullen O'Keefe, and William Saunders.
The team's capacity diminished, raising doubts about the future focus on preventing catastrophic risks associated with AGI.
John Schulman was appointed to lead what's left of the superalignment team, but with increased challenges due to the staff and resource drain.
Key quotes
“I lost trust in OpenAI leadership and their ability to responsibly handle AGI, so I quit.” - Daniel Kokotajlo
“I have been disagreeing with OpenAI leadership about the company’s core priorities for quite some time, until we finally reached a breaking point.” - Jan Leike
“Altman’s threat to hollow out OpenAI unless the board rehired him revealed something about his character.”
“If you shoot at the king and miss, things tend to get awkward.”
“With the safety team gutted, who will make sure OpenAI’s work is safe?”
This summary contains AI-generated information and may have important inaccuracies or omissions.