Particle.news

Download on the App Store

OpenAI Disbands AI Safety Team Following Leadership Departures

The company integrates safety efforts into broader research amid concerns over prioritizing speed in AI development.

  • OpenAI's superalignment team, focused on long-term AI risks, has been dissolved after less than a year.
  • Co-founders Ilya Sutskever and Jan Leike, who led the team, have both resigned citing disagreements with company leadership.
  • The team faced internal challenges, including resource constraints and conflicts over the company's core priorities.
  • Safety research will now be distributed across various teams under new leadership.
  • The move raises questions about OpenAI's commitment to balancing AI innovation with safety.
Hero image