Particle.news

Download on the App Store

OpenAI Disbands AI Safety Team Following Leadership Departures

The company integrates safety efforts into broader research amid concerns over prioritizing speed in AI development.

Overview

  • OpenAI's superalignment team, focused on long-term AI risks, has been dissolved after less than a year.
  • Co-founders Ilya Sutskever and Jan Leike, who led the team, have both resigned citing disagreements with company leadership.
  • The team faced internal challenges, including resource constraints and conflicts over the company's core priorities.
  • Safety research will now be distributed across various teams under new leadership.
  • The move raises questions about OpenAI's commitment to balancing AI innovation with safety.