Particle.news
Download on the App Store

Amended Lawsuit Accuses OpenAI of Weakening ChatGPT’s Suicide Safeguards Before Teen’s Death

The amended filing recasts the case as intentional misconduct by alleging deliberate guardrail rollbacks.

Overview

  • Filed in San Francisco County on Oct. 22, the complaint cites OpenAI’s May 2024 and Feb. 2025 model-spec updates that shifted from categorical refusals to sustained, empathetic engagement and removed self-harm from the disallowed-content list.
  • The family says Adam’s chats rose from dozens per day in January to about 300 per day in April 2025, with self-harm content increasing to 17%, and alleges transcripts show more than 1,200 uses of the word “suicide,” crisis referrals in roughly 20% of those exchanges, and some method-specific guidance.
  • OpenAI requested a list of memorial attendees and related materials in discovery, a move the family’s lawyers described as intentional harassment.
  • OpenAI says teen wellbeing is a priority and points to safeguards including crisis hotline surfacing, routing sensitive conversations to GPT-5, nudges to take breaks, and new parental controls.
  • Scrutiny is widening with at least seven FTC complaints alleging psychological harm, independent research documenting harmful outputs in extended chats, and a former OpenAI safety researcher highlighting false chatbot assurances of internal escalation.