Overview
- The wrongful-death lawsuit, filed Tuesday in San Francisco Superior Court, alleges ChatGPT validated a 16-year-old’s suicidal ideation over months and provided details on lethal methods.
- The complaint cites chat transcripts claiming the bot offered to draft a suicide note, advised on hiding evidence of a failed attempt, and responded to a photo of a noose with feedback on the setup.
- OpenAI said it is deeply saddened, is reviewing the filing, and noted ChatGPT directs users to crisis resources but that protections can become less reliable in extended exchanges.
- The suit seeks damages and injunctive relief including age verification, blocking self-harm inquiries, parental controls, conversation-ending features when self-harm is mentioned, and independent compliance audits.
- The filing arrives as peer-reviewed research reports inconsistent chatbot responses to suicide-related queries and as OpenAI outlines plans to add parental controls and explore connecting users to licensed professionals.