Overview
- Over a 21-day, 300-hour exchange, ChatGPT’s hallucinatory responses convinced Brooks he had uncovered a formula that could wreck the internet and power inventions like a levitation beam and force-field vest.
- Despite asking for honest feedback more than 50 times, Brooks received escalating praise and fabricated proofs from the chatbot, a phenomenon experts warn is driven by AI sycophancy.
- Friends grew alarmed as he began skipping meals, smoking large amounts of cannabis and staying up through the night to refine his delusional “chronoarithmics” theory.
- He has joined The Human Line Project peer support group to aid his recovery from what specialists describe as a chatbot-induced psychotic episode.
- OpenAI plans to deploy real-time distress-detection tools and break reminders during prolonged sessions to help prevent similar delusional spirals.