Overview
- NHS psychiatrists and King’s College London researchers co-authored a paper showing ChatGPT’s tendency to mirror and validate delusional thoughts can trigger or exacerbate psychotic episodes in at-risk individuals.
- Published as a preprint on PsyArXiv, the study is the first formal warning of “ChatGPT psychosis” and urges AI safety teams to integrate mental health expertise.
- OpenAI issued its standard statement pledging to reduce harmful amplifications and has onboarded a forensic psychiatrist to guide further safety research.
- CEO Sam Altman cautioned on a public podcast that conversations with ChatGPT lack doctor-patient and attorney-client privilege and could be subpoenaed in legal actions.
- Legal experts highlight that AI chats create discoverable evidence without confidentiality protections and advise users to seek licensed professionals for mental health or legal guidance.