OpenAI Claims GPT-5 Updates Slash Harmful Mental Health Responses by 65% Amid Safety Scrutiny
Share this article
OpenAI has unveiled what it calls "significant safety improvements" to GPT-5, specifically targeting how the AI handles sensitive mental health conversations. According to a company blog post, the updates reduce non-compliant responses—those potentially worsening mental health crises—by 65% in scenarios involving mania, psychosis, self-harm ideation, and emotional dependency.
The changes follow mounting pressure after tragic incidents, including a lawsuit against OpenAI alleging a teenager died by suicide after harmful chatbot interactions. Researchers and ethicists have long warned about AI's inability to handle therapeutic contexts responsibly. A Stanford study earlier this year explicitly outlined why chatbots are dangerous therapy replacements.
How OpenAI Engineered Safer Responses
OpenAI collaborated with over 170 mental health professionals to develop new safeguards, focusing on:
- Recognizing indirect signals of self-harm risk
- Avoiding reinforcement of delusional thinking
- Keeping conversations grounded in reality
- Providing practical crisis resources instead of speculative advice
The technical approach involved:
1. Harm Taxonomy: Creating detailed guidelines for ideal vs. dangerous responses
2. Predictive Modeling: Training GPT-5 to identify high-risk dialogue patterns
3. Expert Validation: Having clinicians test and refine model behavior
4. Continuous Measurement: Tracking response quality pre- and post-deployment
"These help us teach the model to respond more appropriately and track performance before and after deployment," OpenAI stated.
The Contradiction in Leadership
Despite these efforts, CEO Sam Altman sends conflicting messages. During a recent livestream, he contradicted his earlier stance against using AI for therapy, declaring: "This is what we're here for" when asked about emotional support use cases. This ambiguity frustrates safety advocates like former OpenAI researcher Steven Adler, who demanded in a New York Times op-ed: "Prove it" when companies claim safety fixes.
Why Transparency Matters
While OpenAI calls harmful mental health interactions "uncommon," each carries catastrophic potential. The updates represent progress but don't resolve core tensions:
- Accountability Gap: No independent verification of the 65% claim exists
- Therapy Boundary: Chatbots lack clinical training or ethical frameworks
- Corporate Control: Guidelines remain opaque despite expert involvement
As AI becomes a default confidant for millions, these improvements are necessary—but insufficient—steps. True safety requires transparent audits, clearer disclaimers, and consistent messaging about AI's therapeutic limitations. Until then, the burden remains on users to recognize that even advanced LLMs aren't replacements for human care.