Over a million people every week show ‘suicidal intent’ on ChatGPT

Published October 29, 2025
The ChatGPT app icon on a smartphone in this illustration taken on October 27, 2025. — Reuters
The ChatGPT app icon on a smartphone in this illustration taken on October 27, 2025. — Reuters

SAN FRANCISCO: More than a million ChatGPT users each week send messages that include “explicit indicators of potential suicidal planning or intent”, according to a blog post published by OpenAI on Monday.

The post said more than 800 million people use ChatGPT every week, this translates to about 1.2 million people. The company also estimates that approximately 0.07 per cent of active weekly users show possible signs of mental health emergencies related to psychosis or mania — meaning slightly fewer than 600,000 people.

The issue came to the fore after California teenager Adam Raine died by suicide earlier this year. His parents filed a lawsuit claiming ChatGPT provided him with specific advice on how to kill himself.

OpenAI has since increased parental controls for ChatGPT and introduced other guardrails, including expanded access to crisis hotlines, automatic rerouting of sensitive conversations to safer models, and gentle reminders for users to take breaks during extended sessions.

OpenAI said it has also updated its ChatGPT chatbot to better recognise and respond to users experiencing mental health emergencies, and is working with more than 170 mental health professionals to significantly reduce problematic responses.

The finding, part of an update on how the chatbot handles sensitive conversations, is one of the most direct statements from the artificial intelligence giant on the scale of how AI can exacerbate mental health issues.

In addition to its estimates on suicidal ideations and related interactions, OpenAI also said that about 0.07pc of users active in a given week ­— about 560,000 of its touted 800m weekly users — show “possible signs of mental health emergencies related to psychosis or mania”.

The post cautioned that these conversations were difficult to detect or measure, and that this was an initial analysis.

As OpenAI releases data on mental health issues related to its marquee product, the company is facing increased scrutiny following a highly publicised lawsuit from the family of a teenage boy who died by suicide after extensive engagement with ChatGPT. The Federal Trade Commission last month additionally launched a broad investigation into companies that create AI chatbots, including OpenAI, to find how they measure negative impacts on children and teens.

OpenAI claimed in its post that its recent GPT-5 update reduced the number of undesirable behaviors from its product and improved user safety in a model evaluation involving more than 1,000 self-harm and suicide conversations. The company did not immediately return a request for comment.

“Our new automated evaluations score the new GPT‑5 model at 91pc compliant with our desired behaviors, compared to 77pc for the previous GPT‑5 model,” the company’s post reads.

OpenAI stated that GPT-5 expanded access to crisis hotlines and added reminders for users to take breaks during long sessions. To make improvements to the model, the company said it enlisted 170 clinicians from its Global Physician Network of health care experts to assist its research over recent months, which included rating the safety of its model’s responses and helping write the chatbot’s answers to mental-health related questions.

Published in Dawn, October 29th, 2025

Opinion

Editorial

Pathways to peace
Updated 27 Apr, 2026

Pathways to peace

NEGOTIATIONS to hammer out the 2015 Iran nuclear agreement took nearly two years before a breakthrough was achieved....
Food-insecure nation
27 Apr, 2026

Food-insecure nation

A NEW UN-backed report has listed Pakistan among 10 countries where acute food insecurity is most concentrated. This...
Migration toll
27 Apr, 2026

Migration toll

THE world should not be deceived by a global migration count lower than the highest annual statistics on record —...
Immunity gap
Updated 26 Apr, 2026

Immunity gap

Pakistan’s Big Catch-Up campaign showed progress but also exposed the scale of gaps in routine immunisation.
Danger on repeat
26 Apr, 2026

Danger on repeat

DISASTERS have typically been framed as acts of nature. Of late, they look increasingly like tests of preparedness...
Loose lips
26 Apr, 2026

Loose lips

PAKISTANIS have by now gained something of an international reputation for their gallows humour, but it seems that...