• AI Search
  • Cryptocurrency
  • Earnings
  • Enterprise
  • About TechBooky
  • Submit Article
  • Advertise Here
  • Contact Us
TechBooky
  • African
  • AI
  • Metaverse
  • Gadgets
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
  • African
  • AI
  • Metaverse
  • Gadgets
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
TechBooky
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Home Artificial Intelligence

Study Finds ChatGPT Health Often Misses Emergencies, Fuels Safety Concerns

Paul Balo by Paul Balo
February 28, 2026
in Artificial Intelligence
Share on FacebookShare on Twitter

ChatGPT Health, OpenAI’s dedicated health feature, is facing sharp criticism after an independent study found it frequently failed to recognise medical emergencies and often gave advice that could delay critical care.

Researchers reported that in more than half of urgent cases, the system did not advise an immediate hospital visit, prompting experts to warn the tool could “feasibly lead to unnecessary harm and death”.

The evaluation, published in the February edition of the journal Nature Medicine, is described as the first independent safety assessment of ChatGPT Health. OpenAI launched the feature to limited users in January, promoting it as a way to “securely connect medical records and wellness apps” to generate health advice. According to the company, more than 40 million people ask ChatGPT for health-related information every day.

Lead author Dr Ashwin Ramaswamy and colleagues built 60 realistic patient scenarios that ranged from mild conditions to true emergencies. Three independent physicians reviewed each scenario and agreed, using clinical guidelines, on the level of care needed.

The team then queried ChatGPT Health with each scenario under different conditions such as changing the patient’s gender, adding lab results, or including comments from family members generating nearly 1,000 AI responses. These were compared with the physicians’ recommended actions.

The system performed well in straightforward, clearly defined emergencies such as strokes or severe allergic reactions. But it struggled in more ambiguous or complex cases. In one asthma scenario, ChatGPT Health correctly identified early warning signs of respiratory failure, yet still advised the patient to wait rather than seek emergency treatment.

Overall, in 51.6% of cases where doctors agreed a person needed to go to hospital immediately, ChatGPT Health instead suggested staying home or booking a routine appointment. Alex Ruani, a doctoral researcher in health misinformation mitigation at University College London who was not involved in the study, called that finding “unbelievably dangerous”.

“If you’re experiencing respiratory failure or diabetic ketoacidosis, you have a 50/50 chance of this AI telling you it’s not a big deal,” Ruani said, warning that the reassurance offered by such systems could be deadly if it delays urgent care. In one simulation, she noted, eight times out of 10 — 84% of runs the system sent a suffocating woman to a future appointment “she would not live to see”.

The study also found the model frequently over-reacted in the opposite direction: 64.8% of people described in the scenarios as completely safe were told to seek immediate medical care. Ruani argued that beyond mis-triaging individuals, such behaviour could drive unnecessary pressure on health services.

The researchers observed that ChatGPT Health’s recommendations were highly sensitive to contextual details. The platform was nearly 12 times more likely to minimise symptoms when the simulated patient mentioned that a “friend” believed the issue was not serious. That kind of susceptibility to offhand comments is a concern for those studying AI safety.

“It is why many of us studying these systems are focused on urgently developing clear safety standards and independent auditing mechanisms to reduce preventable harm,” Ruani said.

Ramaswamy, a urology instructor at the Icahn School of Medicine at Mount Sinai in the US, highlighted particular concern around how the system handled suicidal ideation. In one test scenario, a 27-year-old patient reported thoughts of taking “a lot of pills”. When that information was presented alone, ChatGPT Health consistently displayed a crisis intervention banner linking to suicide support services.

But when the researchers added normal lab results with the patient’s words and severity unchanged the behaviour shifted. In 16 attempts under those conditions, the suicide crisis banner did not appear at all. “A crisis guardrail that depends on whether you mentioned your labs is not ready,” Ramaswamy said, adding that such inconsistent protections could be “arguably more dangerous than having no guardrail at all, because no one can predict when it will fail.”

Beyond immediate safety, experts see broader systemic and legal implications. Paul Henman, a digital sociologist and policy specialist at the University of Queensland, described the study as “a really important paper”.

He warned that if ChatGPT Health were widely used in homes, it could both increase unnecessary visits for low-level issues and fail to send people to urgent care when needed, potentially leading to preventable harm and deaths. Henman also pointed to emerging legal risks, noting that cases are already being brought against technology companies in relation to suicide and self-harm after interactions with AI chatbots.

For Henman, a key problem is opacity. “It is not clear what OpenAI is seeking to achieve by creating this product, how it was trained, what guardrails it has introduced and what warnings it provides to users,” he said. “Because we don’t know how ChatGPT Health was trained and what the context it was using, we don’t really know what is embedded into its models.”

An OpenAI spokesperson said the company welcomes independent research evaluating AI systems in healthcare, but argued that the study does not fully reflect how people use ChatGPT Health in practice. The spokesperson added that the model is continually updated and refined.

Ruani countered that even though the research relied on simulated but realistic scenarios, “a plausible risk of harm is enough to justify stronger safeguards and independent oversight”.

With tens of millions of people reportedly turning to ChatGPT for health questions each day, the findings sharpen a debate over how AI tools should be positioned in medicine: as information aids with strict limits, or as more active participants in triage and decision-making. For the researchers behind this study, the answer is clear for now; any system that can tell a suffocating patient to wait days for an appointment needs far stronger checks before being relied on in real-world care.

Related Posts:

  • ChatGPT-Health-3
    OpenAI Launches ChatGPT Health to Power AI-Driven Healthcare
  • 1536x864_cmsv2_85cb6ed8-f81c-5f21-8cb1-78ee23597afe-9605926
    ChatGPT Now Available for Healthcare Providers, OpenAI Says
  • bb3f26e0-82f2-11f0-b74e-adf885bcdc20
    ChatGPT Linked to Suicides in Lawsuits Against OpenAI
  • Screenshot-2026-01-13-at-12.10.58-PM
    ChatGPT Health set for upgrade as OpenAI acquires Torch
  • ChatGPT-Will-Now-Remember-Who-You-Are
    OpenAI Rolls Out New Memory Feature For ChatGPT Plus Users
  • Blog_Asset_Wellbeing
    OpenAI Restricts Personal Advice, adds ChatGPT Break…
  • OpenAI_StudyMode_Flow_7-29
    OpenAI Introduces ChatGPT Study Mode
  • openai
    OpenAI Hires Preparedness Lead to Strengthen Model Safety

Discover more from TechBooky

Subscribe to get the latest posts sent to your email.

Tags: AIChatGPTChatGPT Health
Paul Balo

Paul Balo

Paul Balo is the founder of TechBooky and a highly skilled wireless communications professional with a strong background in cloud computing, offering extensive experience in designing, implementing, and managing wireless communication systems.

BROWSE BY CATEGORIES

Receive top tech news directly in your inbox

subscription from
Loading

Freshly Squeezed

  • Study Finds ChatGPT Health Often Misses Emergencies, Fuels Safety Concerns February 28, 2026
  • ChatGPT Approaches 1 Billion Weekly Users After Rapid Growth February 28, 2026
  • Samsung Rolls Out Satellite Messaging for Galaxy Phones February 28, 2026
  • OpenAI & Google Tech Workers Push Back Against Military AI Contracts February 28, 2026
  • Ultrahuman’s Ring Pro Pushes Smart Ring Battery Life to 15 Days February 28, 2026
  • OpenAI raises $110B from Amazon and others at $730B valuation February 27, 2026
  • Threads Tests Shortcut For Quick DM Conversations February 27, 2026
  • Google Unveils Nano Banana 2 With Faster AI Imaging & Better Text February 27, 2026
  • DeepSeek Blocks US Chipmakers from Its New AI Model February 27, 2026
  • Microsoft Edge to Auto-Launch Copilot from Outlook Links February 27, 2026
  • Warner Bros. Discovery Eyes Paramount-Skydance Deal to Take on Netflix February 27, 2026
  • Block Cuts 4,000 Jobs as Dorsey Warns AI Will Reshape Companies February 27, 2026

Browse Archives

February 2026
MTWTFSS
 1
2345678
9101112131415
16171819202122
232425262728 
« Jan    

Quick Links

  • About TechBooky
  • Advertise Here
  • Contact us
  • Submit Article
  • Privacy Policy
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
  • African
  • Artificial Intelligence
  • Gadgets
  • Metaverse
  • Tips
  • AI Search
  • About TechBooky
  • Advertise Here
  • Submit Article
  • Contact us

© 2025 Designed By TechBooky Elite

Discover more from TechBooky

Subscribe now to keep reading and get access to the full archive.

Continue reading

We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.