epocrates logo
epocrates logo
epocrates logo
  • 0

Journal Article Synopsis

Nat Med

ChatGPT under-triages high‑risk emergencies in structured evaluation

March 5, 2026

card-image

In a stress test of ChatGPT Health, investigators evaluated 960 triage responses spanning 21 clinical domains. Performance followed an inverted U‑shape: the system was most error‑prone in non‑urgent (35%) and emergency scenarios (48%). Among gold‑standard emergencies, it under‑triaged 52%—including diabetic ketoacidosis and emerging respiratory failure—while correctly identifying classic emergencies such as stroke and anaphylaxis. Anchoring bias from minimized symptoms markedly shifted recommendations toward less urgent care (odds ratio, 11.7). Crisis‑intervention messaging also triggered inconsistently across suicidal ideation cases. No significant effects were observed for race, gender, or access barriers. Authors call for prospective validation before broad consumer use.

Clinical takeaway: AI triage tools may miss high‑risk emergencies and show bias‑susceptible variability; clinicians should caution patients against relying on them for urgent decision‑making.

Source:

Ramaswamy A, et al. (2026, February 23). Nat Med. ChatGPT Health performance in a structured test of triage recommendations. https://pubmed.ncbi.nlm.nih.gov/41731097/

Trending icon

TRENDING THIS WEEK

EPOCRATES CME

View Catalog

view all CME activities
learn more about epocrates plus
Clinical FAQ icon

Clinical FAQs

Check out the answers to frequently asked questions about our clinical content.

Download Epocrates from the App StoreDownload Epocrates from the Play Store
About UsFeaturesBusiness SolutionsHelp & Feedback
© 2026 epocrates, Inc.   Terms of UsePrivacy PolicyEditorial PolicyDo Not Sell or Share My Information