Risks of Automated Medical Advice
AI Triage Tools Fail to Flag 52% of Medical Emergencies
Mount Sinai study in Nature Medicine reveals critical safety blind spots in AI-driven symptom checkers.

A smartphone screen showing a medical chatbot with a red warning icon, held in a hospital setting.
Photo: Avantgarde News
Researchers at the Icahn School of Medicine at Mount Sinai have identified significant safety gaps in AI tools used for medical triage [1]. A study published in Nature Medicine found that these systems frequently fail to direct users toward emergency care during life-threatening situations [1][2]. In testing 60 realistic clinical scenarios, the AI under-triaged 52% of cases that physicians determined required immediate emergency intervention [2][3]. The evaluation focused on "ChatGPT Health," a consumer tool launched in January 2026 [1]. While the system correctly identified "textbook" emergencies like strokes, it struggled with nuanced crises such as impending respiratory failure and diabetic ketoacidosis [2]. In many instances, the AI identified dangerous symptoms in its own text but still advised users to wait 24 to 48 hours for an evaluation rather than seeking help immediately [3]. Researchers also flagged inconsistent safeguards for mental health crises [1]. The study revealed that suicide-risk alerts were "inverted," often appearing for lower-risk scenarios while failing to trigger when users described specific plans for self-harm [2]. Experts emphasize that these tools should currently supplement, rather than replace, professional clinical judgment [1].
Editorial notes
Transparency note
Drafted with LLM; human-edited
- AI assisted
- Yes
- Human review
- Yes
- Last updated
Risk assessment
Reviewed for sourcing quality and editorial consistency.
Sources
- 1.↗
mountsinai.org
https://www.mountsinai.org/about/newsroom/2026/research-identifies-blind-spots-in-ai-medical-triage
- 2.↗
nature.com
https://www.nature.com/articles/s41591-026-04297-7
- 3.↗
crainsnewyork.com
https://www.crainsnewyork.com/health-pulse/mount-sinai-study-chatgpt-health-failed-flag-many-life-threatening-cases
Related stories
View allTopics
About the author
Avantgarde News Desk covers risks of automated medical advice and editorial analysis for Avantgarde News.


