AI Health Chatbots Fall Short on Real-World Medical Advice

TL;DR Summary
UK researchers tested GPT-4o, Llama 3, and Command R+ against a control group and found AI health chatbots identified health problems only about one-third of the time and suggested a correct course of action roughly 45% of the time—no better than internet searches. They attribute the gap to miscommunication between humans and AI and warn that chatbots can give wrong diagnoses or miss urgent care cues. With about 1 in 6 US adults asking chatbots about health monthly, the study argues AI isn’t ready to replace clinicians.
- AI Chatbots Give Bad Health Advice, Research Finds Barron's
- A.I. Is Making Doctors Answer a Question: What Are They Really Good For? The New York Times
- Using AI for medical advice 'dangerous', study finds BBC
- Reliability of LLMs as medical assistants for the general public: a randomized preregistered study Nature
- AI-powered apps and bots are barging into medicine. Doctors have questions. Reuters
Reading Insights
Total Reads
1
Unique Readers
1
Time Saved
3 min
vs 4 min read
Condensed
86%
637 → 87 words
Want the full story? Read the original article
Read on Barron's