
A new study out of Harvard is forcing one of the biggest questions in modern healthcare back into the spotlight: what happens when AI starts outperforming doctors in real clinical decisions?
According to research published in Science and reported by major outlets, an advanced AI system based on OpenAI’s reasoning models outperformed human physicians in emergency room triage, one of the most critical and time-sensitive stages of medical care.
In tests involving real patient cases, the AI correctly identified diagnoses about 67% of the time, compared to 50–55% accuracy for doctors working under the same conditions.
When given more detailed patient information, the gap narrowed but still favoured the AI, which reached around 82% accuracy, slightly ahead of experienced clinicians.
That alone would be headline-worthy.
But the deeper implication is what those numbers represent.
Emergency triage is one of the hardest environments in medicine fast-paced, high-pressure, and often operating with incomplete information. It’s where decisions can mean the difference between life and death, and where even experienced physicians can struggle with uncertainty.
And that’s exactly where AI performed best.
Researchers found that the system’s advantage was most pronounced in rapid decision-making scenarios with limited data, suggesting that large language models are particularly effective at synthesizing fragmented information into actionable insights.
The AI also significantly outperformed doctors in treatment planning exercises, scoring 89% compared to 34% for physicians when asked to propose longer-term care strategies.
That’s not just a marginal improvement. It’s a structural one. And it points to something larger happening inside AI.
Modern models are no longer just retrieving information they are reasoning through problems, identifying patterns, and connecting medical knowledge in ways that resemble and in some cases exceed human clinical thinking.
Still, the study comes with important limitations.
The AI was tested primarily on text-based patient records, not live clinical interactions. It did not assess the system’s ability to read physical symptoms, interpret body language, or respond to emotional cues all of which remain central to real-world medicine.
That distinction matters. Because medicine isn’t just diagnosis. It’s judgment, communication, and trust.
Even the researchers behind the study are clear: AI is not replacing doctors anytime soon. Instead, they see it becoming part of a “triadic care model” doctor, patient, and AI working together.
But even that framing represents a major shift.
For decades, medical expertise has been defined by human training, experience, and intuition. Now, for the first time, there’s credible evidence that machines can match and in some cases surpass that expertise in specific domains.
That raises difficult questions.
If AI can diagnose more accurately in certain scenarios, should it be involved in every clinical decision? If it makes a mistake, who is responsible? And perhaps most importantly which is how much should doctors rely on it?
There are also broader risks.
Overreliance on AI could lead to deskilling, where clinicians become dependent on systems they don’t fully understand. At the same time, underuse could mean ignoring tools that could significantly improve patient outcomes.
The balance will be hard to strike. What’s clear is that this study marks a turning point.
Healthcare has long been seen as one of the most resistant sectors to AI disruption not because the technology wasn’t capable, but because the stakes are so high.
That barrier is starting to break. And while AI may not replace doctors, it is rapidly becoming something just as important: A second brain in the room.
One that doesn’t get tired, doesn’t miss patterns easily, and increasingly gets the answer right.
Discover more from TechBooky
Subscribe to get the latest posts sent to your email.







