
Artificial intelligence has outperformed human physicians in high-stakes emergency medicine scenarios, according to a landmark Harvard University study — raising fresh questions about the future role of AI in clinical decision-making at its most critical junctures.
The findings, published in the journal Science, showed that large language models (LLMs) have, in the words of the study’s authors, moved well beyond established benchmarks for clinical reasoning. Independent experts welcomed the results as a meaningful advance in the field.
At the heart of the research was an experiment involving 76 patients admitted to the emergency room of a Boston hospital. Researchers presented both an AI system and pairs of human doctors with identical electronic health records — standard documents typically containing vital signs, basic demographic details, and brief nursing notes explaining the patient’s reason for attendance.
The results were stark. The AI — OpenAI’s o1 reasoning model — correctly identified the exact or near-exact diagnosis in 67% of cases, while the human doctors managed accurate diagnoses only 50% to 55% of the time. The AI’s advantage was most pronounced precisely in the conditions that define emergency triage: rapid decisions made with limited information and no margin for error.
When more detailed patient data was made available, the AI’s diagnostic accuracy climbed to 82%, compared with a 70–79% accuracy rate for experienced human clinicians — though researchers noted this particular gap did not reach statistical significance.
Beyond the Emergency Room:
The AI’s edge extended beyond immediate diagnosis. In a separate component of the study, the o1 model and 46 human doctors were each asked to evaluate five clinical case studies and devise longer-term treatment plans — covering areas such as antibiotic protocols and end-of-life care pathways. The AI scored 89% against a human average of just 34% when doctors relied on conventional resources such as search engines, a difference researchers described as significant.
Despite the headline-grabbing results, the study’s scope carries important limitations. The trials were confined entirely to text-based patient data — the kind that can be communicated through written records. Crucially, the AI was never required to assess the non-verbal dimensions of clinical medicine: a patient’s visible distress, physical appearance, or the instinctive human read of a person in crisis.
African Journalists And Editors Can Now Apply For MTN’s Pan-African Media Innovation Programme
In that sense, the AI functioned less as a frontline clinician and more as a sophisticated second-opinion consultant reviewing case notes from a distance — a distinction that researchers and medical professionals say must be kept firmly in view as the technology continues to develop.
“I don’t think our findings mean that AI replaces doctors,” said Arjun Manrai, one of the lead authors of the study who heads an AI lab at Harvard Medical School. “I think it does mean that we’re witnessing a really profound change in technology that will reshape medicine.”
The rise of diagnostic AI will not spell the end of the physician, but it will fundamentally reshape the relationship between doctor and patient. One of the researchers involved in the Harvard study envisions what they describe as a “triadic care model” emerging over the coming decade — a three-way clinical partnership between patient, human doctor, and artificial intelligence.
The study produced at least one striking illustration of that potential. A patient arrived presenting with a pulmonary embolism — a blood clot on the lungs — and their condition was deteriorating. The attending physicians concluded the anti-coagulant medication was failing. The AI reached a different conclusion. Cross-referencing the patient’s full medical history, it flagged that a diagnosis of lupus had been overlooked as a likely driver of the worsening lung inflammation.
The AI, it turned out, was correct. It is precisely this capacity — to hold an entire patient history in view simultaneously and surface connections that time-pressured clinicians may miss — that researchers say makes AI a potentially powerful complement to human clinical judgment.
Mercedes-Benz GLC 400 4MATIC With EQ Technology Is Here — And It’s Fully Electric
The technology is not merely theoretical. Research published last month found that nearly one in five American physicians are already using AI to assist with diagnosis. In the United Kingdom, a Royal College of Physicians survey found that 16% of doctors are using AI tools daily, with a further 15% doing so weekly. Clinical decision-making ranked among the most commonly cited applications.
Yet adoption has not come without anxiety. The same survey identified AI error and questions of liability as the foremost concerns among British doctors — a worry that sits uncomfortably alongside the billions of pounds being channelled into AI healthcare companies globally.
Expert Reaction: Promise and a Pointed Warning
Senior academics offered a broadly encouraging but carefully measured assessment of the findings.
Professor Ewen Harrison, co-director of the University of Edinburgh’s Centre for Medical Informatics, said the study marked a genuine turning point, arguing that AI systems are no longer simply passing medical exams or performing well on artificial test cases, but are beginning to function as credible second-opinion tools — particularly valuable when clinicians need to consider a broader differential diagnosis and reduce the risk of missing something critical.
A more cautionary note came from Dr Wei Xing, assistant professor at the University of Sheffield’s School of Mathematical and Physical Sciences. Some elements of the study, he suggested, pointed to a subtler risk: that doctors may unconsciously defer to the AI’s conclusions rather than forming independent clinical judgements.
As AI becomes more deeply embedded in routine healthcare settings, that tendency — quiet, incremental, and easy to overlook — could grow into one of the more consequential challenges the profession will need to confront.