Thoughts on Healthcare Markets & Technology

Thoughts on Healthcare Markets & Technology

Clinical AI & Patient Care

What the Harvard ER Study Says About o1 Beating Doctors at Diagnosis, Why It Means Differential Diagnosis Just Stopped Being a Scarce Cognitive Asset, and Where the Money Goes Next

May 05, 2026
∙ Paid

Video Preview

🎧 Listen on Spotify, Apple Podcasts, or the Web Player

0:00
-8:37
Audio playback is not supported on your browser. Please upgrade.

Full episode for paid subscribers available below paywall.

Abstract

A Harvard-led team published in Science evaluating OpenAI’s o1 reasoning model against board-certified physicians on real cases pulled from a Boston emergency department. The headline numbers got picked up by Vox, the Guardian, the SF Chronicle, STAT, and roughly every healthcare Twitter account that has ever quote-tweeted Eric Topol. The popular takeaway is “AI beat doctors at diagnosis.” The popular takeaway is wrong, or at least incomplete enough to be useless for anyone making product, policy, or capital allocation decisions.

What follows is a closer read of what the study actually showed, what the most overlooked finding really implies for clinical workflow design, and why the investable thesis here has very little to do with “AI diagnostic startups.”

Quick coverage map:

  • 76-patient Boston ED dataset, multi-stage evaluation from triage through admission

  • ~67% diagnostic accuracy for o1 at triage vs ~50-55% for physicians, ~80%+ with full context

  • The buried lede: physicians using AI did not outperform AI alone

  • Three-layer physician value stack and which layer just collapsed

  • Triage inversion thesis and why front-door medicine is the highest-leverage deployment

  • Search vs reasoning reframe, and what it means for product design

  • The liability bottleneck nobody is pricing in yet

  • Why the obvious “AI diagnostics” bet is the wrong layer to invest at

Table of Contents

1. Video Preview

2. Podcast, Part I (Free)

3. The setup and what the tweet actually means

4. The numbers, and the caveats most takes skip

5. Why physicians plus AI did not beat AI alone

6. The three-layer model of physician value

7. Diagnosis as infrastructure

8. The triage inversion thesis

9. Search vs reasoning, the real reframe

10. The AI-first differential workflow

11. Liability as the next bottleneck

12. Where the dollars actually flow

13. The punchline

14. The setup and what the tweet actually means

The study tested o1, OpenAI’s reasoning-class model, against attending and resident physicians on 76 cases drawn from the emergency department at a major Boston academic medical center. Cases were structured into stages mirroring the actual ED workflow, starting with triage-level information (chief complaint, vitals, brief history) and progressing through ED workup and ultimately admission. At each stage, both the model and the physician were asked to produce a differential diagnosis and rank candidates. Physicians were blinded to the model’s outputs in the comparison arm.

The methodological detail that matters: this is not a full clinical interaction. There is no patient sitting in front of either the human or the model. There is no live imaging review, no live labs interpretation in the sense of looking at the raw values cycle by cycle, no nursing handoff, no family history clarifying conversation in real time. The inputs are text-based vignettes constructed from chart data, which is closer to a chart review exercise than to actual clinical care. Anyone who has worked in or around ED operations knows that the case on paper looks very different from the case in the room.

The Science Magazine writeup framed this as evidence that o1 reasons clinically at or above physician level. The Science Media Centre piece, which is the most useful secondary read in the entire ecosystem here, pushed back hard on that framing. So did STAT. So did several of the senior clinical informaticists who weighed in on Bluesky and LinkedIn after the paper dropped. The pushback is correct on the merits but does not actually defang what the study showed. What the study showed is still important. It is just important for reasons different from the ones being tweeted about.

The numbers, and the caveats most takes skip

At the triage stage, where the model and the physicians were given only the bare initial intake information, o1 landed correct primary diagnosis around sixty-seven percent of the time. Physicians, both attendings and residents, were closer to fifty to fifty-five percent. As progressively more context got added through the workup stages, both groups improved, with the model and senior physicians converging in the eighty percent plus range by the time full ED workup data was on the table.

A few things to immediately notice. First, the gap is biggest where the information is sparsest. That is counterintuitive only if the prior assumption is that physicians have some special intuition that gets activated at the bedside. The data suggests something closer to the opposite: physicians underperform exactly where breadth of recall is most useful, because at triage the cognitive task is generating a wide net of plausible diagnoses, and humans are systematically narrow-net generators. There is a deep clinical reasoning literature on premature closure, the failure mode where the diagnosing physician latches onto an early hypothesis and stops searching the space. Singh and colleagues have published repeatedly on this; their estimates put the population-level diagnostic error rate somewhere around twelve million American adults per year experiencing some form of misdiagnosis, with downstream mortality contributions in the forty to eighty thousand range depending on which methodology you trust. The Boston study is, in effect, a controlled demonstration of why those numbers exist. Doctors are not generating wide enough differentials at the front door.

Second, the convergence at later stages is real and underdiscussed. Once the chart includes labs, imaging reads, and a focused history, the physician advantage at pattern integration largely catches up. The model does not lose ground; the human gains. So the right read is not “AI is better than doctors” but “AI is dramatically better than doctors at the early-information end of the workflow, and roughly comparable once enough data is on the table.” That is a much more interesting finding than the tweet version.

Third, this is text in, text out. There is no imaging interpretation in this study. There is no real-time signal handling. There is no conversational diagnosis where the model would have to decide what to ask next. So the result generalizes to chart-mediated reasoning, not to autonomous clinical care. STAT was right to flag this. The Science Media Centre was right to flag it. Anyone using this paper to argue that physicians are obsolete in any near-term sense is either trolling or has not read the methods section.

Why physicians plus AI did not beat AI alone

The single most important finding in the paper, and the one that has been least discussed publicly, is that the human-plus-AI condition did not outperform AI alone. This deserves to be sat with for a minute. Every health tech pitch deck for the past three years has assumed an additive model: physician judgment plus machine breadth equals better than either alone. That is the entire premise of “copilot” framing. That is the premise of the FDA’s evolving thinking on clinical decision support. That is the premise of the augmented intelligence positioning the AMA has been pushing.

What the data showed is that when physicians had access to the model’s outputs, they did not pick up the additional accuracy the model would have delivered on its own. The most plausible mechanism is automation bias combined with anchoring. When physicians saw the model’s top-ranked diagnoses, they used those as a starting point and pruned, but they also prematurely accepted incorrect rankings and failed to catch model errors as often as one would hope. They sometimes also discounted correct AI suggestions when those suggestions conflicted with the human’s prior. The net effect was a wash, or close enough to a wash that the additive collaboration thesis is now empirically on the back foot.

This finding is not unique to this paper. Radiology has been wrestling with it for the better part of a decade, and there is a meaningful literature on how radiologists with CAD assistance often perform comparably to or marginally worse than radiologists without it, depending on prevalence, signal quality, and trust calibration. The Boston ED finding is the same shape of result but in cognitive diagnosis rather than image read. It suggests that “human in the loop” is not automatically a quality multiplier, and may, under certain configurations, be a quality reducer relative to AI alone. That is a deeply uncomfortable conclusion for anyone whose product strategy assumes the physician seat is the one paying the bill.

The three-layer model of physician value

To make sense of where this lands economically, it helps to decompose what physicians actually get paid for. The traditional value stack has three layers, roughly. The first layer is diagnosis: figuring out what is wrong. The second is decision-making: choosing the appropriate treatment, weighing risk, integrating patient preference and goals. The third is execution: actually performing care, whether that is procedural skill, longitudinal management, or care coordination.

Layer one is what the study just commoditized. Or, more precisely, the study added another data point to a trend line that has been visible since GPT-4 cleared roughly ninety percent on USMLE Step questions in 2023. The benchmarks have been signaling for two years that knowledge recall and pattern matching across a wide medical literature is no longer a scarce skill. The Boston study moved the goalposts from “can pass a board exam” to “can outperform attendings on real cases at the high-uncertainty end of the workflow.” That is a meaningful jump in evidentiary strength.

Layer two and layer three are still firmly in human hands. Decision-making in the face of uncertainty, particularly when patient values, social context, payer constraints, and risk tolerance are all in play, is not something current LLMs do credibly. They can summarize options. They cannot sit with a sixty-eight-year-old recently widowed patient and decide whether aggressive chemo is the right call. Execution, especially procedural execution and longitudinal relationship work, is even further from being touched.

So the right read on the study is not “doctors are obsolete.” The right read is “the diagnostic step, which has historically been a meaningful chunk of why physicians are paid what they are paid, just got pulled out of the value stack and turned into a near-zero marginal cost utility.” Whatever fraction of physician comp is implicitly compensating for layer one is now exposed to repricing. Estimating that fraction is hard, but if you look at how E&M coding levels work, with diagnostic complexity being a primary driver of code level and therefore reimbursement, it is not small. The 2021 office visit E&M revisions explicitly anchored level selection on medical decision-making complexity, of which differential diagnosis is a core input. When the marginal cost of generating a high-quality differential goes to roughly zero, the question of why the physician is being paid at level five becomes more pointed.

Diagnosis as infrastructure

The cleaner mental model for what is happening is that diagnosis is shifting from being an artisanal individual skill to being a piece of infrastructure. This is not a new pattern. It happened to security analysis when Bloomberg terminals went mainstream in the 1980s. Before Bloomberg, knowing the price of a corporate bond was a skill, and the analysts who could remember the spreads got paid for it. After Bloomberg, that knowledge was a utility, and analyst comp moved up the stack to interpretation, modeling, and origination. It happened to legal research when Westlaw and LexisNexis got their full text databases working. Before Westlaw, finding the relevant case was a skill, and junior associates billed hours for it. After, finding the case was a utility, and the human work moved to argument construction and judgment. It happened to radiology, partially, with PACS and the early generation of CAD tools. The radiologist’s value moved from “find the lesion” toward “integrate the finding into clinical context.”

Diagnosis is now in the early phase of the same shift. The right way to think about o1, or whatever next-generation model gets deployed at the bedside, is as the medical equivalent of a Bloomberg terminal for differential diagnosis. The value question becomes: who owns the terminal, and who controls distribution into the workflow? Not: who is the best diagnostician.

If that framing is correct, then the strategic moves that matter are the ones that determine the distribution layer. The model itself becomes a commodity over a long enough horizon. The training data, the integration into the EHR, the specific UX in the order entry flow, the audit trail, and the contractual relationship with the health system are where defensibility lives. This is consistent with how Epic and Oracle Health have been positioning, and with why Microsoft’s twenty-billion-dollar Nuance acquisition has aged considerably better than skeptics predicted at the time. Distribution into the moment of care is the moat. Model quality is table stakes.

User's avatar

Continue reading this post for free, courtesy of Thoughts on Healthcare.

Or purchase a paid subscription.
© 2026 Healthcare Markets & Technology · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture