<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Healthcare Markets & Technology: Clinical AI & Patient Care]]></title><description><![CDATA[AI applications in clinical settings]]></description><link>https://www.onhealthcare.tech/s/clinical-ai-and-patient-care</link><image><url>https://substackcdn.com/image/fetch/$s_!Wr7p!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7280dcad-05ec-4956-97c3-9faecb031e7a_1024x1024.png</url><title>Healthcare Markets &amp; Technology: Clinical AI &amp; Patient Care</title><link>https://www.onhealthcare.tech/s/clinical-ai-and-patient-care</link></image><generator>Substack</generator><lastBuildDate>Thu, 23 Apr 2026 09:06:25 GMT</lastBuildDate><atom:link href="https://www.onhealthcare.tech/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Healthcare Markets & Technology]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rustythreek1@gmail.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rustythreek1@gmail.com]]></itunes:email><itunes:name><![CDATA[Special Interest Media]]></itunes:name></itunes:owner><itunes:author><![CDATA[Special Interest Media]]></itunes:author><googleplay:owner><![CDATA[rustythreek1@gmail.com]]></googleplay:owner><googleplay:email><![CDATA[rustythreek1@gmail.com]]></googleplay:email><googleplay:author><![CDATA[Special Interest Media]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[GPT-Rosalind Lands: What OpenAI’s First Domain-Specific Life Sciences Model, the Codex Life Sciences Plugin & the Trusted Access Program Actually Mean]]></title><description><![CDATA[Table of Contents]]></description><link>https://www.onhealthcare.tech/p/gpt-rosalind-lands-what-openais-first</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/gpt-rosalind-lands-what-openais-first</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Sat, 18 Apr 2026 11:59:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Bkmo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00e23750-9b80-4d0b-86d7-952f2e62caf0_1290x716.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Table of Contents</h2><p>What actually shipped on April 16</p><p>Benchmarks, with the appropriate skepticism</p><p>The plugin is the real story</p><p>Trusted access, biosecurity, and why the gate matters</p><p>Where this lands inside the pharma stack</p><p>What gets compressed, what gets created, what gets killed</p><p>Read-throughs for tools, data, and services companies</p><p>Angel and seed-stage implications</p><p>The quiet part about moats</p><p>Caveats, open questions, and the boring stuff nobody wants to talk about</p><p>So what now</p><h2>Abstract</h2><p>- OpenAI shipped GPT-Rosalind on April 16, 2026, its first purpose-built domain model, aimed at biochemistry, genomics, and protein engineering</p><p>- Access is gated via a trusted-access program; launch partners include Amgen, Moderna, Thermo Fisher Scientific, Allen Institute, plus a Los Alamos collab on protein and catalyst design</p><p>- Claimed benchmark results: 0.751 pass rate on BixBench, beats GPT-5.4 on 6 of 11 LABBench2 tasks, and in a Dyno Therapeutics eval on unpublished RNA sequences, best-of-10 submissions cleared the 95th percentile of human experts on sequence-to-function prediction and roughly 84th percentile on sequence generation</p><p>- A Life Sciences research plugin for Codex connects the model to 50+ scientific tools and public bio databases, which is arguably more commercially important than the model weights themselves</p><p>- Preview phase does not consume tokens or credits for approved orgs, meaning the effective price is zero for the enterprise tier, which will distort willingness-to-pay data across the entire biotech software market for roughly 6 to 12 months</p><p>- Read-through for founders: data-access wrappers, lit-review tools, and protocol-design copilots with no proprietary data are now at existential risk; differentiated wet-lab data, closed-loop experimentation, regulated workflows, and vertical systems of record are relatively safer</p><p>- Read-through for angels: pause any check into a pure RAG-over-PubMed startup, underwrite biotech software against a post-Rosalind baseline rather than a GPT-4-era baseline, and lean into companies producing new, non-public scientific data</p><p>- Caveats: dual-use risk is non-trivial, no fully AI-discovered drug has cleared phase 3, and OpenAI&#8217;s benchmark numbers are self-reported against evals where OpenAI had training-time knowledge of the tasks</p><h2>What actually shipped on April 16</h2><p>OpenAI pushed out three things in a single announcement, and the health tech crowd keeps conflating them. Separating the three is how the analysis gets interesting.</p><p>The first is GPT-Rosalind itself, a frontier reasoning model in a new Life Sciences series. It is designed to support evidence synthesis, hypothesis generation, experimental planning, and multi-step scientific workflows across biochemistry, genomics, and protein engineering. &#65532; Named after Rosalind Franklin, which is a nice bit of historical housekeeping given the Nobel committee&#8217;s 1962 miss. The model is available in ChatGPT, Codex, and the API, but you cannot just sign up for it.</p><p>The second is the gating layer, which OpenAI calls the trusted access program. Eligibility is restricted to qualified enterprise customers in the US working on health-relevant research, with governance and safety oversight controls in place. Launch partners named publicly are Amgen, Moderna, Thermo Fisher Scientific, and the Allen Institute, plus an existing collaboration with Los Alamos National Laboratory on AI-guided protein and catalyst design. During the preview, usage does not consume existing credits or tokens for approved orgs, subject to abuse guardrails. The pricing part is worth staring at for a minute. OpenAI is effectively giving the model away to pharma at the moment, which is a fairly aggressive land grab and is going to wreck price discovery for every startup trying to sell AI-for-biotech software to the same buyers.</p><p>The third is the Life Sciences research plugin for Codex, published to GitHub. The plugin connects models to over 50 scientific tools and data sources, &#65532; including human genetics, functional genomics, protein structure, and clinical evidence data. Quietly, OpenAI said it is also making the connectors and the plugin more broadly available for use with mainline models, not just Rosalind. That matters more than the model itself. More on that below.</p><h2>Benchmarks, with the appropriate skepticism</h2><blockquote><p>The benchmark numbers are noteworthy but worth reading carefully, because every model vendor publishes whatever makes their thing look good.</p></blockquote>
      <p>
          <a href="https://www.onhealthcare.tech/p/gpt-rosalind-lands-what-openais-first">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Goodfire AI and the Billion Dollar Bet on Neural Network Interpretability: Why Reverse Engineering Foundation Models Matters for Health Tech Investors Watching the Life Sciences AI Stack Take Shape]]></title><description><![CDATA[Table of Contents]]></description><link>https://www.onhealthcare.tech/p/goodfire-ai-and-the-billion-dollar</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/goodfire-ai-and-the-billion-dollar</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Fri, 17 Apr 2026 10:19:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!fD4z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Table of Contents</h2><p>Abstract</p><p>The Setup: What Even Is This Company</p><p>The Steam Engine Problem and Why Interpretability Matters Now</p><p>Inside the Ember Platform: What the Tech Actually Does</p><p>The Life Sciences Play: Alzheimer&#8217;s Biomarkers, Evo 2, and Mayo Clinic</p><p>The Business: Funding, Valuation, and Who Wrote the Checks</p><p>The Team Card</p><p>Where This Fits in the Health Tech Investment Landscape</p><p>The Bull Case and the Bear Case</p><p>So What</p><h2>Abstract</h2><p>- Goodfire is a San Francisco based AI research lab and public benefit corporation focused on mechanistic interpretability, the science of reverse engineering neural networks to understand how they work internally</p><p>- Founded in 2023 by Eric Ho (CEO), Dan Balsam (CTO), and Tom McGrath (Chief Scientist, formerly of Google DeepMind&#8217;s interpretability team)</p><p>- Raised $209M total across three rounds: $7M seed (Aug 2024), $50M Series A (Apr 2025, led by Menlo Ventures with Anthropic participating), $150M Series B (Feb 2026, led by B Capital, valued at $1.25B)</p><p>- Core product is Ember, a model design environment that provides programmatic access to neural network internals for feature steering, hallucination reduction, and behavior modification</p><p>- Key health/life sciences milestones: identified novel Alzheimer&#8217;s biomarkers by reverse engineering Prima Mente&#8217;s epigenetic model (first natural science finding from foundation model interpretability), decoded Arc Institute&#8217;s Evo 2 genomic model (published in Nature), collaboration with Mayo Clinic on genomic medicine, and TIME magazine feature (Apr 2026) on genetic disease diagnosis</p><p>- Claimed results: 58% hallucination reduction in LLMs at 90x lower cost than LLM-as-judge approaches, 30% improvement in viable candidate materials from diffusion models</p><p>- ~51 employees as of Jan 2026, team includes researchers from OpenAI, DeepMind, Harvard, Stanford</p><p>- Investors include B Capital, Menlo Ventures, Lightspeed, Anthropic, Salesforce Ventures, Eric Schmidt, DFJ Growth, Wing Venture Capital, South Park Commons</p><p>- Health tech relevance: interpretability positions as a critical enabling layer for any AI system deployed in clinical, diagnostic, or life sciences contexts where &#8220;trust the black box&#8221; is not an acceptable answer</p><h2>The Setup: What Even Is This Company</h2><p>Goodfire is one of those companies that requires you to think about two or three things at once, which is probably why it gets less coverage in health tech circles than it deserves. On the surface, it looks like a pure AI safety play. San Francisco research lab, public benefit corporation, bunch of former OpenAI and DeepMind researchers doing deep technical work on how neural networks function internally. And yeah, that is what they do. But the health and life sciences applications that have come out of this work are some of the most interesting things happening at the intersection of AI and biomedicine right now, and the angel investing community should be paying very close attention to the downstream implications.</p><p>The company was founded in 2023 by Eric Ho, Dan Balsam, and Tom McGrath. McGrath is probably the name that matters most from a credibility standpoint if you care about the research pedigree, because he founded the interpretability team at Google DeepMind before leaving to cofound Goodfire. Balsam serves as CTO and has publicly called interpretability &#8220;the most important problem in the world,&#8221; which is the kind of statement that either makes you roll your eyes or lean in depending on your priors about where AI is headed. Ho is the CEO and the one doing most of the public talking, including a Bloomberg interview where he said what the AI industry is doing right now is &#8220;quite reckless.&#8221; That quote probably did not endear him to the scaling labs, but it tracks with the company&#8217;s overall thesis.</p><p>So what is the thesis? It goes something like this: every major engineering discipline in human history has been gated by fundamental science. You could build steam engines before thermodynamics, but they were wildly inefficient and you could not predictably improve them because nobody understood why they worked. AI is at that exact inflection point. The scaling labs (OpenAI, Google, Anthropic, etc.) are building increasingly powerful systems with very limited understanding of what goes on inside the models. This means nobody can reliably predict when these systems will fail, nobody can surgically fix specific failure modes, and nobody can extract the knowledge that these models have clearly learned from training data but keep locked inside a black box. Goodfire exists to change that by building the science and tooling for mechanistic interpretability, which is basically the discipline of reverse engineering neural networks to figure out what individual components do and how they interact.</p><h2>The Steam Engine Problem and Why Interpretability Matters Now</h2><p>The steam engine analogy that Ho keeps using is actually pretty good, so it is worth sitting with for a second. Before thermodynamics gave engineers a theoretical framework for understanding heat and energy transfer, improving steam engines was basically trial and error. You would change something, see if it worked, change something else. Sound familiar? That is more or less how the entire AI industry trains and fine-tunes models today. You adjust training data, tweak hyperparameters, run RLHF, do some eval benchmarks, and hope for the best. The industry term for this, which Goodfire uses frequently, is &#8220;guess and check.&#8221; Their pitch is that interpretability is the thermodynamics that turns AI development from alchemy into precision engineering.</p><p>This framing lands differently depending on whether you are thinking about chatbots or clinical decision support. If Claude or ChatGPT hallucinates a restaurant recommendation, the stakes are low. If a genomic foundation model makes a pathogenicity prediction that influences a clinical decision, the stakes are very high. And this is where the health tech angle gets interesting, because the FDA and CMS are both moving toward requiring more explainability from AI systems deployed in healthcare settings. The regulatory trajectory is pretty clearly pointing toward a world where &#8220;we do not know why the model made that prediction&#8221; stops being an acceptable answer in clinical contexts. Goodfire is building the toolkit that could become essential infrastructure for anyone trying to deploy AI in regulated health markets.</p><p>The company self-identifies as part of a new category they call &#8220;neolabs,&#8221; which are research-first AI companies pursuing fundamental breakthroughs in training methodology that the scaling labs have mostly neglected because they have been too busy racing to make models bigger. Whether the neolab framing sticks as a category label remains to be seen, but the underlying observation is correct: there has been a massive resource allocation toward making models larger and a relatively tiny investment in understanding them. Ho has pointed out that there are probably fewer than 150 full-time interpretability researchers in the world. For a technology that is being deployed across healthcare, finance, defense, and basically every other consequential domain, that number is absurdly small.</p><h2>Inside the Ember Platform: What the Tech Actually Does</h2><p>The flagship product is called Ember, and it is essentially a model design environment (their term) that gives developers and researchers programmatic access to the internal mechanisms of neural networks. To understand what this means, you need a quick primer on the underlying science.</p><p>Neural networks consist of artificial neurons that individually have simple designs but interact in enormously complex ways. Tens of thousands of neurons might be involved in generating a single prompt response. The challenge is that individual neurons do not map neatly to individual concepts. This is the superposition problem: neurons contribute to multiple features simultaneously, so the conceptual representations inside a model are all tangled up between physical components. The field of mechanistic interpretability has developed tools called sparse autoencoders (SAEs) that can disentangle these representations and extract human-interpretable features from model activations. A feature might correspond to a concept like &#8220;formal tone&#8221; or &#8220;medical terminology&#8221; or &#8220;protein secondary structure.&#8221; It depends entirely on the model and the training data.</p><p>Ember takes these research techniques and packages them into a platform with several practical capabilities. Feature steering lets you tune model internals to shape how an AI model thinks and responds. They have built an &#8220;Auto Steer&#8221; mode that finds relevant features and activation strengths from a short prompt, which basically means you can tell the system what behavior you want changed and it figures out which internal knobs to turn. One of the more compelling demos has been conditional feature steering for jailbreak prevention: by detecting jailbreak patterns and amplifying the model&#8217;s refusal features, they showed dramatically increased robustness to adversarial attacks without affecting normal performance, latency, or cost.</p><p>On the diagnostic side, Ember provides tools for identifying why models behave in specific ways. Their SPD method works by identifying model components that may be involved in generating a response and removing them one by one. If removing a component does not affect the output, researchers can conclude it is not part of the relevant processing chain. Think of it like lesion studies in neuroscience, where you figure out what brain regions do by observing what happens when they are damaged. Same logic, applied to artificial neural networks.</p><p>They also claim a 58% reduction in LLM hallucinations by using interpretability to guide model training, at roughly 90x lower cost per intervention compared to LLM-as-judge approaches, with no degradation on standard benchmarks. If those numbers hold up across diverse deployments, that is a genuinely significant result. Hallucination reduction has been one of the hardest problems in making LLMs production-ready for high-stakes applications, and most existing approaches involve expensive post-hoc filtering or additional model calls that add latency and cost. A method that targets the internal mechanisms responsible for hallucination and fixes them at the training level is a fundamentally different and more elegant approach.</p><h2>The Life Sciences Play: Alzheimer&#8217;s Biomarkers, Evo 2, and Mayo Clinic</h2><p>Alright, here is where things get really interesting for the health tech crowd. Goodfire has three major life sciences collaborations that showcase different aspects of what interpretability can do for biomedicine, and each one represents a different flavor of value creation.</p><p>The Prima Mente collaboration produced what Goodfire calls the first major finding in the natural sciences obtained from reverse engineering a foundation model. Prima Mente built an AI model that analyzes cell-free DNA (cfDNA) fragments to detect Alzheimer&#8217;s disease. cfDNA is DNA that floats freely in the bloodstream after cells die and release their contents, and it carries epigenetic marks that reflect the cellular environment it came from. Prima Mente trained their model (called Pleiades) on this data and got good predictive performance, but could not explain what the model was actually learning. Enter Goodfire. By applying their interpretability toolkit, Goodfire&#8217;s researchers discovered that the model was primarily relying on cfDNA fragment length as a diagnostic signal. This finding was not previously documented in scientific literature. The fragment length pattern represents a novel class of Alzheimer&#8217;s biomarkers surfaced entirely through AI interpretability.</p><p>Think about what happened here. A neural network trained on biological data learned something about disease mechanisms that human scientists had not identified. The knowledge was trapped inside the black box. Interpretability tools opened the box, extracted the insight, and made it available for traditional scientific validation. Goodfire frames this as &#8220;model-to-human knowledge transfer,&#8221; and it is a genuinely new paradigm for scientific discovery. The model becomes a source of testable hypotheses rather than just a prediction machine.</p><p>The Arc Institute collaboration focused on Evo 2, a genomic foundation model trained on DNA sequences. Goodfire decoded Evo 2&#8217;s internal representations and found features that map onto known biological concepts, from coding sequences to protein secondary structure. This work was published in Nature. The interesting thing here is not just that the model learned biology (you would hope it did, given the training data) but that interpretability tools could recover the conceptual structure. They literally found the tree of life embedded in the model&#8217;s activation patterns.</p><p>The Mayo Clinic collaboration, announced in September 2025, takes the genomic interpretability work into a clinical research context. The stated goal is to reverse engineer advanced genomics foundation models to understand what they have learned about genomic relationships, disease mechanisms, and biological processes. Dan Balsam&#8217;s framing of this was pretty direct: generative AI has made enormous progress in modeling complex biological systems, but clinical deployment remains blocked because there is a disconnect between model predictions and real-world biological understanding. Interpretability is the bridge. Mayo Clinic has a financial interest in the technology, which tells you something about how seriously they are taking this.</p><p>Then just this week, TIME magazine ran a feature on Goodfire&#8217;s work with Mayo Clinic researchers using Evo 2 to predict which genetic mutations cause disease and, critically, to explain why. The approach achieved state-of-the-art performance on pathogenicity prediction with interpretable-by-design outputs. Given that the cost of genome sequencing has dropped to around $100 per genome, the bottleneck is increasingly shifting from data generation to data interpretation. A tool that can predict pathogenic variants and provide mechanistic explanations is exactly what the precision medicine ecosystem needs. There are caveats, of course. Stanford&#8217;s James Zou has pointed out that finding known biological concepts inside a model does not guarantee the model was actually using those concepts to make its predictions. Clinical validation requires larger trials across diverse populations and FDA approval. But the direction of travel is clear.</p><h2>The Business: Funding, Valuation, and Who Wrote the Checks</h2><p>The funding trajectory tells its own story. Seed round of $7M in August 2024, led by Lightspeed. Series A of $50M in April 2025, less than a year after founding, led by Menlo Ventures with Anthropic as a notable participant. Then Series B of $150M in February 2026, led by B Capital, with a $1.25B valuation. Total funding: $209M across three rounds.</p><p>The cap table is worth examining because of what it signals about market conviction. Anthropic, which is probably the most credible voice in AI safety and the company that literally pioneered constitutional AI, participated in the Series A. That is Dario Amodei&#8217;s shop putting money behind the belief that external interpretability research has commercial value. Eric Schmidt personally invested in the Series B. Salesforce Ventures came in on the B round as well, which suggests enterprise AI buyers see interpretability tooling as a procurement category they will eventually need. B Capital, which led the B round, has over $9B in AUM and focuses on technology and healthcare. The general partner who led the deal, Yanda Erlich, was formerly COO and CRO at Weights and Biases, which means he watched thousands of ML teams struggle with model behavior and presumably concluded that the interpretability layer was the missing piece.</p><p>The valuation jump from wherever it was at Series A to $1.25B at Series B is aggressive for a company with around 51 employees and what appears to be relatively early commercial traction. This is not a SaaS business with predictable recurring revenue (at least not yet). It is a research-first organization that is converting scientific breakthroughs into a platform while simultaneously pursuing fundamental research. The Series B press release explicitly says the funding will support green-field research into new interpretability methods alongside product development and partnership scaling. That is an unusual capital allocation mix for a company raising at unicorn valuations, and it suggests investors are pricing in the platform option value rather than near-term revenue.</p><h2>The Team Card</h2><p>For a 51-person company, the research bench is unusually deep. Tom McGrath founded interpretability at DeepMind. Nick Cammarata was a core contributor to the original interpretability team at OpenAI. Leon Bergen is a professor at UC San Diego who is on leave to work at Goodfire. The broader team includes researchers from Harvard, Stanford, and top ML engineering talent from OpenAI and Google. Mark Bissell and Myra Deng (Head of Product, formerly at Palantir working with health systems) have been doing the public technical evangelism on how the platform translates from research to production deployments.</p><p>The Palantir connection through Deng is actually interesting for health tech investors to note. Palantir has significant health system deployments, and Deng&#8217;s background in forward-deployed engineering at health systems means she has firsthand experience with the gap between what AI can do in a research setting and what it takes to deploy in clinical environments. That translational experience is exactly what you want on the product team of a company trying to move from research papers to production tools in healthcare.</p><h2>Where This Fits in the Health Tech Investment Landscape</h2><p>The angel investing question here is not really about whether to invest in Goodfire itself (it is way past the stage where most angel syndicates would participate, having already raised $209M). The question is about what Goodfire&#8217;s emergence means for the broader health tech AI stack and where the downstream investment opportunities are.</p><p>A few things jump out. First, interpretability as a category is becoming real. When Anthropic invests in your Series A and Eric Schmidt writes a personal check for your Series B, the market is telling you that &#8220;understanding what AI models actually do internally&#8221; is transitioning from academic curiosity to commercial necessity. For health tech investors, this means any portfolio company deploying foundation models in clinical or regulatory-sensitive contexts should be thinking about interpretability tooling as part of their technical architecture. The question to ask founders is not just &#8220;what model are you using&#8221; but &#8220;can you explain what the model learned and why it makes specific predictions.&#8221;</p><p>Second, the model-to-human knowledge transfer paradigm that Goodfire demonstrated with the Alzheimer&#8217;s biomarkers is potentially a massive unlock for biotech and diagnostics. The basic idea is that AI models trained on large biological datasets may have already learned things about disease biology that human researchers have not discovered yet. Interpretability provides the extraction mechanism. If this paradigm scales, we could see a wave of startups building on top of interpretability-enabled scientific discovery, using AI models as hypothesis generation engines and then feeding those hypotheses into traditional wet lab validation pipelines. That is a very different (and potentially much faster) drug discovery and diagnostics development cycle than what exists today.</p><p>Third, the regulatory angle matters more than most people appreciate. CMS has been tightening requirements around AI transparency in healthcare. The EU AI Act has explicit provisions for high-risk AI systems in healthcare. The FDA&#8217;s approach to AI/ML-based software as a medical device keeps evolving toward greater explainability requirements. A company that can provide interpretability-as-a-service for healthcare AI deployments is positioned to become critical infrastructure. Goodfire might do this directly, or (more likely) the techniques and tooling they develop will get embedded in the compliance and deployment stacks of health AI companies across the ecosystem.</p><p>Fourth, and this is more speculative, the convergence of interpretability with genomic foundation models could reshape how we think about precision medicine. If you can reverse engineer what a genomic model learned about variant pathogenicity and generate mechanistic explanations, you have a path toward AI-augmented genetic counseling at scale. The cost of sequencing keeps dropping. The bottleneck is interpretation. Interpretability applied to genomic AI models directly addresses that bottleneck. Health tech investors should be watching for startups that sit at this intersection.</p><h2>The Bull Case and the Bear Case</h2><p>The bull case is pretty straightforward. AI is eating healthcare. Regulatory and clinical requirements demand explainability. Goodfire is building the foundational science and tooling for AI explainability. They have the best team in the world for this specific problem, early proof points in life sciences, institutional partnerships with places like Mayo Clinic and Arc Institute, and enough capital to sustain a long research program. If interpretability becomes as essential to AI deployment as testing and monitoring are to software deployment (which seems likely), the market opportunity is enormous and Goodfire has a massive head start.</p><p>The bear case requires a bit more nuance. Research-first companies have historically struggled to convert scientific breakthroughs into sustainable commercial businesses. The gap between &#8220;we can do cool things with interpretability in a controlled research setting&#8221; and &#8220;here is a product that reliably improves model behavior across diverse production deployments with predictable unit economics&#8221; is real and has killed many promising startups. The $1.25B valuation prices in a lot of future execution. There is also the question of whether the scaling labs (OpenAI, Anthropic, Google) build sufficient interpretability tooling internally and make third-party solutions less necessary. Anthropic in particular has been doing serious interpretability research of its own, and the fact that they invested in Goodfire&#8217;s Series A could be read either as validation of external interpretability companies or as a hedge that keeps a potential competitor close.</p><p>There is also a timing question specific to healthcare. The regulatory requirements for AI explainability in clinical settings are clearly tightening, but the exact timeline and stringency of those requirements remain uncertain. If regulators move slowly, the commercial pull for interpretability tooling in healthcare could take longer to materialize than the bull case assumes. And the Stanford criticism from James Zou is worth taking seriously: finding biological concepts inside a model is different from proving the model used those concepts for its predictions. The validation requirements for clinical applications of interpretability-derived insights will be rigorous, and rightly so.</p><h2>So What</h2><p>For health tech angels and entrepreneurs, Goodfire represents something bigger than any single company. It represents the maturation of a new layer in the AI infrastructure stack that is particularly relevant to healthcare. The days of deploying black-box AI in clinical settings and hoping for the best are numbered, and the companies that figure out how to make AI transparent, steerable, and debuggable in healthcare contexts are going to capture enormous value.</p><p>The smartest thing an early-stage health tech investor can do right now is not necessarily try to chase Goodfire&#8217;s cap table (that ship has sailed for most angel syndicates). It is to understand the interpretability paradigm deeply enough to spot the startups that will build on top of it. Someone is going to build the interpretability-first diagnostics company. Someone is going to build the interpretability-enabled clinical decision support platform. Someone is going to build the regulatory compliance layer that uses interpretability techniques to satisfy FDA requirements for AI/ML-based medical devices. Those companies do not all exist yet, or they are early enough that angels can still get in.</p><p>Meanwhile, Goodfire keeps publishing research, signing partnerships with places like Mayo Clinic, and hiring researchers from the labs that built the foundation models everyone else is trying to deploy. Whether the $1.25B valuation proves prescient or premature will depend on execution, but the underlying bet, that understanding AI is as important as building AI, looks increasingly sound. Especially in a domain like healthcare where the consequences of not understanding what your model is doing can be measured in patient outcomes rather than just customer churn.&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fD4z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fD4z!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!fD4z!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!fD4z!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!fD4z!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fD4z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fD4z!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!fD4z!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!fD4z!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!fD4z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd476db80-1e5a-4b4f-be16-0e3bf39c3fb2_1200x600.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[NVIDIA Just Helped Map 31 Million Protein Complexes and the Health Tech Investment Implications Are Enormous]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/nvidia-just-helped-map-31-million</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/nvidia-just-helped-map-31-million</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Fri, 10 Apr 2026 12:58:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2_06!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>- NVIDIA, Google DeepMind, EMBL-EBI, and Seoul National University expanded the AlphaFold Protein Structure Database (AFDB) from monomeric protein structures to proteome-scale quaternary (complex) structures, predicting over 31 million homo- and heteromeric protein complexes across 4,777 proteomes</p><p>- 1.8 million high-confidence homodimer structures are now publicly available through AFDB, with the full 31M set coming for bulk download</p><p>- GPU-accelerated infrastructure running on H100 DGX Superpod clusters, using MMseqs2-GPU for multiple sequence alignment and TensorRT plus cuEquivariance for deep learning inference, enabled this scale of computation</p><p>- The work used STRING database physical interaction annotations to define biologically relevant heterodimer candidates, yielding ~8M heterodimer predictions with 57K tentatively high-confidence results</p><p>- Clustering of high-confidence complexes showed extreme concentration: the top 1% of structural representatives account for ~25% of all complexes, and ~9% of clusters are conserved across superkingdoms</p><p>- Downstream applications include drug target validation, variant interpretation at protein interfaces, generative protein design benchmarking, and systems-level structural biology</p><p>- This represents a foundational shift in the computational drug discovery stack with significant implications for health tech founders and investors evaluating companies in structural biology, protein engineering, and AI-driven therapeutics</p><h2>Table of Contents</h2><p>Why Protein Complexes Matter More Than Monomers</p><p>What Actually Got Built Here</p><p>The GPU Infrastructure Story</p><p>Confidence Calibration and the Heterodimer Problem</p><p>What the Clustering Reveals About Biology</p><p>The Drug Discovery and Health Tech Investment Angle</p><p>What This Means for Founders Building in This Space</p><p>Where This Goes Next</p><h2>Why Protein Complexes Matter More Than Monomers</h2><p>So AlphaFold2 was a massive deal. No dispute there. The Nobel Prize, the database of 200M+ predicted monomeric protein structures, the complete transformation of computational structural biology. But here is the thing that has been nagging at people in this space for years now: proteins almost never work alone. They form complexes. Dimers, trimers, big gnarly multi-subunit assemblies. The biological action happens at the interfaces between proteins, not just within the isolated 3D fold of a single chain. And for most of those interfaces, structural information has been basically nonexistent at any kind of useful scale.</p><p>The Protein Data Bank, which houses experimentally determined structures, covers a tiny fraction of known protein-protein interactions. For most organisms, the number of experimentally resolved multimeric structures is one to three orders of magnitude below what you would need to do serious systems biology or structure-based drug design against interaction surfaces. This is not a minor gap. This is the gap. If you are trying to interpret variants at protein interfaces, or validate drug targets that depend on complex formation, or benchmark generative protein design models, you have been operating with one hand tied behind your back.</p><p>What the NVIDIA-DeepMind-EMBL-EBI-Seoul National University collaboration just shipped is a direct assault on that bottleneck. Over 31 million predicted homo- and heteromeric protein complexes. 1.8 million of them classified as high-confidence and now surfaced through the AlphaFold Database at alphafold.com. For health tech investors and founders working anywhere near the structural biology stack, this is worth understanding in detail because it changes the ground truth assumptions underpinning a lot of computational approaches in drug discovery and protein engineering.</p><h2>What Actually Got Built Here</h2><p>The team predicted 23.4 million homodimers derived from 4,777 proteomes in UniProt, including 16 model organisms and 30 WHO global health proteomes. Then they added approximately 7.6 million heterodimer candidates extracted from the STRING database using physical protein-protein interaction annotations. That is a staggering combinatorial space. The heterodimer problem in particular is nasty because the number of possible pairwise interactions grows quadratically with proteome size. You cannot just do all-against-all predictions for large proteomes and expect to finish before the heat death of the universe.</p><p>Their approach to scoping the heterodimer set was pragmatic. They used STRING interaction evidence to filter down to physically interacting pairs, restricted to the same proteome (no inter-proteome complexes), and focused on dimers with a maximum combined sequence length of 3,000 amino acids. Critically, they did not filter by STRING score threshold for their initial computation, choosing coverage over precision. The literature suggests that filtering for STRING scores above 700 further reduces inputs while increasing prediction quality, but the team wanted maximum coverage for these priority proteomes and left that tighter filter as an option for downstream users.</p><p>For MSA generation, they used ColabFold&#8217;s search tool with the MMseqs2-GPU backend, keeping only the best hit per taxon based on alignment score. This is a clever orthology filter that prevents paralogous sequences from diluting the evolutionary signals that AlphaFold-Multimer needs to predict complex formation accurately. For heterodimers, they just concatenated the homodimer MSAs without pairing, which sounds lazy but actually held up well in their validation. They compared taxonomy-based pairing against simple concatenation and found that additional pairing did not clearly yield better predictions, especially at higher confidence thresholds.</p><p>Structure prediction ran through either ColabFold or an accelerated OpenFold implementation. Both used the same parameters: one set of weights from AlphaFold Multimer (model_1_multimer_v3), four recycles with early stopping, and no relaxation. The choice to skip relaxation saves compute without meaningfully hurting accuracy for the purposes of database-scale prediction. On a benchmark set of 125 X-ray resolved PDB homodimers released after AlphaFold2 was introduced (minimizing training data leakage), OpenFold accelerated with TensorRT and cuEquivariance matched ColabFold interface accuracy. The accelerated pipeline hit 75.4% usable predictions (DockQ above 0.3) compared to ColabFold at 73%, with mean DockQ scores of 0.647 versus 0.637. Not a massive difference, but the throughput gains from the accelerated stack are where the real story is.</p><h2>The GPU Infrastructure Story</h2><p>This is where things get genuinely interesting from an infrastructure perspective. The team ran on H100 DGX Superpod clusters and faced the classic HPC problem of maximizing GPU utilization across two workloads (MSA generation and structure prediction) that scale very differently.</p><p>For MSA generation with MMseqs2-GPU, the GPUs are only used during the ungapped filter stages. The subsequent alignment stages are multithreaded CPU processes. So you end up with a lot of GPU idle time if you just run one job at a time. Their solution was to stagger multiple colabfold_search processes per GPU, monitoring output to kick off the next one as soon as the GPU was free from the previous run. On a DGX H100 node, they found that three staggered processes could increase overall throughput by up to 25%, though individual chunks process more slowly due to CPU oversubscription. Not a perfect solution but a pragmatic one.</p><p>Chunk sizing matters here too. Smaller chunks mean more per-process overhead (database loading takes a couple minutes even on fast storage), while larger chunks take longer to finish and risk hitting SLURM wall time limits. For their setup with a 4-hour wall time limit, chunks of 300 sequences worked well. They also found that pre-staging databases on node-local SSDs helped throughput.</p><p>For structure prediction with ColabFold, they got higher throughput by packing homodimers of equal length into batches sorted by MSA depth in descending order. This reduces JAX recompilations, which is a surprisingly big deal for throughput at scale. This trick does not work for heterodimers where chain lengths differ, which is annoying. For OpenFold, the recompilation problem does not exist, but sequence length still drives execution time, so they reserved longer sequences for individual jobs and overlapped CPU-bound featurization of the next query with GPU-bound inference of the current one.</p><p>The broader SLURM orchestration story involved packing multiple predictions per node, matching GPU memory to sequence length, separating short versus long sequence queues, and monitoring GPU memory fragmentation. Asynchronous I/O helped avoid disk bottlenecks. None of this is glamorous work but it is the kind of systems engineering that determines whether a project like this takes three months or three years.</p><h2>Confidence Calibration and the Heterodimer Problem</h2><p>This section is arguably the most important for anyone who wants to actually use these structures, because confidence calibration is where the monomeric AlphaFold experience breaks down for complexes.</p><p>For monomers, pLDDT (predicted Local Distance Difference Test) gives you a pretty good sense of per-residue confidence. Above 70 is generally good, above 90 is great. But for complexes, the problem is fundamentally harder. You need to assess not just whether each chain is folded correctly but whether the interface between chains is plausible and positioned in the right pocket. That requires evaluating global and per-chain confidence metrics alongside local confidence metrics at the interface. Way more dimensions, way less training data to calibrate against.</p><p>The team evaluated four scoring metrics against a curated ground truth set of 1,968 PDB homodimers and 2,211 PDB monomers (as negative controls), all released after AlphaFold2&#8217;s training cutoff. They looked at ipTM (interface predicted TM-score), ipSAEmin (the minimum of the bidirectional interaction prediction Score from Aligned Errors), LISmin (Local Interaction Score), and pDockQ2. Of these, ipSAEmin showed the cleanest distributional separation between true homodimers and monomers, and the most stable F1 plateau across cutoffs.</p><p>They settled on a high-confidence threshold of ipSAEmin at or above 0.6, pLDDT average at or above 70, and backbone clashes at or below 10. This yielded precision of 0.859, recall of 0.655, and F1 of 0.744. Roughly 7% of homodimer predictions passed this filter, giving 1.8 million high-confidence homodimers. The AFDB website further categorizes these into &#8220;very high confidence&#8221; (ipSAEmin at or above 0.8, about 973K entries), &#8220;confident&#8221; (0.7 to 0.8, about 439K), and &#8220;low confidence&#8221; (0.6 to 0.7, about 343K).</p><p>Here is where it gets tricky. When they applied the same homodimer-derived thresholds to the 7.6 million heterodimer predictions, only about 57,000 passed. That is a tiny fraction, and the heterodimers that did pass showed a strong bias toward homodimer-like properties: smaller length differences between chains, higher inter-chain sequence identity. This is a real caveat. The current filtering criteria may be systematically excluding biologically real heterodimeric complexes that just happen to look less like homodimers. The team explicitly flags these 57K as &#8220;tentatively high-confidence&#8221; and says further calibration is needed before releasing a more representative heterodimer set.</p><p>For health tech investors, this matters because a lot of the most therapeutically interesting protein-protein interactions are heteromeric. Drug targets at heterodimer interfaces, signaling pathway complexes, antibody-antigen interactions. The homodimer expansion is valuable and immediately useful, but the heterodimer story is where the bigger drug discovery value lives, and it is not fully baked yet.</p><h2>What the Clustering Reveals About Biology</h2><p>The team clustered all 1.8 million high-confidence structures using Foldseek Multimercluster, which compressed the dataset roughly 8-fold into about 225,000 clusters. Of these, about 87,000 were non-singletons (had at least one other member). The distribution of cluster sizes is telling.</p><p>The top 1% of non-singleton cluster representatives cover approximately 25% of all entries, and the top 20% cover approximately 82%. This is a power law distribution that means predicted complex space is concentrated around a relatively small number of recurrent structural solutions. Nature keeps reusing the same interfaces. For protein engineering and generative design, this is useful information because it tells you where the structural density is and where genuinely novel folds might be hiding.</p><p>Clusters without any detectable PDB multimer match were more frequent among smaller clusters. The biggest clusters tend to overlap with known multimeric structures, which makes sense since the most common biological solutions are also the most experimentally characterized. The rare clusters, the ones with fewer members and no PDB match, are potentially the most interesting from a basic science perspective. These are predicted complex structures that nobody has crystallized or cryo-EM&#8217;d yet.</p><p>The taxonomic analysis is fascinating. About 9% of non-singleton clusters contain members from at least two different superkingdoms (bacteria, archaea, eukaryotes). These complexes likely originated in a common ancestor and have been maintained as universal building blocks of cellular life for billions of years. That is remarkable evolutionary conservation. Archaea and bacteria showed higher prediction success rates than eukaryotes, likely because prokaryotic proteins tend to be shorter, more compact, and richer in homo-oligomeric assemblies. Eukaryotic proteins are longer, more multi-domain, and more often participate in heteromeric complexes that are harder to predict.</p><h2>The Drug Discovery and Health Tech Investment Angle</h2><p>There are several concrete downstream applications that flow from having 1.8 million (and eventually 31M+) predicted complex structures publicly available. The first and most obvious is variant interpretation at protein interfaces. When you find a variant of uncertain significance through genomic sequencing, the question is always whether it affects protein function. If the variant sits at a protein-protein interface in a predicted complex, that is immediately informative in a way that monomeric structure alone cannot be. This matters for clinical genomics companies, rare disease diagnostic platforms, and anyone building tools for variant classification.</p><p>Drug target validation gets a boost too. Lots of drug targets depend on protein complex formation for their biological function. Having structural hypotheses for those complexes, even at moderate confidence, gives computational chemists and medicinal chemists a starting point for structure-based drug design at interfaces. Interface-directed drug design is harder than targeting a well-defined binding pocket on a monomer, but it is also where some of the most compelling therapeutic opportunities live, particularly in oncology and immunology.</p><p>Generative protein design benchmarking is another big one. Companies building protein design platforms (de novo binders, engineered enzymes, designed protein therapeutics) need benchmark datasets to validate their models. This dataset provides 1.8 million complex structures with calibrated confidence metrics. That is a serious training and benchmarking resource for anyone in the generative bio space.</p><p>Systems-level structural biology is the broader scientific play. Being able to overlay structural information onto interaction networks from resources like STRING creates a new kind of structural systems biology that was previously impossible at proteome scale. For health tech companies building knowledge graphs or multi-omic analysis platforms, this is another data layer to integrate.</p><p>The infrastructure itself is also investable. The fact that NVIDIA is shipping MMseqs2-GPU, cuEquivariance, and TensorRT as freely available libraries (Apache 2.0 licensing), and offering inference microservices through NIMs for MSA search and protein folding, means the barrier to running these kinds of analyses is dropping fast. A startup that would have needed six months and a million dollars in compute to run a large-scale complex prediction campaign can now potentially do it in weeks for much less. That changes the economics of computational structural biology startups.</p><h2>What This Means for Founders Building in This Space</h2><p>If you are founding or building a company anywhere in the structural biology or computational drug discovery stack, this release changes a few things worth thinking about.</p><p>First, the data moat argument for structural prediction companies just got weaker. If NVIDIA and DeepMind are going to keep expanding AFDB with complex structures at this pace, and the inference tools are freely available, then simply having predicted structures is not a defensible position. The value has to come from what you do with the structures: interpretation, design, integration into clinical or drug development workflows. The raw prediction layer is being commoditized in real time.</p><p>Second, the confidence calibration problem for heterodimers is an open research and commercial opportunity. The team explicitly acknowledged that their homodimer-derived thresholds do not work well for heterodimers. If someone builds better confidence metrics or better models for heteromeric complex prediction, that is genuinely differentiable right now. Companies like Protai (which NVIDIA has highlighted as using AlphaFold with proteomics and NVIDIA NIM for complex prediction in drug discovery) are already operating in this space.</p><p>Third, the integration opportunity is enormous. Most drug discovery and clinical genomics platforms have been built on monomeric structure assumptions. Retrofitting them to incorporate complex structure information, especially with calibrated confidence, is nontrivial engineering and science. There is real value in being the integration layer that makes complex structure predictions actionable for therapeutic development or clinical interpretation.</p><p>Fourth, compute economics continue to favor GPU-native approaches. The staggered MSA generation, the sequence packing tricks, the decoupled pipeline architecture described in this work represent significant systems engineering knowledge. Startups that understand how to run these workloads efficiently on modern GPU clusters will have meaningful cost advantages over those that treat compute as a black box.</p><p>The case studies in the paper are worth reading in full because they illustrate the kinds of biological insights that only emerge from complex prediction. There is a transcription elongation factor from Dictyostelium that has a completely fragmented, low-confidence monomeric prediction (pLDDT of 50.56) but forms a clean, high-confidence domain-swapped homodimer (pLDDT of 86.06). The fold literally does not exist without the partner chain. There is a membrane protein from a fungal pathogen where the monomeric prediction is mediocre but the dimeric model properly defines the membrane boundaries. There is a Mycoplasma transcriptional regulator where the monomer prediction is garbage (pLDDT of 56) but the dimer rescues it to high confidence (pLDDT of 85). These are not edge cases. For some proteins, monomeric prediction provides an incomplete or actively misleading structural picture. That has real implications for anyone relying on AlphaFold monomer predictions as the ground truth for their analysis.</p><h2>Where This Goes Next</h2><p>The team has been explicit that this is not the final state. The full 31M predictions (including the 21M+ homodimers below the high-confidence threshold and the 7.5M+ heterodimers) will be released for bulk download. Better heterodimer confidence calibration is coming. The prediction tools themselves continue to improve: OpenFold3, Boltz-2, and NVIDIA&#8217;s own Proteina are all advancing the frontier for complex structure prediction accuracy.</p><p>The convergence of GPU-accelerated inference, improved prediction models, and large-scale public databases is creating a new baseline for computational structural biology. For health tech investors, the question is no longer whether accurate protein complex structures will be widely available. They will. The question is who builds the most valuable applications on top of that infrastructure. Drug discovery platforms that can exploit interface-level structural information. Clinical genomics tools that interpret variants in the context of complex formation. Protein engineering companies that design novel interactions using these structures as templates. Biosecurity and pandemic preparedness applications that leverage pathogen-host interaction predictions from WHO priority proteomes.</p><p>The AlphaFold Database expansion from monomers to complexes is not just an incremental database update. It is a shift in what kind of structural biology is computationally accessible at population scale. For anyone investing in or building companies at the intersection of AI, structural biology, and therapeutics, ignoring this would be a mistake.&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2_06!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2_06!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2_06!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2_06!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2_06!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2_06!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg" width="1290" height="1824" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:1824,&quot;width&quot;:1290,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2_06!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2_06!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2_06!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2_06!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce31a1fd-b2b6-4692-bbfc-448cf25e522b_1290x1824.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[From Fringe to Formulary: How Integrative Medicine, Peptides, and the D2C Biomarker Stack Are Reshaping the Boundaries of Evidence-Based Care]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/from-fringe-to-formulary-how-integrative</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/from-fringe-to-formulary-how-integrative</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Thu, 09 Apr 2026 12:48:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Wr7p!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7280dcad-05ec-4956-97c3-9faecb031e7a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>This essay examines how historically marginalized medical approaches, including integrative medicine modalities, peptide therapeutics, precision nutrition, and direct-to-consumer biomarker-driven supplement protocols, are being selectively absorbed, validated, or rejected by mainstream U.S. healthcare infrastructure. The core thesis is that mainstreaming is not ideological adoption but a filtration process driven by measurement capability, regulatory tolerance, and reimbursement mechanics. Key areas covered include:</p><p>- Market demand for integrative approaches (~37% of U.S. adults, $30B+ annual spend) massively outpacing insurance coverage</p><p>- NIH/NCCIH institutional pivot toward Whole Person Health measurement infrastructure (~$170M annual budget) </p><p>- The VA Whole Health system as the largest real-world proof of concept for scaled integrative care delivery</p><p>- Clinical evidence stratification across modalities (chronic pain, oncology supportive care, mental health, sleep)</p><p>- Insurance and reimbursement dynamics including Medicare Advantage flexibility, opioid crisis catalysis, and structural CPT code gaps</p><p>- The peptide landscape split between FDA-approved therapeutics (GLP-1 agonists) and gray-zone compounded longevity peptides (BPC-157, Thymosin Alpha-1, CJC-1295/Ipamorelin) facing 503A/503B regulatory tightening</p><p>- The emerging &#8220;precision holistic medicine&#8221; stack integrating biomarkers, wearables, continuous glucose monitoring, and algorithmic supplement protocols</p><p>- mD2C lab and supplement ecosystems (Function Health, InsideTracker, Levels) operating outside traditional reimbursement and evidence thresholds</p><p>- A bifurcated future model: clinical medicine (evidence-based, reimbursed, slow), consumer precision health (fast, personalized, weakly evidenced), and an emerging hybrid layer where validated lifestyle and digital interventions integrate into standard care pathways</p><h2>Table of Contents</h2><p>Reframing Alternative Medicine</p><p>Market Demand vs System Resistance</p><p>NIH and the Infrastructure of Scientific Legitimacy</p><p>The VA Whole Health System as Scaled Proof of Concept</p><p>The Evidence Base: Where It Works and Where It Does Not</p><p>Insurance, Policy, and the Reimbursement Question</p><p>Peptides in the System</p><p>The Precision Holistic Medicine Stack</p><p>D2C Lab Testing and the Supplement Economy</p><p>Where These Worlds Collide</p><p>Critiques, Limitations, and Contrarian Takes</p><p>The Bifurcated Future of Healthcare</p><p>Conclusion</p><h2>Reframing Alternative Medicine</h2><p>There is a version of this essay that opens with some breathless claim about how ancient wisdom is finally being recognized by modern medicine. That is not this essay. The actual story is way less romantic and way more interesting. What is happening to integrative, holistic, and historically fringe medical approaches in the U.S. is not a cultural awakening. It is a decomposition process. Specific practices are being pulled apart, isolated into testable components, run through clinical trial infrastructure, and either validated narrowly or discarded. The stuff that survives gets a CPT code, a reimbursement pathway, and a slot in a clinical guideline. The stuff that does not keeps living in the cash-pay economy, where roughly 37 percent of American adults are already spending north of 30 billion dollars a year, mostly out of pocket, on complementary and integrative health approaches. That is not a fringe market. That is a parallel health economy operating at scale with almost zero insurance penetration. The question for investors, operators, and policymakers is not whether integrative medicine matters. Patients have already voted with their wallets. The question is which components cross the evidentiary and economic threshold into reimbursable clinical care, which stay in the consumer health lane, and which collapse entirely under regulatory scrutiny. That sorting process is the actual story.</p><h2>Market Demand vs System Resistance</h2><p>The demand side of this equation is not ambiguous. National Health Interview Survey data consistently shows roughly one in three U.S. adults using some form of complementary or integrative health approach. Chronic pain, which affects about 20 percent of the adult population, is the single largest demand driver, but mental health, sleep, and metabolic concerns are not far behind. Annual out-of-pocket spending on these approaches has been estimated at over 30 billion dollars, which is a genuinely remarkable number when you consider that almost none of it is reimbursed by commercial insurance or traditional Medicare. What that spending pattern tells you is that patients are not waiting for the healthcare system to catch up. They are building workarounds. They are paying cash for acupuncture, functional medicine consultations, supplement stacks, peptide protocols, and precision nutrition programs because the traditional system either does not offer these services or actively gatekeeps them behind referral chains and prior authorization barriers that make access impractical. The system resistance is structural, not ideological. Fee-for-service payment models reward discrete procedures and pharmacologic interventions. They do not reward longitudinal lifestyle coaching, behavioral modification programs, or multidisciplinary care coordination, which is exactly what most integrative approaches require. There is a deep misalignment between how these interventions are delivered and how the payment system is architected, and that misalignment explains far more of the adoption gap than any debate about scientific legitimacy.</p><h2>NIH and the Infrastructure of Scientific Legitimacy</h2><p>If the demand side is clear, the legitimacy side has been the bottleneck for decades. And the single most important institutional actor on that front is the National Center for Complementary and Integrative Health, which operates within NIH on a budget of roughly 170 million dollars annually. That is not a huge number relative to NIH total spend, but the strategic direction matters more than the dollar figure. NCCIH has undergone a meaningful pivot in recent years, moving away from studying individual alternative modalities in isolation and toward what it calls Whole Person Health. This is not just branding. It represents a shift toward systems biology frameworks, non-pharmacologic intervention research, and behavioral integration studies that treat the patient as a complex adaptive system rather than a collection of organ-specific disease states. The most consequential output of this pivot may be the Whole Person Health Index, a nine-domain validated measurement tool covering sleep, stress, diet, social connectedness, physical activity, and related domains. It is designed for both clinical and research integration and is slated for deployment in national surveys. This matters enormously because the historical barrier to integrative medicine legitimacy has always been measurement. You cannot bill for what you cannot measure. You cannot build clinical guidelines around what you cannot quantify. You cannot run a value-based contract on outcomes you have no validated instrument to assess. NCCIH is essentially building the measurement infrastructure that the field has lacked, and if the Whole Person Health Index achieves broad adoption in clinical research, it creates the evidentiary scaffolding that downstream reimbursement decisions require. The phrase that captures this dynamic precisely: if you cannot measure it, you cannot reimburse it. NIH is trying to solve the measurement problem.</p><h2>The VA Whole Health System as Scaled Proof of Concept</h2><p>The single most important real-world case study for integrative medicine at scale is not an academic medical center pilot program or a Silicon Valley health startup. It is the Veterans Health Administration. The VA Whole Health system represents the largest implementation of integrative care delivery in the United States, and it happened not because the VA suddenly became philosophically aligned with holistic medicine but because of a policy mandate. The Comprehensive Addiction and Recovery Act, known as CARA, created a legislative requirement for non-opioid pain management alternatives within the VA system. That mandate, driven by the opioid crisis, opened the door for acupuncture, yoga, meditation, nutrition counseling, and health coaching to be embedded directly into VA care pathways. The VA did not bolt these services on as optional wellness add-ons. It redesigned care delivery around personalized health plans that integrate these modalities alongside conventional treatment. Chiropractors, acupuncturists, massage therapists, and health coaches now operate as members of clinical teams within the system. Reported outcomes include improved patient-reported quality of life metrics, reduced opioid utilization, and increased patient engagement and satisfaction scores. The evidence is largely observational and self-reported at this point, which matters for the purists, but the operational proof of concept is significant. The VA demonstrated that integrative care can scale within a large, bureaucratic, risk-averse delivery system when there is a policy mandate, a structured delivery model, and institutional commitment. The key insight for investors and operators is that the VA model worked because it embedded integrative services into existing care infrastructure rather than creating a parallel track. Integrative medicine succeeds when it becomes invisible, woven into standard care pathways rather than marketed as an alternative. That is the design principle that separates scalable models from boutique experiments.</p><h2>The Evidence Base: Where It Works and Where It Does Not</h2><p>Honesty about the evidence base is critical here because the temptation to overstate clinical support for integrative modalities is real and counterproductive. The strongest evidence clusters around four domains: chronic pain, oncology supportive care, mental health conditions including anxiety, depression, and PTSD, and sleep disorders. Within those domains, the modalities with the most robust trial data include acupuncture for pain and insomnia and anxiety, mindfulness-based interventions for stress and depression and PTSD, and tai chi and yoga for musculoskeletal and functional outcomes. But robust is a relative term. Many of the studies in this space are small, heterogeneous in design, and vulnerable to placebo and expectancy effects. Blinding is often impractical or impossible. Effect sizes tend to be modest. And reproducibility across populations and settings is inconsistent. None of that is disqualifying, but it does mean that the evidence case for most integrative modalities is not based on superiority over conventional treatment. It is based on a different value proposition entirely: favorable safety profiles, additive effects when combined with standard care, and potential system-level cost offsets through reduced utilization of high-cost interventions like opioids, emergency department visits, and surgical procedures. The honest framing is that these therapies win on risk-adjusted marginal benefit, not on standalone efficacy. That framing is less exciting than claiming acupuncture cures everything, but it is far more useful for clinical decision-making and reimbursement design. It also maps cleanly onto value-based care incentives, where the goal is total cost of care reduction and patient satisfaction improvement, not necessarily head-to-head superiority on any single clinical endpoint.</p><h2>Insurance, Policy, and the Reimbursement Question</h2><p>Reimbursement is where idealism meets spreadsheets, and the dynamics here are worth understanding in detail because they determine what actually scales. The dominant strategy among payors who engage with integrative medicine at all is narrow coverage. That means reimbursement tied to specific clinical indications, most commonly chronic low back pain, delivered via standardized protocols with measurable endpoints. Payors fund procedures, not philosophies, and the distinction matters. A commercial insurer will cover acupuncture for chronic pain documented by ICD-10 codes and delivered in a defined number of sessions. That same insurer will not cover a holistic wellness program that claims to improve vitality and balance energy flow. The specificity requirement is the filter, and it is the reason why broad integrative medicine adoption through insurance channels has been so slow despite massive consumer demand. The opioid crisis was the single most important policy catalyst for integrative reimbursement expansion. When opioid prescribing became a public health emergency, regulators and payors were forced to search for non-pharmacologic pain management alternatives. That search opened doors for acupuncture, physical therapy and behavioral therapy hybrids, and mindfulness-based pain programs that might otherwise have remained uncovered for another decade. Integrative medicine did not create its own reimbursement pathway. It piggybacked on a public health crisis that made the status quo politically and clinically untenable. Medicare Advantage has emerged as the most interesting sandbox for integrative benefit experimentation. MA plans have supplemental benefit flexibility that traditional Medicare lacks, and some plans now include fitness programs, wellness coaching, and in certain cases acupuncture coverage as supplemental offerings. The strategic logic is member acquisition and retention in a competitive enrollment market, not ideological commitment to holistic care, but the effect is the same: real-world testing of integrative benefits within a managed care framework. The structural barriers remain significant though. Many integrative services lack dedicated CPT codes. Credentialing and training standards vary dramatically across states and modalities. Delivery models for longitudinal lifestyle interventions do not fit neatly into fee-for-service visit structures. And the fundamental misalignment between episodic billing and continuous care models means that even when there is clinical evidence and patient demand, the payment plumbing often cannot accommodate the service.</p><h2>Peptides in the System</h2><p>Now we get to the part of the essay that is going to make the compliance departments nervous. Peptide therapeutics represent one of the most interesting and most confused categories in the broader integrative and precision medicine landscape. The confusion stems from the fact that peptides exist on a spectrum that ranges from fully FDA-approved blockbuster drugs to gray-market compounded substances prescribed via telehealth clinics with minimal oversight, and the public conversation tends to lump everything together. On the legitimate pharmaceutical end, you have GLP-1 receptor agonists like semaglutide, which have transformed metabolic medicine and represent one of the largest drug categories by revenue globally. You have insulin analogs. You have growth hormone-related peptides approved for specific clinical indications. These are rigorously tested, fully reimbursable, and integrated into standard care pathways. Nobody debates whether semaglutide is real medicine. Then there is the gray zone, and it is large and growing. Peptides like BPC-157, Thymosin Alpha-1, TB-500, and CJC-1295 combined with Ipamorelin are prescribed through longevity clinics and compounding pharmacies for indications including tissue recovery, inflammation reduction, sleep optimization, and general anti-aging. The mechanistic rationale for many of these peptides is genuinely interesting. They tend to have high receptor specificity, lower systemic toxicity profiles compared to broad-spectrum pharmaceuticals, and they mimic endogenous signaling pathways, which is an elegant therapeutic approach in theory. The problem is that theory and evidence are not the same thing. Most of these peptides lack large-scale randomized controlled trials. Long-term safety data is sparse to nonexistent. Dosing protocols are not standardized. Manufacturing quality varies dramatically across compounding sources. And the intellectual property incentives that drive pharmaceutical development are largely absent, which means nobody with deep pockets has a strong economic reason to fund the expensive trials that would resolve the evidence questions. The regulatory environment is tightening. The FDA has been increasing scrutiny of compounding pharmacies under the 503A and 503B frameworks, and several peptides have been removed from allowable bulk compounding lists. This trend is likely to accelerate. The strategic read for investors is that most gray-zone peptides face a binary future: either they get absorbed into formal pharmaceutical development pipelines with proper trials and regulatory approval, or they get systematically pushed out of semi-legal clinical use as compounding regulations tighten. The middle ground of widespread off-label compounded prescribing is probably not sustainable over a five to ten year horizon. Peptides as a category represent the closest bridge between biohacker medicine and pharmaceutical-grade therapeutics, but crossing that bridge requires capital, time, and regulatory patience that most of the current ecosystem does not have.</p><h2>The Precision Holistic Medicine Stack</h2><p>There is a new category emerging that does not have a clean name yet, so for purposes of this analysis call it the precision holistic medicine stack. It represents the convergence of biomarker data, lifestyle inputs, environmental signals, and behavioral tracking into individualized intervention protocols. It borrows infrastructure from precision medicine, philosophy from integrative and functional medicine, and go-to-market strategy from consumer technology. The architecture has three layers. The data acquisition layer includes blood biomarkers, hormone panels, microbiome sequencing, continuous glucose monitoring, and wearable-derived metrics like heart rate variability, sleep staging, and activity patterns. The interpretation layer involves algorithmic scoring systems that translate raw biomarker data into actionable recommendations, often using what practitioners call functional or optimal ranges rather than the disease-threshold reference ranges used in conventional clinical medicine. The intervention layer includes supplements, peptides, dietary modifications, exercise protocols, sleep hygiene programs, and behavioral change frameworks. The epistemological shift embedded in this stack is significant and worth flagging explicitly. Clinical medicine defines normal by the absence of disease. The precision holistic stack defines optimal by deviation from a theoretically ideal biomarker profile. These are fundamentally different frameworks, and the gap between them creates both opportunity and risk. The opportunity is earlier intervention, catching metabolic dysfunction or hormonal imbalance or micronutrient deficiency before it progresses to diagnosable disease. The risk is overtesting, overinterpretation, and overtreatment of normal biological variation that has no clinical significance. Both things can be true simultaneously, and the honest assessment is that the current evidence base does not clearly resolve which is dominant.</p><h2>D2C Lab Testing and the Supplement Economy</h2><p>The consumer-facing infrastructure for this precision holistic stack is already at scale. Companies like Function Health, InsideTracker, and Levels have built direct-to-consumer lab testing platforms that bypass primary care gatekeeping entirely. The model works like this: consumer orders an expanded biomarker panel online, blood draw happens at a partner lab or via mobile phlebotomy, results are analyzed using proprietary algorithms, and personalized recommendations are generated, often including supplement protocols available for purchase through the platform. The business model incentives here are worth understanding clearly because they shape what gets recommended. These companies generate significant revenue from supplement sales, which are high-margin recurring products. That creates an inherent tension between the clinical interpretation function and the commercial imperative to convert biomarker results into product purchases. This is not necessarily nefarious but it is structurally important because the same dynamic does not exist in traditional clinical medicine, where the interpreting physician typically does not have a financial interest in the specific supplement the patient buys. The expanded biomarker panels used by these platforms go well beyond standard clinical panels. They include micronutrient levels, inflammatory markers like high-sensitivity CRP and homocysteine, advanced lipid subfractions, hormone panels, and metabolic intermediates that most primary care physicians would not order in a routine visit. Whether that additional data is clinically useful for asymptomatic individuals is genuinely debated. The causal chain from biomarker deviation to supplement intervention to measurable health outcome has not been validated by randomized controlled trials for most of these protocols. The regulatory environment is permissive because supplements are regulated under the Dietary Supplement Health and Education Act of 1994, which imposes minimal pre-market efficacy requirements. Companies can sell personalized supplement stacks based on biomarker data without demonstrating that those stacks actually improve outcomes. The result is an ecosystem that is scaling rapidly precisely because it is unconstrained by the evidence thresholds and reimbursement requirements that govern clinical medicine. That is both its advantage and its vulnerability.</p><h2>Where These Worlds Collide</h2><p>The interesting analytical question is where the mainstream clinical system and this emerging consumer precision health ecosystem actually converge, and where they remain fundamentally separate. The convergence points are real. Precision nutrition is increasingly backed by NIH-funded research and is moving toward clinical integration. Continuous glucose monitoring, originally a diabetes management tool, is crossing over into metabolic health optimization for non-diabetic populations and is entering clinical workflows in endocrinology and primary care. Behavioral interventions for sleep, stress, and exercise are evidence-supported and increasingly embedded in value-based care models. Wearable-derived data is beginning to inform clinical decision-making, particularly in cardiology and sleep medicine. These are legitimate areas of overlap where consumer health innovation and clinical medicine are meeting on shared evidentiary ground. The non-convergence zones are equally real. Peptide longevity clinics operating on compounded gray-market substances are not converging with hospital-based care. Supplement mega-stacks prescribed based on expanded biomarker panels without RCT validation are not being adopted by academic medical centers. Functional diagnostic frameworks that use proprietary reference ranges without published validation studies are not gaining traction in evidence-based clinical guideline development. The pattern is clear: the mainstream system absorbs components that can be measured, standardized, and reimbursed. Everything else stays in the consumer health economy, operating on cash-pay models with rapid iteration cycles and weak evidence floors. Hospitals are evidence-bound, liability-constrained, and reimbursement-driven. The D2C ecosystem is consumer-driven, narrative-driven, and optimized for speed of product iteration. These are fundamentally different operating systems, and the expectation that they will fully merge is probably wrong.</p><h2>Critiques, Limitations, and Contrarian Takes</h2><p>Any honest analysis of this space has to grapple with the contrarian case, and there are several strong versions of it. First, the evidence quality problem is real and not easily dismissed. Many studies supporting integrative modalities are small, poorly controlled, and vulnerable to bias. The standard counterargument, that real-world evidence and safety profiles should matter alongside RCTs, has merit but also has limits. Real-world evidence is subject to confounding, selection bias, and placebo effects that are particularly pronounced in interventions where patient expectation plays a large role. Being low-risk does not make something effective, and scaling low-efficacy interventions system-wide has opportunity costs. Second, the concern about medicalization of wellness is legitimate from both directions. Critics worry that absorbing wellness practices into clinical medicine dilutes scientific rigor and creates reimbursement claims for interventions with marginal benefit. Proponents argue that the biomedical model is too narrow and that expanding the care model to include behavioral and lifestyle interventions produces better population health outcomes. Both sides have reasonable points, and the resolution probably looks less like one side winning and more like continuous negotiation at the boundary. Third, cost effectiveness remains an open question for most integrative interventions. The theoretical case for total cost of care reduction through upstream prevention and reduced utilization of high-cost acute services is intuitive but not yet proven at scale for most modalities. The VA Whole Health data is suggestive but not definitive, and extrapolating VA results to commercial populations requires caution given the unique characteristics of the veteran population. Fourth, and this is the spiciest take, most of the precision holistic medicine stack may be signal-poor. High noise in biomarker interpretation, weak causal linkage between supplement interventions and outcomes, and the fundamental challenge of optimizing complex biological systems based on snapshot blood panels all suggest that the current generation of D2C precision health tools may be overselling what they can deliver. Supplements are not the same as interventions in the clinical sense, and treating them as equivalent is a category error that the market has not yet corrected. Fifth, the peptide boom may partially collapse under regulatory pressure. As the FDA continues tightening compounding pharmacy oversight, many of the peptides currently available through longevity clinics may become inaccessible outside formal pharmaceutical channels, and the formal pharmaceutical pathway requires capital and time that most of these molecules will never attract. The contrarian synthesis is that mainstream medicine may be right to resist most of this, and the components that do cross over will do so only after clearing evidence bars that the current consumer health ecosystem largely ignores.</p><h2>The Bifurcated Future of Healthcare</h2><p>The most likely forward state over the next five to ten years is not convergence into a single unified system. It is a bifurcated or possibly trifurcated model. The first tier is clinical medicine as it exists today, evidence-based, reimbursed, integrated into delivery systems, and slow to adopt new modalities. This tier will continue to selectively absorb integrative components that clear evidentiary and economic thresholds, particularly in chronic pain management, oncology supportive care, mental health, and metabolic disease. Reimbursement evolution in this tier will likely follow bundled payment models, longitudinal care arrangements, and outcomes-based contracts that create financial incentives for non-pharmacologic interventions. The second tier is the consumer precision health economy, operating on cash-pay models with rapid product iteration, personalized protocols, and weak evidence requirements. This tier includes D2C lab testing, supplement protocols, gray-zone peptide clinics, and wellness-adjacent services that do not need or seek reimbursement. It will continue to grow as long as consumer demand and disposable income support it, but it is vulnerable to regulatory action on the peptide side and to reputational damage if high-profile adverse events occur. The third tier, and the most interesting one for investors, is the emerging hybrid layer where validated lifestyle, behavioral, and digital interventions integrate into clinical care pathways. This includes precision nutrition informed by continuous monitoring, digital therapeutics with FDA clearance, AI-guided behavioral interventions, wearable-derived clinical data, and structured health coaching embedded in value-based care models. This hybrid layer is where the actual transformation happens because it combines the rigor and reimbursability of clinical medicine with the personalization and patient engagement advantages of the consumer health stack. The expansion areas in this hybrid tier over the next decade will likely center on chronic pain management using multimodal non-pharmacologic protocols, oncology supportive care integrating mind-body interventions, mental health treatment combining digital therapeutics with behavioral coaching, and metabolic disease management using continuous monitoring and precision nutrition. The data layer underneath all of this is going to explode. Continuous monitoring, behavioral phenotyping from passive digital data, and AI-guided intervention selection are going to create a density of patient-level data that current clinical systems are not architected to handle. The delivery systems that figure out how to ingest, interpret, and act on that data within value-based care frameworks will have significant competitive advantages.</p><h2>Conclusion</h2><p>The future of this space is not alternative versus conventional. That framing was always a false binary, and it is becoming increasingly irrelevant as the sorting process accelerates. What is actually happening is a filtration process where biology, data, and economics determine what survives. The healthcare system is not absorbing holistic medicine wholesale. It is selectively ingesting components that can be measured, standardized, and reimbursed, while the rest forms a rapidly growing parallel health economy that operates on entirely different rules. For investors, the actionable insight is to focus on the hybrid layer, the companies and models building validated, measurable, reimbursable interventions that borrow from the integrative and precision health toolkit but meet clinical evidence and regulatory standards. That is where venture-scale outcomes live. The consumer health tier will produce some big consumer brands but carries regulatory and evidence risk that makes underwriting harder. The pure clinical tier will continue absorbing modalities one narrow indication at a time, slowly but with durable reimbursement once adopted. The companies that will matter most are the ones building the connective tissue between these tiers: the measurement tools, the data infrastructure, the care coordination platforms, and the evidence generation engines that translate promising approaches into clinically integrated, economically viable care delivery. That is not a wellness story. It is a healthcare infrastructure story, and it is one of the more interesting ones playing out right now.&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;</p><p></p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tWBm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tWBm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 424w, https://substackcdn.com/image/fetch/$s_!tWBm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 848w, https://substackcdn.com/image/fetch/$s_!tWBm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!tWBm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tWBm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg" width="360" height="180" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:180,&quot;width&quot;:360,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tWBm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 424w, https://substackcdn.com/image/fetch/$s_!tWBm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 848w, https://substackcdn.com/image/fetch/$s_!tWBm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!tWBm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6a8c832-1ca5-4f45-8f80-eee98a825eab_360x180.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[Clinical Trials Are the New Bottleneck: AI Drug Discovery Has Created an Evidence Infrastructure Crisis]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/clinical-trials-are-the-new-bottleneck</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/clinical-trials-are-the-new-bottleneck</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Wed, 01 Apr 2026 12:25:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!78v_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d31b8ce-c49f-430a-9b46-82e4071bd6fc_434x585.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>The core argument: AI has dramatically compressed preclinical drug discovery, but clinical development timelines remain stuck. The bottleneck has shifted from molecule identification to evidence generation. The next durable health-tech companies won&#8217;t discover drugs. They&#8217;ll prove they work.</p><h3>Key claims:</h3><p>- AI-driven structure-based and generative methods have increased preclinical throughput substantially, pushing more candidates into already-strained development pipelines</p><p>- Long-run clinical success rates only recently began recovering after decades of decline, per 2025 Nature Communications data &#8211; meaning the industry hasn&#8217;t solved translational efficiency at scale</p><p>- FDA&#8217;s 2025 draft guidance on externally controlled trials effectively outlines a technical stack (phenotype normalization, covariate harmonization, temporal alignment, endpoint ontology mapping) that nobody has fully built yet</p><p>- A 2025 Nature Medicine TrialTranslator study showed real-world oncology survival is often ~6 months worse than RCT results, and ~1 in 5 real-world patients don&#8217;t qualify for phase 3 trials</p><p>- 2025 FedECA paper in Nature Communications introduces federated external control arms for distributed settings &#8211; a direct blueprint for privacy-preserving comparator networks</p><p>- TrialGPT (Nature Communications, 2024) and successor systems suggest patient-trial matching is solvable, but recruitment alone is too narrow a moat without a broader trial-state architecture</p><p>- Five infrastructure layers where durable category-defining companies will likely form: comparator infrastructure, phenotype infrastructure, continuous measurement infrastructure, model assurance infrastructure, and adaptive protocol infrastructure</p><h2>Table of Contents</h2><p>The paradox nobody talks about out loud</p><p>Why this bottleneck exists now</p><p>The technical stack academia is quietly building</p><p>Regulators are opening the door but raising the bar</p><p>What this means for founders, CROs, and venture underwriting</p><p>The five infrastructure layers that matter</p><h2>The paradox nobody talks about out loud</h2><p>Here&#8217;s a thing that should be more disorienting than it is: the AI-in-biopharma narrative has largely convinced the industry, investors, and the trade press that the hard part of drug development is finding good molecules. Structure-based design, generative chemistry, AlphaFold derivatives, target ID from genomic embeddings &#8211; all of it has gotten genuinely impressive. Not vaporware impressive. Actually impressive. Some of these tools are running live in discovery programs at major sponsors right now.</p><p>The problem is that making the front end of discovery faster is a little like widening the on-ramp to a highway that&#8217;s already gridlocked. You don&#8217;t actually get more cars to their destination any faster. You just fill up the backup.</p><p>Clinical development is the backup. It takes, on average, somewhere between six and ten years from first-in-human to approval, and the bottleneck during most of that period is not computational. It&#8217;s not even primarily biological in the narrow sense. It&#8217;s evidentiary. The industry is slow at assembling the kind of regulatory-grade, causally defensible, generalizable evidence packages that the FDA and its international equivalents actually need to say yes with confidence. And as AI accelerates the front end, that evidentiary bottleneck becomes more acute, not less.</p><p>This is the paradox that people in clinical development talk about quietly but that rarely makes it into the funding narratives for health-tech companies. Everyone wants to fund the &#8220;AI for drug discovery&#8221; story because it maps to a familiar venture pattern: scientific insight, clever model, platform play, licensing revenue or acquisition. The evidence generation story is harder to pitch because the product is less romantic. It involves phenotype normalization, external control cohort assembly, federated data governance, digital twin validation frameworks, and adaptive master protocol software. None of that fits on a TED slide.</p><blockquote><p>But that&#8217;s exactly where the alpha is.</p></blockquote><h2>Why this bottleneck exists now</h2><p>To understand why this is happening now specifically, it helps to separate two different timelines that are running at very different speeds.</p><p>The first timeline is discovery throughput. Over the past five or six years, the combination of structure prediction, generative molecular design, and large-scale biological embedding has genuinely changed the rate at which credible drug candidates can be identified. The number of AI-discovered compounds entering clinical trials &#8211; while still small relative to the total industry pipeline &#8211; is growing. More importantly, the capital and talent flowing into &#8220;AI-native&#8221; biotech is enormous, which means the pipeline of candidates heading toward IND filings is expanding.</p><p>The second timeline is development infrastructure. This one has barely moved. The FDA&#8217;s average review timeline hasn&#8217;t compressed dramatically. Phase 2 and phase 3 success rates, while having shown some recent improvement after decades of decline per 2025 Nature Communications analysis, are still deeply sobering &#8211; roughly 10 to 15 percent of candidates that enter phase 1 ultimately reach approval depending on the therapeutic area. Enrollment velocity is still plagued by the same problems it was twenty years ago: sites are overextended, patients are hard to identify, eligibility criteria are often written for clean populations that don&#8217;t exist in the wild, and sponsors regularly discover late in development that their trial population doesn&#8217;t look much like the patients who will actually use the drug if it&#8217;s approved.</p><p>The structural cause of this divergence is that discovery innovation has been driven by computational methods that are relatively cheap to deploy and iterate, while development innovation requires touching the actual trial infrastructure: sites, IRBs, patient populations, regulatory submissions, comparator data sets, endpoint definitions. That stuff is slow, bureaucratic, and deeply institutional. It&#8217;s not the kind of thing you can iterate on quickly with gradient descent.</p><p>What makes this moment distinct is that multiple academic and regulatory threads are now converging on a shared diagnosis, and regulators themselves are beginning to acknowledge that the evidence generation stack needs to be rebuilt rather than patched. That combination &#8211; a growing pipeline of AI-discovered candidates, a strained development infrastructure, and a regulatory environment signaling conditional openness to new methodologies &#8211; is the setup for a very large business opportunity.</p><h2>The technical stack academia is quietly building</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/clinical-trials-are-the-new-bottleneck">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[NVIDIA’s Healthcare Stack Is the Picks and Shovels Play You’ve Been Waiting For]]></title><description><![CDATA[Table of Contents]]></description><link>https://www.onhealthcare.tech/p/nvidias-healthcare-stack-is-the-picks</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/nvidias-healthcare-stack-is-the-picks</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Tue, 31 Mar 2026 12:11:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!toV1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Table of Contents</h2><p>Section 1: The Inflection Point Is Already Here</p><p>Section 2: BioNeMo and the Drug Discovery Revolution</p><p>Section 3: MONAI and the Medical Imaging Flywheel</p><p>Section 4: Isaac for Healthcare and the Robotics Buildout</p><p>Section 5: Holoscan and the Edge Intelligence Layer</p><p>Section 6: Parabricks and the Genomics Data Deluge</p><p>Section 7: Clara, NIM, and the Open Source Bet</p><p>Section 8: What This Means for Investors and Founders</p><h2>Abstract</h2><p>This essay examines NVIDIA&#8217;s healthcare and life sciences platform stack as an investment and competitive landscape thesis for health tech entrepreneurs and angel investors. Drawing on NVIDIA&#8217;s 2026 State of AI in Healthcare and Life Sciences survey (600+ respondents, fielded Aug-Sept 2025) and the company&#8217;s product ecosystem documentation, the piece argues that NVIDIA has quietly built the most comprehensive AI infrastructure layer in healthcare, and that understanding each component, BioNeMo, MONAI, Isaac for Healthcare, Holoscan, Parabricks, Clara, and NIM, is now table stakes for anyone deploying capital or building companies in the space.</p><h3>Key data points from the survey:</h3><p>- 70% of healthcare/life sciences orgs actively using AI, up from 63% in 2025</p><p>- 69% using generative AI/LLMs, up from 54%</p><p>- 85% of orgs increasing AI budgets in 2026</p><p>- 85% of management-level respondents report AI increased annual revenue</p><p>- 44% of management say AI boosted annual revenue by more than 10%</p><p>- 57% of medtech orgs report ROI from medical imaging AI</p><p>- 46% of pharma/biotech report ROI from drug discovery AI</p><p>- 47% either actively using or assessing agentic AI</p><p>- 82% say open-source models are moderately to extremely important to their AI strategy</p><p>- Hybrid computing for AI workloads rose from 35% to 43% year over year</p><h2>Section 1: The Inflection Point Is Already Here</h2><p>The 2026 NVIDIA State of AI in Healthcare and Life Sciences survey is the kind of data that should make any health tech investor put down whatever deck they&#8217;re reading and pay attention. Seven out of ten healthcare and life sciences organizations are now actively using AI. That&#8217;s not piloting, not exploring, not forming a committee to assess strategic readiness, actually using it. That number was 63% a year prior. The jump isn&#8217;t noise. It&#8217;s an industry crossing a threshold.</p><p>What&#8217;s more interesting than the headline number is where the growth came from. The payers and providers segment, which historically moves about as fast as a fax machine, jumped 13 percentage points year over year, from 43% to 56% active AI usage. Hospitals and insurance companies are now majority AI users by this measure. That&#8217;s not a small thing. Payers and providers represent the largest slice of U.S. healthcare spending by a wide margin, and they&#8217;ve been the laggard cohort in every digital health wave since the EMR rollouts of the early 2010s. When that segment starts moving, the infrastructure underneath it matters a lot.</p><p>The survey also surfaced something that more or less confirms what anyone building in health tech has been observing in the field: generative AI and LLMs blew past predictive analytics as the top AI workload category. Sixty-nine percent of respondents cited gen AI as their primary focus, up from 54% the prior year. Data analytics and data science came in at 65%, predictive analytics at 51%, and agentic AI, newly tracked this year, debuted at 47%. That agentic number is going to be important. More on that later.</p><p>The revenue story is the part that should accelerate capital deployment. Eighty-five percent of management-level respondents said AI increased their annual revenue. Eighty percent said it reduced annual costs. Forty-four percent said the revenue increase exceeded 10% annually. For small companies specifically, 56% reported more than 10% revenue lift from AI. These are not marginal improvements on the margin of a spreadsheet. At the portfolio company level, these are the kinds of numbers that change valuations, extend runways, and make the next fundraise significantly easier.</p><p>Budget intentions for 2026 are equally unambiguous. Eighty-five percent of respondents said their AI budgets will increase. Nearly half said budgets will grow more than 10% year over year. The shift in how that money gets allocated is also telling: in 2025, 47% of respondents said identifying new AI use cases was a top spending priority. In 2026, that number dropped to 37%. Meanwhile, optimizing existing AI workflows and production cycles jumped from 34% to 47% as the top spending category. The industry has found its use cases. Now it&#8217;s scaling them. That&#8217;s a very different market than it was 18 months ago, and it means the infrastructure layer underneath those production deployments is about to get a lot more important.</p><blockquote><p>That infrastructure layer is predominantly built on NVIDIA.</p></blockquote><h2>Section 2: BioNeMo and the Drug Discovery Revolution</h2><p>If you&#8217;re an investor in biotech-adjacent AI or a founder building anything in the drug discovery or precision medicine space, BioNeMo is the framework you need to understand cold. It&#8217;s NVIDIA&#8217;s platform specifically built for AI-driven drug discovery, and it represents a genuine architectural shift in how preclinical R&amp;D gets done.</p><p>The traditional drug discovery pipeline is one of the most inefficient processes in all of industry. Average time from target identification to approved therapy runs somewhere between 12 and 15 years. Average cost is north of 2 billion dollars depending on the study, with failure rates exceeding 90% in clinical trials. Most of that failure happens because the preclinical computational work simply wasn&#8217;t good enough to predict what would happen in a human body. BioNeMo attacks that problem at the source.</p><p>The platform runs generative AI models on NVIDIA GPUs and is designed to let models navigate what the documentation calls vast biochemical universes, meaning the combinatorial space of possible molecular structures, protein interactions, and binding predictions that would take human researchers lifetimes to explore manually. The system can design candidate drug molecules and predict molecular interactions at atomic precision, compressing discovery timelines from years to months in favorable cases.</p><p>The architecture is worth understanding in some detail because it tells you what kinds of startups can actually build on top of it. BioNeMo has three distinct layers. The NIM microservices layer delivers pre-trained state-of-the-art models through standardized APIs, meaning a team of five engineers can access world-class molecular simulation capabilities without standing up massive infrastructure. The BioNeMo Framework layer is the adaptation layer, where scientists and developers can fine-tune models on proprietary molecular or genomic data. This is where the moat gets built for commercial companies. The BioNeMo Blueprints layer operationalizes entire workflows into what NVIDIA describes as self-learning loops of design, make, test, and learn, essentially autonomous research cycles that iterate without constant human input.</p><p>For investors, the strategic implication is that the old drug discovery model, where value was concentrated in massive R&amp;D operations with thousands of bench scientists, is getting disaggregated. A small team with access to BioNeMo, a compelling proprietary dataset, and a focused therapeutic hypothesis can now do work that would have required a mid-size pharma company&#8217;s computational biology department a decade ago. That changes the addressable competitive landscape for biotech startups considerably. It also means the venture returns math on early-stage biotech AI companies looks different than it did five years ago, both in terms of capital efficiency going in and in terms of partnership and acquisition interest from large pharma on the way out.</p><p>The NVIDIA survey data from pharma and biotech organizations reinforces this. Literature review and analysis was the top agentic AI use case at 55% for that segment. Drug discovery and biomarker identification came in at 48%. These aren&#8217;t exploratory pilots anymore. Nearly half of pharma and biotech respondents have AI agents running in their discovery workflows. Forty-six percent of that segment reported ROI from their drug discovery AI investments.</p><h2>Section 3: MONAI and the Medical Imaging Flywheel</h2><p>Medical imaging is the use case where healthcare AI ROI is most clearly established, and MONAI, the Medical Open Network for AI, is the open-source framework underpinning a significant share of that activity. Fifty-seven percent of medtech respondents in the NVIDIA survey reported ROI from AI in medical imaging. That&#8217;s the highest confirmed ROI rate of any specific use case across any segment in the entire report.</p><p>MONAI has 6.5 million downloads as of the latest data, has been cited in over 4,000 peer-reviewed papers, and has won more than 20 international medical AI competitions, frequently outperforming proprietary tools. Those are credibility numbers that matter when you&#8217;re trying to get a hospital system&#8217;s IT governance committee to approve a new AI vendor. The fact that the underlying framework is open-source, well-documented, and academically validated is a genuine adoption accelerant in a segment that treats vendor risk with extreme caution.</p><p>The technical architecture of MONAI is what makes it interesting from a buildout perspective. It provides domain-optimized tooling across the full imaging pipeline, from interactive 3D segmentation to multimodal vision-language models that can integrate imaging data with clinical text and other modalities. That last piece, multimodal integration, is where the real clinical value gets generated. A model that can look at a CT scan and simultaneously contextualize findings against a patient&#8217;s clinical notes, lab values, and medication history is a fundamentally different tool than a model that just classifies images. MONAI&#8217;s architecture is designed to support that kind of integration.</p><p>For the medtech segment specifically, the survey data shows medical imaging at 61% as the top use case, followed by clinical decision support at 42% and diagnostic testing including disease diagnosis and risk prediction at 34%. The medtech segment was also the one where computer vision ranked as the top AI workload area at 59%, ahead of generative AI. That&#8217;s counterintuitive relative to the rest of the industry but makes perfect sense when you think about what medtech companies actually build. CT scanners, MRI machines, pathology slide analyzers, and ultrasound systems are fundamentally computer vision applications running on specialized hardware.</p><p>The imaging AI market is also one of the few areas in health tech where the reimbursement pathway is reasonably established. The FDA has cleared or authorized over 950 AI-enabled medical devices as of early 2026, and a growing number of those have CPT codes for reimbursement. Founders building imaging AI companies on MONAI-based infrastructure are entering a market with a defined regulatory playbook and actual payment mechanisms, which is not something you can say about most digital health categories. That combination of technical maturity, regulatory precedent, and demonstrated ROI is why imaging continues to attract disproportionate capital relative to its share of total healthcare AI activity.</p><h2>Section 4: Isaac for Healthcare and the Robotics Buildout</h2><p>The robotics angle on healthcare AI is probably the least appreciated opportunity in the current investment landscape, partly because the timelines are longer and the capital requirements are higher, and partly because most health tech investors don&#8217;t come from a robotics background. NVIDIA&#8217;s Isaac for Healthcare platform is worth understanding regardless, because it&#8217;s defining the development environment for what will likely be a very large market segment over the next decade.</p><p>Isaac for Healthcare is a simulation and deployment platform that gives medical robotics developers a complete end-to-end pipeline from virtual environment construction through AI model training to real-world hardware deployment. The workflows currently supported are a useful guide to where commercial activity is concentrating. Robotic surgery and surgical assistant robotics using the SO-ARM101 manipulator are the furthest along, with full pipelines for data collection, policy training, and deployment. Robotic ultrasound, telesurgery with haptic feedback and low-latency video streaming, and hospital automation workflows are also supported.</p><p>The concept of a digital twin of a hospital, where care teams can simulate procedures and train AI models before any patient is involved, is no longer a speculative idea. The Isaac platform makes it operational. Developers can build sim-ready assets that mirror real hospital environments, run AI model training against those synthetic environments, and then deploy trained models to actual hardware. The implications for clinical trial design, staff training, and surgical outcomes research are significant and mostly unappreciated outside of a relatively small circle of surgical robotics investors.</p><p>For investors, the key strategic question around surgical robotics AI is whether the value accrues at the hardware layer or the software layer. The Intuitive Surgical model has historically concentrated value at the hardware layer through platform lock-in, but that model is under meaningful competitive pressure from newer entrants building software-defined surgical systems on open platforms. Isaac for Healthcare is explicitly designed to support the software-defined model, where the intelligence of the surgical system is continuously updated through software rather than through hardware replacement cycles. That&#8217;s a fundamentally better business model for recurring revenue, and it&#8217;s the direction the market is moving.</p><p>The hospital automation workflow, called Rheo in the Isaac documentation, is also worth flagging. Autonomous hospital logistics, medication delivery, specimen transport, and environmental services robots are a category that has been commercially challenging historically because the environments are complex and unpredictable. Simulation-trained robotics on Isaac infrastructure addresses the training data problem directly, generating synthetic environments that cover the edge cases a real-world training program would take years to encounter organically. The operational cost savings potential in hospital logistics is substantial given nursing labor costs and the administrative burden that non-clinical tasks impose on clinical staff.</p><h2>Section 5: Holoscan and the Edge Intelligence Layer</h2><p>Holoscan is the NVIDIA product that gets the least attention from health tech investors, which is a mistake given its technical position. It&#8217;s a multimodal AI sensor processing platform designed specifically for real-time inference on streaming data at the edge. In healthcare terms, that means running AI directly on medical devices during procedures, not sending data to a cloud, not adding latency, not depending on network connectivity, doing inference in the operating room or at the point of care in real time.</p><p>The surgical video workflow is the clearest current application. Holoscan enables low-latency AI processing of surgical video feeds with real-time tool detection and segmentation. The modular pipeline architecture means that medtech companies can integrate specific AI models for their procedure type without rebuilding the underlying infrastructure. The Holoscan Sensor Bridge extends this to arbitrary sensor types, handling high-bandwidth data from diverse sensors over Ethernet with a standard API and open software built on an FPGA interface.</p><p>The HoloHub repository is where the reference applications live, and browsing it gives a good sense of where commercial development is concentrating. End-to-end surgical video, body pose estimation, integration with 3D Slicer for surgical planning, and augmented reality volume rendering via Magic Leap are among the available workflows. The 3D Slicer integration in particular is interesting because 3D Slicer is already deeply embedded in surgical planning workflows at academic medical centers, meaning Holoscan can slot into existing institutional infrastructure rather than requiring greenfield adoption.</p><p>The broader strategic significance of Holoscan is that it enables a class of medical AI applications that cloud-dependent architectures fundamentally cannot support. Real-time intraoperative guidance, where a surgeon needs AI feedback within milliseconds of a camera movement, cannot tolerate cloud round-trip latency. Real-time monitoring of critically ill patients where response time is measured in seconds, not minutes, requires edge inference. These are the high-value, high-stakes applications where AI actually changes clinical outcomes rather than just administrative efficiency, and Holoscan is the platform specifically built for them.</p><p>The NVIDIA survey data on hybrid computing is relevant here. The shift from 35% to 43% using hybrid computing for AI workloads year over year, concurrent with cloud-only dropping from 41% to 35%, reflects exactly the trend Holoscan is positioned to capture. Healthcare organizations are figuring out that some workloads need to live at the edge, and they&#8217;re building infrastructure to support that. For founders building device-adjacent AI companies, the question of which inference platform to build on has a fairly clear answer at this point.</p><h2>Section 6: Parabricks and the Genomics Data Deluge</h2><p>Parabricks is NVIDIA&#8217;s GPU-accelerated genomics software suite, and the market context for it is almost comically large. The genomics field is heading toward tens of exabytes of sequencing data in the coming decade as sequencing costs continue their exponential decline. The cost to sequence a human genome has dropped from roughly 100 million dollars in 2001 to under 200 dollars in 2026. The problem has completely flipped: getting the sequence is now the easy part. Making sense of it is where the bottleneck is.</p><p>Parabricks handles the secondary analysis layer, taking raw sequencing output and doing the alignment, variant calling, and related processing that turns raw reads into interpretable genomic data. The GPU acceleration cuts processing runtimes from hours to minutes for standard whole-genome sequencing pipelines. On a practical level, that&#8217;s the difference between genomic results being available for clinical decision-making during a hospitalization versus arriving days after discharge. In neonatal intensive care, oncology, and rare disease settings, that timeline difference is clinically material.</p><p>The CUDA-X Data Science suite, formerly RAPIDS, handles single-cell and tertiary analysis, which is where population genomics and research applications live. Combined with Parabricks for secondary analysis and GPU-accelerated primary analysis during sequencing itself, NVIDIA now has coverage across the entire genomics computational pipeline.</p><p>The investment angle here is mostly about what Parabricks makes possible downstream rather than Parabricks itself as an investment target. When genomic data can be processed in near-real-time at scale, the market for clinical genomics interpretation, population health genomics, and genomically-informed drug target identification all expand substantially. Any startup operating in those application layers benefits directly from the infrastructure improvement. The pharma and biotech segment ranked genomic applications at 44% as the second most common AI use case in the NVIDIA survey, just behind drug discovery at 57%. That&#8217;s a meaningful indicator of where R&amp;D capital is flowing.</p><h2>Section 7: Clara, NIM, and the Open Source Bet</h2><p>The open-source question in healthcare AI is worth addressing directly because the NVIDIA survey data on it is unusually strong. Eighty-two percent of respondents said open-source models and software were moderately to extremely important to their AI strategy. Fifty-seven percent said they were very or extremely important. This is not a fringe preference. It&#8217;s the dominant strategic orientation of the people actually building and deploying healthcare AI.</p><p>The logic is not hard to follow. Healthcare AI applications tend to be highly specific. An imaging AI for detecting early-stage pancreatic cancer in CT scans has a fundamentally different training distribution than one for detecting pneumonia in chest X-rays. A clinical documentation model tuned on oncology notes performs differently than one tuned on emergency department encounters. General-purpose foundation models are a starting point, not a finish line. Fine-tuning on proprietary clinical data, using open-source frameworks that allow full customization without licensing constraints, is how organizations build AI that actually works in their specific context.</p><p>Clara is NVIDIA&#8217;s family of open models and tools purpose-built for scientific discovery, medical imaging, and biology and chemistry research. It includes models, development recipes, and evaluation frameworks across imaging, biology, and drug discovery domains. The strategic positioning of Clara as an openly accessible platform is a deliberate move to accelerate ecosystem adoption, and it&#8217;s working given the MONAI download numbers and the peer-reviewed citation count.</p><p>NIM microservices are the delivery mechanism for accessing NVIDIA&#8217;s most capable models through standardized APIs without requiring teams to manage underlying infrastructure. For health tech startups, NIM is particularly valuable because it removes the infrastructure overhead from model deployment, letting small engineering teams focus on the application layer rather than on GPU cluster management. The combination of NIM for access to foundation capabilities and Clara or BioNeMo for domain-specific fine-tuning gives a startup team a genuinely competitive technical starting point without enterprise-scale infrastructure.</p><p>The implication for the build-vs-buy question that every health tech founder faces is increasingly clear. Buying access to general-purpose AI from large model providers and building on top of it without domain customization produces mediocre results in clinical applications. Building everything from scratch is prohibitively expensive for most startups. The open-source fine-tuning path, using NVIDIA&#8217;s domain-specific frameworks as the foundation and layering in proprietary clinical data, is where the best risk-adjusted technical outcomes are happening. The survey data on where ROI is concentrating supports this: the organizations reporting the highest AI ROI are the ones applying specific AI to distinct use cases, not the ones deploying general-purpose tools broadly.</p><h2>Section 8: What This Means for Investors and Founders</h2><p>The picture that emerges from the NVIDIA platform ecosystem combined with the 2026 survey data is of an infrastructure stack that is simultaneously mature enough to build on and early enough that the application layer is not yet crowded. That&#8217;s a genuinely rare combination in health tech, which tends to alternate between underdeveloped infrastructure that makes building too hard and overcrowded application markets where differentiation is nearly impossible.</p><p>For angel investors and syndicate leads specifically, the portfolio construction implications break down across a few dimensions. First, any company building clinical AI applications that is not building on top of GPU-accelerated infrastructure, whether NVIDIA-based or otherwise, should be asked hard questions about how they plan to compete as inference demands scale. The survey data showing hybrid computing adoption at 43% and climbing reflects a market that is normalizing GPU infrastructure costs. That normalization reduces the infrastructure moat of companies that built early on proprietary compute and increases the importance of application-layer differentiation.</p><p>Second, the agentic AI debut at 47% usage or assessment is the number to watch. Agentic AI systems that can autonomously reason, plan, and execute multi-step healthcare tasks represent a qualitative jump in what AI can do in clinical and research settings. The current top use cases, knowledge management and retrieval at 46%, literature review at 38%, and internal process optimization at 37%, are mostly back-office. The interesting commercial territory is what happens when agentic AI moves into clinical workflows in earnest. The regulatory environment, HIPAA, FDA, and similar frameworks, remains the primary constraint on that transition, with 40% of respondents citing regulatory compliance as the top factor influencing their agentic AI implementation approach. Founders who can navigate that regulatory surface with defensible governance frameworks are building a real moat.</p><p>Third, the small company revenue data deserves more attention than it usually gets. Fifty-six percent of small healthcare AI companies reported more than 10% annual revenue growth attributable to AI, versus 44% for large companies. That&#8217;s counterintuitive given that larger organizations have more resources, more data, and more infrastructure. The explanation is probably that small companies are better at applying AI to a single well-defined problem rather than trying to boil the ocean, which maps to the survey&#8217;s broader finding that specific AI applied to distinct use cases outperforms general-purpose deployment on ROI metrics. For early-stage investors, this is an argument for funding companies with narrow, well-defined wedge applications over companies pitching broad platform plays.</p><p>The infrastructure inequality finding between large and small organizations is the one genuine caution flag in the data. Forty percent of small healthcare AI companies cited budget as their top challenge. Thirty-three percent cited data size constraints for model training. These are structural disadvantages that don&#8217;t go away on their own, and they&#8217;re why the open-source ecosystem matters so much for small company competitiveness. NIM microservices, MONAI, BioNeMo, Parabricks, and the Clara model families collectively give small teams access to capabilities that would have required dedicated ML infrastructure teams just a few years ago. The playing field is leveling in terms of model access, but capital constraints on compute and data acquisition remain real.</p><p>The overall thesis is not complicated. NVIDIA has built the most comprehensive AI infrastructure stack specifically targeting healthcare and life sciences. It spans drug discovery through BioNeMo, medical imaging through MONAI, surgical and hospital robotics through Isaac for Healthcare, edge inference through Holoscan, genomics through Parabricks, and open model access through Clara and NIM. The survey data shows an industry that has crossed the adoption inflection point, is generating measurable ROI, is increasing budgets, and is shifting from experimentation to production scaling. The companies building serious clinical AI applications in 2026 and beyond are building on this infrastructure whether they realize it or not, and the investors who understand the stack have a meaningful edge in evaluating who is building on solid ground versus who is building on sand.</p><p>The picks-and-shovels metaphor gets overused in tech investing but it&#8217;s actually apt here. In the 1849 California Gold Rush, the people who got rich selling picks and shovels did so because they got paid regardless of which miner struck gold. NVIDIA&#8217;s position in healthcare AI infrastructure is structurally similar. Every drug discovery company that finds a novel molecule using BioNeMo, every radiology AI company that deploys on MONAI, every surgical robotics company that trains on Isaac, every genomics company running Parabricks in production, they all run on NVIDIA. The question for founders is how to build a durable application-layer business on top of that infrastructure. The question for investors is which of those application-layer bets are most likely to generate asymmetric returns given a market that is clearly in the early innings of a sustained scaling cycle.</p><p>The NVIDIA survey&#8217;s closing observation is probably the right one to end on. The researchers predict that by 2027, healthcare AI will shift from predominantly predictive analytics toward more consistent deployment of agentic systems capable of reasoning across patient populations, clinical trials, and care workflows simultaneously. That transition, if it happens anywhere close to that timeline, will be the most significant shift in how clinical decisions get made since evidence-based medicine became the standard of care in the 1990s. The infrastructure to support it already exists. The regulatory frameworks are catching up. The capital is moving in. This is the moment to be paying very close attention.&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!toV1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!toV1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 424w, https://substackcdn.com/image/fetch/$s_!toV1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 848w, https://substackcdn.com/image/fetch/$s_!toV1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!toV1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!toV1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg" width="1290" height="688" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:688,&quot;width&quot;:1290,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!toV1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 424w, https://substackcdn.com/image/fetch/$s_!toV1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 848w, https://substackcdn.com/image/fetch/$s_!toV1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!toV1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdac3107-56dd-4c92-84dd-70d2bf338bc3_1290x688.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[The MATCH Monopoly and What It Actually Means for Health Tech]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-match-monopoly-and-what-it-actually</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-match-monopoly-and-what-it-actually</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Sat, 28 Mar 2026 13:22:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!DcpO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>The House Judiciary Subcommittee&#8217;s investigation into the National Resident Matching Program (NRMP) antitrust exemption and its downstream effects on the physician workforce, health system economics, and health tech investment.</p><h3>Key facts:</h3><p>- NRMP established 1952; congressional antitrust exemption passed 2004 (15 U.S.C. 37b)</p><p>- In 2025, 52,498 medical students applied for 43,237 residency positions; ~9,000 went unmatched</p><p>- Average PGY1 resident salary: ~$68,166 in 2025; Medscape reports average across all years ~$75k</p><p>- Resident pay 2020-2024 did not keep pace with inflation; over 70% of residents say they need at least a 26% raise</p><p>- AAMC projects a US physician shortage of up to 86,000 by 2036</p><p>- Resident Physician Shortage Reduction Act of 2025 proposes 14,000 new Medicare-funded GME slots over 7 years</p><p>- House Judiciary Subcommittee convened formal hearing May 14, 2025 titled &#8220;The MATCH Monopoly: Evaluating the Medical Residency Antitrust Exemption&#8221;</p><p>- Repealing the exemption would not condemn the NRMP but would open it to antitrust scrutiny under Sherman Act Section 1</p><h3>Why health tech investors should care:</h3><p>- Physician supply directly constrains addressable markets for telehealth, AI clinical tools, and virtual-first care models</p><p>- GME funding reform and workforce disruption create real enterprise software and infrastructure opportunities</p><p>- Wage suppression dynamics and resident workflow are underexplored verticals for health tech founders</p><h2>Table of Contents</h2><p>The Setup: What the MATCH Actually Is</p><p>The 2004 Exemption and Why Congress Is Back on It</p><p>The Wage Problem: Residents Are Getting Paid Less Than Subway Managers</p><p>The Supply Side: Where the Bottleneck Really Lives</p><p>What Happens If the Exemption Goes Away</p><p>The Health Tech Angle: Why This Should Land in Your Investment Thesis</p><p>The Bottom Line</p><h2>The Setup: What the MATCH Actually Is</h2><p>Most people outside medicine have no idea the MATCH exists, which is kind of wild given that it controls how roughly 43,000 doctors a year enter the workforce. The National Resident Matching Program has been running since 1952 and is the infrastructure layer that pairs graduating medical students with residency training programs. It is, functionally, the only game in town. Students submit ranked lists of programs they want to train at. Programs submit ranked lists of students they want. An algorithm developed originally in the 1950s and later revised following Alvin Roth&#8217;s Nobel-winning work on market design runs the match. Everyone finds out on Match Day. You either matched or you didn&#8217;t. There is no negotiation, no counter-offer, no chance to say you got a better deal elsewhere. You get assigned and you go.</p><p>It is worth pausing on the Alvin Roth connection for a second because it matters. Roth&#8217;s contribution was making the algorithm &#8220;applicant-optimal,&#8221; meaning if a more preferred program was available, the algorithm would find it. This was genuinely a real improvement over the prior setup and gave the NRMP legitimate academic cover for decades. The problem is that a well-designed algorithm can still produce bad market outcomes if the broader market conditions surrounding it are distorted. And that is where things get messy.</p><p>The system sits on top of an accreditation monopoly that almost nobody talks about. ACGME, the Accreditation Council for Graduate Medical Education, controls which residency programs are legitimate. If you want to be a licensed physician in virtually any specialty in the US, you need an ACGME-accredited residency. ACGME and the NRMP together form what the House Judiciary Subcommittee called a bottleneck to the physician workforce. That framing is a little dramatic but not entirely wrong. The two organizations function in sequence: ACGME certifies the programs, NRMP fills them. Miss either gate and you are out of the physician pipeline.</p><p>The 2004 Exemption and Why Congress Is Back on It</p><p>The antitrust exemption that shields the MATCH from federal and state antitrust enforcement did not materialize from nowhere. It came directly from a 2002 class action lawsuit filed on behalf of medical residents alleging that the NRMP constituted a violation of Sherman Act Section 1. The argument was pretty straightforward from an antitrust standpoint: a centralized, closed hiring system in which employers share compensation information and applicants cannot negotiate independently looks a lot like horizontal wage-fixing, which courts have historically treated as per se illegal. Congress, apparently deciding that defending this setup in court was too risky and that the matching system was net-positive for medical education, passed the Pension Funding Equity Act in 2004 with a rider tucking in the exemption. Codified at 15 U.S.C. 37b, the statute simply declares it not unlawful to sponsor, conduct, or participate in a graduate medical education residency matching program.</p><p>That worked fine for about 20 years. Then in March 2025, Rep. Scott Fitzgerald&#8217;s House Judiciary Subcommittee on the Administrative State, Regulatory Reform, and Antitrust sent letters to basically everyone involved: ACGME, the AMA, the American Osteopathic Association, AAMC, the NRMP itself, and a handful of academic medical centers including Stanford, Duke, MedStar-Georgetown, and Philadelphia College of Osteopathic Medicine. The letters requested documents on the matching algorithm, salary information sharing, restrictions on resident mobility, and concerns from residents or programs about NRMP and ACGME conduct. The deadline was March 28, 2025, a turnaround that implied the subcommittee was not messing around.</p><p>The formal hearing followed on May 14, 2025 under the title &#8220;The MATCH Monopoly: Evaluating the Medical Residency Antitrust Exemption.&#8221; The NRMP notably did not testify. Witnesses included a geriatric medicine specialist and clinical professor, an attorney specializing in health law, a senior fellow from AEI, and a physician with dual training in medicine and public health policy. The testimony split roughly as expected: critics argued the system restricts mobility and suppresses wages; defenders warned that blowing up a 70-year-old matching infrastructure without a clear alternative would create chaos, class inequities in residency placement, and potentially fewer residents overall as institutional willingness to participate fractured. Both sides made legitimate points.</p><h2>The Wage Problem: Residents Are Getting Paid Less Than Subway Managers</h2><p>This one deserves its own section because it is genuinely jarring once you look at the numbers. In 2025, the mean PGY1 resident salary is $68,166 according to AAMC data. The median sits around $66,986. Medscape&#8217;s broader survey, averaging across all training years, puts it closer to $75,000. These numbers sound okay until you factor in that most residents are working 60 to 80 hours a week, many in urban metros with serious cost-of-living exposure, and all of them entered training carrying somewhere between $200,000 and $350,000 in medical school debt on average. When residents have literally calculated their hourly rate and compared it to minimum wage, and the math has come out unfavorable, you have a structural compensation problem.</p><p>Resident pay increased about 6.5% in 2025, which sounds like progress but was preceded by years of basically flat growth. Medscape&#8217;s own data shows that from 2020 to 2024, resident salary increases did not keep pace with general US inflation. That means the real wage for medical residents was declining during a period of extraordinary cost-of-living pressure. Over 70% of residents surveyed by Medscape say they need at least a 26% raise. A third say they need 51% or more, particularly after accounting for debt service.</p><p>The antitrust case against the MATCH on wages is that in a free market, residents would be able to negotiate competing offers. One of the residents who provided testimony to the subcommittee stated pretty directly that if he had been able to receive and negotiate offers from multiple institutions in parallel, he would have secured better compensation from Stanford. That is a real and specific harm. The NRMP&#8217;s rules prohibit programs from extending employment commitments outside the match timeline, which means even if a program wants to pay more to attract talent, the mechanism for doing so competitively is structurally blocked. The Council of Teaching Hospitals further facilitates salary information sharing among programs, which critics argue creates a de facto floor that serves institutional interests rather than resident ones.</p><h2>The Supply Side: Where the Bottleneck Really Lives</h2><p>Here is where things get legitimately complicated and where the antitrust framing may actually be misleading. In 2025, the NRMP ran its largest match ever: 43,237 positions offered, up 4.2% from 2024, with 877 more primary care slots than the prior year. Internal medicine alone placed 11,750 positions. That sounds like supply is growing, and it is. But 52,498 students applied for those positions. About 9,000 went unmatched. That gap is the physician shortage in its most concentrated form.</p><p>The supply constraint is not primarily an NRMP problem. It is a Medicare funding problem. In 1997, Congress capped the number of residency positions Medicare would fund per hospital. That cap has not moved meaningfully in almost 30 years. Teaching hospitals are the primary employers of residents and they depend heavily on GME funding from Medicare to make the economics of training work. The cap created a ceiling on resident supply that has nothing to do with the NRMP or the matching algorithm. You can redesign the match a thousand different ways and you still run into the same federally imposed cap on how many trained physicians can enter the workforce per year.</p><p>AAMC projections are straightforward and alarming: a shortage of up to 86,000 physicians by 2036. That is not a rounding error. It is the result of an aging population requiring more care, a significant chunk of the existing physician workforce approaching retirement, and a training pipeline constrained by a 1997 funding mechanism that no one has had the political will to meaningfully update. The Resident Physician Shortage Reduction Act of 2025, introduced by Reps. Sewell and Fitzpatrick with bipartisan Senate support, proposes adding 14,000 Medicare-supported GME slots over seven years starting in 2026, with 2,000 positions per year. It also makes the Rural Residency Planning and Development program permanent with $12.7 million annually. Good bill. Prior versions failed in 2019, 2021, and 2023. There is no particular reason to assume 2025 is different, but the political environment around physician shortages has shifted enough that it might actually move this time.</p><h2>What Happens If the Exemption Goes Away</h2><p>Repealing the 2004 antitrust exemption would not kill the NRMP. This is probably the most important clarification to make because it gets lost in the political framing. Removing the exemption would open the NRMP to antitrust scrutiny, meaning plaintiffs could bring cases and courts would evaluate whether specific practices cause anticompetitive harm. Under Sherman Act Section 1, most agreements are subject to a rule of reason analysis that weighs procompetitive benefits against anticompetitive effects. A well-designed matching algorithm that creates genuine efficiencies in a complex labor market would likely survive that scrutiny even without an explicit exemption. The problem for the NRMP is that several specific practices, particularly the restrictions on parallel negotiation and the salary information-sharing architecture, might not survive. Those are the pieces most likely to face successful legal challenges post-repeal.</p><p>The NRMP&#8217;s own December 2025 letter to the medical education community made its position clear: the system creates fair, efficient placement by ensuring all positions and all applicants are available simultaneously, preventing the kind of scramble and exploding-offer dynamic that plagued medical hiring before 1952. That is a legitimate point. The pre-NRMP era was reportedly a disaster of early offers, pressure tactics, and students being forced to commit before they had meaningful information. Nobody actually wants to go back to that. But the argument that historical dysfunction justifies a blanket antitrust exemption in perpetuity is a stretch, and Congress clearly agrees.</p><p>The more interesting scenario is not full repeal but targeted reform. Several hearing participants suggested the better path is to modify the exemption rather than eliminate it entirely. Specifically, allowing residents to engage in some degree of parallel negotiation around compensation, while preserving the central matching mechanism for placement, might address the wage suppression concerns without introducing the chaos that full repeal would risk. That version of reform is politically messier because it requires Congress to actually write substantive policy rather than just rescind an exemption, but it is probably the outcome that produces the best results for residents, programs, and patients.</p><h2>The Health Tech Angle: Why This Should Land in Your Investment Thesis</h2><p>This is the part that most coverage of the MATCH antitrust story completely ignores, which makes sense because most health policy coverage is written for policy people and not for founders and investors. But the downstream implications for the health tech stack are real.</p><p>Start with physician supply as a fundamental constraint on addressable markets. Every investment thesis in telehealth, AI-assisted diagnostics, virtual-first primary care, and remote monitoring implicitly assumes a certain density of physician capacity. When that capacity is constrained by a structurally broken training pipeline, the thesis has to account for substitution effects, automation curves, and mid-level provider expansion that would not otherwise be as prominent. Nurse practitioners and physician assistants have been absorbing demand that would historically go to physicians, and that trend accelerates in a supply-constrained environment. That is relevant to companies thinking about clinical workflow design, scope-of-practice tooling, and care team orchestration software.</p><p>Second, the GME funding reform conversation is a legitimate enterprise software opportunity that nobody is building for yet. If the Resident Physician Shortage Reduction Act passes and 2,000 new positions a year start flowing to hospitals, those hospitals need infrastructure to administer expanded residency programs, track GME funding compliance, manage accreditation requirements, and optimize resident scheduling. ACGME compliance and GME financial management are genuinely painful administrative domains. There is a real VRAM or MedHub-adjacent opportunity for a founder who wants to own GME administration software in a world where the number of training programs and positions is growing faster than administrative capacity.</p><p>Third, if the antitrust exemption is repealed or materially reformed, resident compensation will almost certainly increase. The wage suppression effect of the current system, even accounting for the NRMP&#8217;s defenders, is real enough that loosening the negotiation constraints would move average resident pay up. That is relevant to benefits platforms, financial wellness tools, and resident-facing fintech companies that have historically treated residents as low-priority because their compensation was both low and non-negotiable. A market where residents are making $90,000 to $120,000 instead of $68,000 looks meaningfully different from a product and monetization standpoint. Companies like Laurel Road, which specifically targets the medical trainee segment for student loan refinancing and banking, are already in this space, but it is not crowded.</p><p>Fourth and maybe most interesting from a longer-term horizon: the physician shortage math creates a structural tailwind for AI clinical decision support and ambient AI documentation tools that is independent of whatever happens with the MATCH. If the US is genuinely going to be 86,000 physicians short by 2036, the system will compensate through some combination of scope-of-practice expansion, telehealth penetration, and productivity tools that let existing physicians handle higher patient volumes with lower cognitive and administrative burden. Companies building AI scribes, pre-charting automation, order set optimization, and AI-assisted care navigation are not just nice-to-haves in a shortage environment; they become infrastructure. The shortage is the tailwind. Investors who are modeling physician supply as a stable input into healthcare demand equations are probably underestimating how disruptive a genuine shortage will be to care delivery models and therefore to which tech categories win.</p><h2>The Bottom Line</h2><p>The MATCH is not going away. The NRMP has too much institutional legitimacy, too many genuine defenders in academic medicine, and too long a track record of producing reasonably efficient placements for Congress to simply delete it. The more likely outcome is a targeted modification of the antitrust exemption, probably accompanied by some version of the Resident Physician Shortage Reduction Act that expands GME slots incrementally over seven years while preserving the core matching infrastructure. Residents will probably see modest compensation gains as a result, though not the 51% some of them are asking for. The physician shortage will continue getting worse in the near term regardless, because even 14,000 new slots over seven years is nowhere near the 86,000-physician gap that AAMC is projecting.</p><p>For health tech investors and founders, the MATCH story is really two stories running in parallel. The first is a labor market reform story about whether residents will be treated more like the highly skilled professionals they are rather than captive trainees in a system designed primarily for institutional convenience. The second is a supply shock story about what happens to the entire US healthcare delivery infrastructure when the doctor pipeline has been structurally constrained for decades and is now catching up to demand curves that nobody planned for. Both stories create dislocation. Dislocation creates opportunity. The question is whether your firm is positioned to see the opportunity or whether it is still modeling physician supply as a fixed variable in a static healthcare market. It is anything but.&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;&#8203;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DcpO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DcpO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DcpO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DcpO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DcpO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DcpO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg" width="1290" height="810" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:810,&quot;width&quot;:1290,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DcpO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DcpO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DcpO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DcpO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f974cc6-fefa-433b-8a29-1dee166ab5e8_1290x810.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[The Elon Terrawatt Announcement Nobody in Health Tech Is Taking Seriously Enough]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-elon-terrawatt-announcement-nobody</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-elon-terrawatt-announcement-nobody</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Mon, 23 Mar 2026 11:49:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GKYE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33c02ab1-c30e-43c8-ada3-4b8db9e008e0_596x336.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>Elon Musk&#8217;s April 2025 announcement of the &#8220;Terrafab&#8221; project, a joint venture between Tesla, xAI, and SpaceX to build an advanced semiconductor fabrication facility in Austin, Texas, aimed at producing a terawatt of compute per year, and what this means for healthcare AI infrastructure, health tech investment, and the broader trajectory of computational medicine.</p><h3>Key claims assessed:</h3><p>- Current global AI compute output is roughly 20 gigawatts per year; all existing fabs combined represent about 2% of what the Terrafab targets</p><p>- Space-based solar AI compute may undercut terrestrial compute costs within 2-3 years</p><p>- Edge inference chips optimized for Optimus humanoid robots could reach production volumes of 1-10 billion units per year vs. 100M vehicles globally today</p><p>- The fab includes in-house lithography mask production, enabling a chip design iteration loop with no known global equivalent</p><p>- Healthcare AI infrastructure is directly in the blast radius of this shift, whether or not health tech investors are paying attention</p><p>Why it matters for health readers: compute constraints are already the binding limit on clinical AI deployment at scale, and this announcement represents a potential step-change in the supply curve that will reprice everything from EHR automation to genomic inference to surgical robotics.</p><h2>Table of Contents</h2><p>What Actually Got Announced (And Why the Framing Was Weird)</p><p>The Compute Constraint Nobody Talks About in Health Tech</p><p>Edge Inference and the Optimus Variable</p><p>Space Compute Is Not Sci-Fi, It&#8217;s a Cost Curve</p><p>What the Iteration Loop Means for Medical Chip Design</p><p>How This Lands for Health Tech Investors</p><p>The Long Game: Kardashev, Abundance, and What It Means for Healthcare Economics</p><h2>What Actually Got Announced (And Why the Framing Was Weird</h2><p>So Musk opened with Kardashev civilizations and galactic expansion, which understandably caused a lot of eyerolls. That framing probably caused most serious health tech operators and investors to tune out around minute three and go back to their IRR models. That would be a mistake. Buried inside the cosmic rhetoric was a genuinely substantive industrial announcement with near-term implications that are very hard to dismiss once you actually read the transcript carefully.</p><p>The core of it: Tesla, xAI, and SpaceX are jointly building what they&#8217;re calling the Terrafab, starting with an advanced semiconductor fabrication facility in Austin. The word &#8220;advanced&#8221; is doing a lot of work there. This is not a packaging plant or an assembly operation. The claim is that a single building will house all of the equipment needed to produce logic chips, memory chips, perform packaging, run testing, and crucially, manufacture the lithography masks themselves. That last part is what makes this unusual. Lithography mask production is typically a completely separate, highly specialized operation that sits upstream of fab operations and represents one of the most capital-intensive and time-sensitive bottlenecks in chip development. Putting that inside the same building as the fab closes a loop that doesn&#8217;t currently exist anywhere in the world at this scale.</p><p>The context Musk gave for why this is necessary is actually pretty clean arithmetic. Current global AI compute output is roughly 20 gigawatts per year. The Terrafab project targets a terawatt of compute, which is 1,000 gigawatts. That means the existing output of every semiconductor manufacturer on earth combined, including TSMC, Samsung, and everyone else, represents about two percent of what this project is targeting. He said explicitly that he has told those suppliers he will buy every chip they can make, and they&#8217;re still not expanding fast enough to close the gap. So the framing is: build the fab or don&#8217;t have the chips. That&#8217;s it. That&#8217;s the whole logic.</p><p>For anyone in health tech, the instinct might be to treat this as a hyperscaler problem, something relevant to OpenAI or Google but not to a Series B digital health company or a health system trying to deploy clinical AI. That instinct is wrong, and the rest of this essay is essentially an argument for why.</p><h2>The Compute Constraint Nobody Talks About in Health Tech</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-elon-terrawatt-announcement-nobody">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Clinical Reasoning vs. Documentation: The Next Battleground for Medical LLMs]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/clinical-reasoning-vs-documentation</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/clinical-reasoning-vs-documentation</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Fri, 20 Mar 2026 12:19:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qk5d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12cdcf80-c472-4a43-b663-7c6428f3b6c9_850x594.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>The first wave of healthcare AI scored decisive wins in documentation automation. Ambient scribes, coding copilots, and summarization layers delivered clear ROI by solving a well-bounded problem: compress high-entropy clinical inputs into structured, billable outputs. That layer is now saturating. The next frontier is harder, more valuable, and genuinely unsolved: augmenting clinical reasoning itself.</p><h3>This essay covers:</h3><p>- Why documentation AI succeeded and why reasoning AI is fundamentally different</p><p>- The three hard architectural requirements current LLMs only partially meet (state representation, hypothesis generation, uncertainty quantification)</p><p>- Why next-token predictors structurally struggle with clinical cognition</p><p>- Emerging architectures trying to bridge the gap (tool-augmented reasoning, graph-based inference, persistent memory layers)</p><p>- Failure modes unique to reasoning-adjacent systems</p><p>- Why current benchmarks like MedQA are nearly useless for evaluating actual reasoning</p><p>- The economic argument for why reasoning AI creates durable moats that documentation AI cannot</p><p>- A framework for thinking about AI&#8217;s role: Advisor vs. Cognitive Extender vs. Autonomous Reasoner</p><h2>Table of Contents</h2><p>The Documentation Win and Why It&#8217;s Running Out</p><p>Compression vs. Inference: A Real Distinction</p><p>Three Requirements That Break Current LLMs</p><p>Why the Architecture Itself Is the Problem</p><p>What&#8217;s Actually Being Built to Fix This</p><p>The Failure Modes Nobody Talks About</p><p>Evaluation Is Broken and Everyone Knows It</p><p>Three Paradigms for AI&#8217;s Role in Reasoning</p><p>The Economic Case for Betting on Reasoning</p><p>What a Computable Differential Actually Looks Like</p><h2>The Documentation Win and Why It&#8217;s Running Out</h2><p>If you&#8217;ve been paying attention to where healthcare AI dollars have gone over the past four years, the pattern is pretty obvious. Ambient scribes, prior auth automation, clinical note summarization, revenue cycle coding assist. Every major health system has piloted at least one of these. Most are in some phase of deployment. Nuance DAX, Abridge, Suki, Nabla, and a handful of EHR-native products from Epic and Oracle have collectively reshaped how clinicians think about administrative burden. That&#8217;s not nothing. It&#8217;s actually a big deal.</p><p>The ROI story for this category is clean and defensible. Physicians were spending somewhere between one and three hours per day on documentation depending on specialty. Ambient scribes demonstrably cut that. KLAS Research data from 2023 showed DAX users saving an average of 7 minutes per note. Multiply that across a 20-patient day and you&#8217;re talking real productivity gains. Payers and health systems could quantify it. CFOs could model it. Procurement decisions got made.</p><p>But here&#8217;s the uncomfortable reality underneath that success: documentation AI is fundamentally a compression problem. It takes high-entropy inputs, which is to say the rambling, overlapping, sometimes contradictory content of a clinical encounter, and transforms them into low-entropy structured outputs. Progress notes. HCC codes. Discharge summaries. The model doesn&#8217;t need to understand what&#8217;s clinically happening. It needs to recognize patterns in language and map them onto the expected structure of a clinical document. That&#8217;s a solved class of problem. Transformers are exceptionally good at it.</p><p>The saturation dynamic is already visible in the market. Ambient scribe functionality is becoming table stakes. Epic and Oracle are bundling it natively. Gross margins are compressing. The differentiation thesis for pure-play documentation vendors is getting harder to sustain. That doesn&#8217;t mean the category is dead. It means the wave has crested and the smart capital is looking at what comes next.</p><h2>Compression vs. Inference: A Real Distinction</h2><p>The phrase &#8220;clinical reasoning&#8221; gets thrown around a lot, sometimes loosely, so it&#8217;s worth being precise about what it actually means and why it&#8217;s a different beast from documentation.</p><p>Documentation is a compression problem. Clinical reasoning is an inference problem. Those are not the same thing, and conflating them has led to a lot of overpromising in the AI health space.</p><p>Here&#8217;s what inference under clinical uncertainty actually involves. A 58-year-old with chest pain and new dyspnea walks into the ED. The clinician is not searching a knowledge base. They are constructing a probabilistic model in real time. What&#8217;s the prior probability of ACS in this demographic given this symptom cluster? How does the troponin trend update that probability? What does the absence of pleuritic component tell me about PE likelihood? If the D-dimer comes back mildly elevated in the context of a recent long flight, how much does that shift things? This is Bayesian updating applied under time pressure across a partially observed, dynamically evolving data set.</p><p>Documentation systems don&#8217;t need to do any of this. They just need to accurately capture and reformat what happened. The clinician already did the reasoning. The scribe just records it.</p><p>The reason this distinction matters for investors and builders is that it defines the size and durability of the value being created. Compressing a clinical note saves time. Improving diagnostic inference changes outcomes. And changing outcomes is where healthcare actually spends money. The US misdiagnosis rate hovers around 12 million adults per year according to data published in BMJ Quality and Safety. Diagnostic error contributes to somewhere between 40,000 and 80,000 deaths annually by most estimates. The economic footprint of that problem dwarfs the administrative burden problem by a considerable margin.</p><h2>Three Requirements That Break Current LLMs</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/clinical-reasoning-vs-documentation">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The AI Factory Is Jensen Huang’s Most Important Keynote in a Decade: Implication for Healthcare]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-ai-factory-is-jensen-huangs-most</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-ai-factory-is-jensen-huangs-most</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Tue, 17 Mar 2026 10:56:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!d2CR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa324838-6eaa-45b1-8edb-92e40038f424_5120x2880.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>This essay unpacks what Jensen Huang laid out at GTC 2026 and why the implications for health tech investors and founders are almost certainly underappreciated right now. The core claim is that the shift from application-layer software to AI factory infrastructure and agent operating systems is not an incremental upgrade cycle. It is a platform extinction event for the majority of legacy SaaS business models, including a large swath of health tech. The essay covers the token economy thesis, what the Vera Rubin hardware launch and the OpenClaw phenomenon actually mean for the software stack above them, where the real moats are forming, and how healthcare specifically should be thinking about the next five years of infrastructure spend, agent deployment, and equity creation.</p><h3>Key data points referenced:</h3><p>- Computing demand increased 1 million times in the past two years per Huang</p><p>- Inference compute demand is roughly 100,000x higher than training for modern reasoning models</p><p>- NVIDIA Blackwell and Rubin lines have 500 billion dollars in orders in 2026, heading to 1 trillion in 2027</p><p>- Vera Rubin delivers 35x token throughput improvement over Hopper at equivalent power, plus another 35x via Groq LPU integration for high-value inference tiers</p><p>- OpenClaw became the most popular open-source project in human history within a few weeks of launch, surpassing Linux&#8217;s 30-year growth trajectory</p><p>- NVIDIA&#8217;s autonomous vehicle platform now covers 18 million vehicles produced annually across seven OEM partners</p><h2>Table of Contents</h2><p>Why This Keynote Is Different</p><p>The Token Economy and What It Means to Price Software</p><p>Vera Rubin, Groq, and the Hardware Stack You Need to Understand</p><p>OpenClaw Is Not a GitHub Curiosity, It Is the New OS Layer</p><p>What 80 Percent of Applications Disappearing Actually Means for Health Tech</p><p>Where the Moats Are Forming and What Founders Should Build</p><p>The Capital Allocation Question for Health Tech Investors</p><h2>Why This Keynote Is Different</h2><p>Jensen Huang has given a lot of keynotes. He is not generally known for understatement. But the GTC 2026 presentation felt different from the prior years of chip announcements dressed up in leather jacket theater, and the difference is worth dwelling on before getting into the specifics. In past years the narrative was essentially: GPUs are faster, training is cheaper, here are some demo apps. In 2026 the narrative was a full-stack worldview about what the computing paradigm itself is becoming, and it carried implications that extend well past NVIDIA&#8217;s own product line into virtually every software category that exists. For health tech specifically, where the software stack is unusually deep and unusually sticky due to regulatory complexity and clinical workflow lock-in, the implications are somewhere between extremely interesting and genuinely alarming depending on where your equity sits.</p><p>The framing Huang used was a transition from retrieval-based computing to generative computing, from data storage to token production, and from application software to intelligent agent systems. These are not marketing phrases. They describe a real architectural discontinuity that has already started playing out and that will accelerate sharply as the Vera Rubin hardware cycle hits hyperscaler and enterprise deployments through 2026 and 2027. The trillion dollar order book NVIDIA is sitting on is not speculative. It reflects capital allocation decisions that have already been made by the largest buyers of infrastructure on earth, and those buyers are not building more of what they already had. They are building something categorically different.</p><p>For investors and founders who have been operating in health tech for the last decade, the honest question is whether the companies in your portfolio or on your term sheets are positioned to thrive in the world Huang described, or whether they are positioned to be among the 80 percent of applications that disappear. That framing should be taken seriously rather than dismissed as hyperbole.</p><h2>The Token Economy and What It Means to Price Software</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-ai-factory-is-jensen-huangs-most">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The HIMSS Conference Nobody Actually Attended​​​​​​​​​​​​​​​​]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-himss-conference-nobody-actually</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-himss-conference-nobody-actually</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Wed, 11 Mar 2026 20:11:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!UI0M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70c89104-2f74-40c0-ac97-c1eb8b514ade_1920x1080.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>The conference floor at HIMSS is already kind of a simulation. Anyone who has walked the Las Vegas Convention Center in March, badge scanning through vendor booths while someone in a logo polo tries to schedule a &#8220;brief 30-minute follow-up,&#8221; knows the theater of it. The whole thing is performative sales infrastructure dressed up as knowledge transfer. Which makes the following thought experiment less crazy than it sounds: what if the humans just stopped going, and sent their AI agents instead?</p><p>This is not a piece about the metaverse. That hype cycle is mostly dead and nobody is pouring another drink for it. This is about something that is actually being built right now, across a dozen enterprise software stacks simultaneously, and the question is just whether the health tech industry is going to be intentional about it or stumble into it backwards.</p><h3>Key Arguments:</h3><p>- The primary value of conferences like HIMSS is top-of-funnel discovery and partnership exploration, not the content sessions</p><p>- AI agents are already capable of executing the semantic matching, negotiation scaffolding, and follow-up workflows that conferences are supposed to accelerate</p><p>- A new business model exists for conference operators who want to survive the transition rather than be bypassed by it</p><p>- The technical architecture is largely available today using existing LLM infrastructure, RAG systems, and multi-agent orchestration frameworks</p><p>- The human role shifts from &#8220;attendee&#8221; to &#8220;deal closer&#8221; with significant productivity implications for enterprise health tech sales teams</p><p>Data Points Referenced: HIMSS 2024 attendance (~28,000 registered), average enterprise sales cycle in health IT (12-18 months), cost per attended lead at major health tech conferences ($800-$2,000+ all-in per meeting), LLM context window capabilities relevant to negotiation tasks.</p><h2>Table of Contents</h2><p>The Conference as a Broken Sales Primitive</p><p>What AI Agents Can Actually Do Today</p><p>The Architecture of an Agent-to-Agent Conference</p><p>Business Models for the Operator and the Attendee</p><p>The Human Closes the Deal</p><p>Why HIMSS Should Build This Before Someone Else Does</p><h2>The Conference as a Broken Sales Primitive</h2><p>Here is what HIMSS actually is if you strip away the educational programming, the keynotes, the Harold Klieman Foundation awards, and the slightly depressing buffet situation: it is a very expensive, very slow, very geographically concentrated CRM event. Companies spend north of $100,000 just to have a booth presence when you factor in the exhibit space, shipping, booth build, hotel rooms, flights, meals, and the opportunity cost of pulling your best people off their actual jobs for a week. In return, they get badge scans, some conversations that were already scheduled two months in advance anyway, and a pile of business cards that someone will fail to enter into Salesforce before February is over.</p><p>The content sessions are largely a sideshow. The real work of HIMSS happens in the Starbucks line, the suite meetings at the Wynn, and the private dinners that the companies with real budgets host on Tuesday night. The conference is functioning as a temporal and geographic forcing function. It gets the right people in the same place at the same time, which is genuinely valuable, but the mechanism is preposterously inefficient relative to the outcome.</p><p>To be specific about the numbers: HIMSS 2024 drew approximately 28,000 registered attendees. Of those, a meaningful fraction are there in explicitly commercial roles, meaning they are either selling, buying, or evaluating partnerships. If you assume conservatively that the average enterprise health tech company spends $150,000 to attend, and that they walk away with 40 qualified conversations, the cost per meaningful interaction runs somewhere between $800 and $2,000 when you account for the ones that go nowhere. That is an extraordinarily high price for what is essentially a discovery mechanism, a way to find out that a company you had never heard of has a data asset or a workflow capability that maps onto a problem you are actively trying to solve.</p><p>The tragedy is not that conferences exist. The tragedy is that the discovery layer, the part where you figure out who you should be talking to and why, is done almost entirely by humans wandering around a floor plan with a printed map, hoping they bump into the right booth. In what other domain would you accept that level of signal-to-noise for that price?</p><h2>What AI Agents Can Actually Do Today</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-himss-conference-nobody-actually">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Pipes Are Finally Moving: Why Clinical Event Streaming Is the Infrastructure Bet Nobody Took Seriously Enough]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-pipes-are-finally-moving-why</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-pipes-are-finally-moving-why</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Sun, 08 Mar 2026 13:18:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!SXSK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7a6e59b-815d-4e3f-8d89-37c99621734f_1280x720.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>This essay covers the architectural shift from batch ETL to event-driven clinical data pipelines, with specific focus on EHR event streaming, real-time deterioration detection, and operational alert routing. Topics include why healthcare streaming is categorically harder than fintech equivalents, what the dominant infrastructure stack looks like, and where the actual venture opportunity sits.</p><h3>Key Points:</h3><p>- Batch ETL has been healthcare&#8217;s dominant data pattern for 30+ years and it was always wrong for clinical work</p><p>- Event-driven architectures using Kafka, Flink, and similar tooling are now viable in hospital environments</p><p>- Clinical data has 5-6 structural problems fintech doesn&#8217;t have: schema chaos, contextual validity, documentation timing lag, regulatory constraints, human-in-loop requirements, and no equivalent to atomic transactions</p><p>- Real-time deterioration detection (sepsis, respiratory failure) is the clearest clinical proof case, with measurable mortality impact</p><p>- The venture opportunity is not the streaming infrastructure itself but the clinical logic layer sitting on top of it</p><p>- Health systems have the data and the willingness; they largely lack the engineering talent to build this internally</p><h2>Table of Contents:</h2><p>The Batch ETL Hangover</p><p>What Event-Driven Actually Means in a Hospital Context</p><p>The Stack: Kafka, Flink, and the Middleware Nobody Talks About</p><p>Why This Is Not Fintech (And Fintech People Keep Getting Burned)</p><p>Deterioration Detection as the Canonical Use Case</p><p>Streaming Into Operations: The Alert Routing Problem</p><p>Where the Venture Opportunity Actually Lives</p><h2>The Batch ETL Hangover</h2><p>The honest story of healthcare data infrastructure is that it was designed around the billing cycle, not the patient. Everything in the legacy stack optimizes for the claim. Data gets captured, batched, transformed, and exported in windows that align with when someone needs to submit something to a payer or generate a report for the board. The clinical workflow was an afterthought, and the infrastructure reflected that priority ordering pretty faithfully for about three decades.</p><p>Batch ETL as a pattern made sense for a certain era. You pull data from an EHR at 2am, transform it, load it into a warehouse, and analysts run reports in the morning. That is genuinely fine for retrospective quality analysis, population health dashboards, and financial reconciliation. Nobody needs real-time data to figure out how the readmission rate trended in Q3. The problem is that a lot of healthcare use cases are not retrospective. A patient who is septic at 11pm cannot wait for the 2am batch. A nurse who needs to know a patient&#8217;s current medication list before administering something cannot work from yesterday&#8217;s extract. The infrastructure was fundamentally mismatched with the urgency of clinical reality, and the industry just kind of tolerated that mismatch for a long time because building something better was genuinely hard and the incumbents had no financial incentive to do it.</p><p>What changed is a combination of three things happening more or less simultaneously. Epic and the other major EHR vendors started building real API surfaces, particularly after CMS started mandating interoperability. Cloud infrastructure got cheap enough that health systems could actually consider running Kafka clusters without having to justify a seven-figure capex line item. And a generation of engineers who had built streaming systems in fintech and adtech started showing up in healthcare companies with a reasonable question: why is this so much worse than what we built at Stripe? Those three forces together created the conditions for the current architectural transition.</p><p>The transition is real and it is happening, but it is not happening uniformly. What you see in the market right now is a fairly wide spectrum ranging from academic medical centers with genuinely sophisticated real-time pipelines down to rural critical access hospitals still faxing things. The opportunity space lives in that gap, and understanding the architecture is table stakes for understanding where the real leverage points are.</p><h2>What Event-Driven Actually Means in a Hospital Context</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-pipes-are-finally-moving-why">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The AI clinical infrastructure company: why the real money in Health AI isn’t in the models]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-ai-clinical-infrastructure-company</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-ai-clinical-infrastructure-company</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Sun, 22 Feb 2026 13:01:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Wr7p!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7280dcad-05ec-4956-97c3-9faecb031e7a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>This essay makes the case that the most durable and defensible business in health AI over the next decade isn&#8217;t building foundation models &#8211; it&#8217;s building the deployment, governance, and validation infrastructure that makes those models safe to use in clinical settings. The argument draws on the collective action problem facing health systems, the lessons of enterprise infrastructure companies like Azure, and the structural advantages of coalition-based data networks in regulated industries.</p><h3>Key claims:</h3><p>- Foundation model competition is already a race to the bottom among the world&#8217;s best-capitalized companies. Health tech entrepreneurs don&#8217;t need to be in that race.</p><p>- Health systems lack the internal capability to deploy AI safely and at scale. This isn&#8217;t a gap a few vendors are going to close &#8211; it&#8217;s a systemic infrastructure deficit.</p><p>- The clinical deployment and governance layer is the Azure of health AI. It sits between the model and the bedside and does the unglamorous, high-margin work of making AI actually work.</p><p>- Coalition-based deployment networks create compounding data and validation advantages that point solutions can&#8217;t replicate.</p><p>- This is a venture-backable company with a realistic path to $500M+ ARR.</p><h2>Table of Contents</h2><p>The Setup: Why Health Systems Can&#8217;t Do This Themselves</p><p>The Mistake Everyone&#8217;s Making: Betting on Models</p><p>What &#8220;Clinical Infrastructure&#8221; Actually Means</p><p>The Coalition Play: Network Effects in a Regulated Industry</p><p>The Business Model and Why It Works</p><p>What Could Kill This</p><p>The Investment Case</p><h2>The Setup: Why Health Systems Can&#8217;t Do This Themselves</h2><p>Start with a number: there are roughly 6,000 hospitals in the United States, operating across about 900 health systems of meaningful size. Every single one of them is going to need to figure out AI over the next decade. Not because they want to, necessarily, but because the economics of healthcare delivery are going to force their hand. Labor is the biggest cost driver in hospital operations &#8211; somewhere north of 55-60% of total expenses &#8211; and AI is the only plausible lever that bends that curve without destroying quality. So this isn&#8217;t optional. The question is execution.</p><p>Here&#8217;s where it gets complicated. Deploying AI in a clinical environment is not like deploying AI in a SaaS product or a financial services workflow. The regulatory surface area alone is enormous. FDA has started asserting jurisdiction over clinical decision support software in ways that are still being litigated in real time. HIPAA creates data handling requirements that most generic infrastructure can&#8217;t meet out of the box. And beyond compliance, there&#8217;s the clinical validation problem, which is its own category of hard. A model that performs well on training data and even on test data from an academic medical center can fail in surprising ways when it hits the actual workflow of a community hospital in rural Tennessee with different patient demographics, different EHR configurations, and different clinical protocols. The history of health tech is littered with products that worked in the demo and died in deployment.</p><p>Health systems know this. Ask any CIO at a regional health system what keeps them up at night and AI governance is going to be on the list, right next to cybersecurity and Epic upgrade cycles. They&#8217;re getting pitched constantly by AI vendors and they&#8217;re sophisticated enough to know that the pitch is usually ahead of the reality. The honest ones will tell you they don&#8217;t have the internal capability to evaluate AI tools rigorously, much less build the deployment infrastructure to operationalize them safely. A mid-sized health system might have a dozen data scientists and maybe one or two people with any real ML background. That&#8217;s not enough to build model fine-tuning pipelines, clinical validation frameworks, governance documentation, change management playbooks, and the monitoring infrastructure to catch model drift over time. It&#8217;s especially not enough to build all of that and then maintain it across dozens of use cases simultaneously.</p><p>So you have a demand signal that is clear and growing, an internal capability gap that is not going to close anytime soon, and a vendor landscape that is mostly offering models rather than infrastructure. That&#8217;s the setup.</p><h2>The Mistake Everyone&#8217;s Making: Betting on Models</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-ai-clinical-infrastructure-company">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The $145M Federal Subsidy Nobody in Health Tech Is Talking About Yet]]></title><description><![CDATA[Table of Contents]]></description><link>https://www.onhealthcare.tech/p/the-145m-federal-subsidy-nobody-in</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-145m-federal-subsidy-nobody-in</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Tue, 17 Feb 2026 15:29:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!zFe7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c040e45-ffbd-4e6f-8982-5ec194303a5a_1290x1726.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Table of Contents</h2><p>Abstract</p><p>The Setup: Why the Healthcare Workforce Is a Burning Platform</p><p>What DOL Actually Announced (and Why the Details Are Weird in a Good Way)</p><p>The Playbook: How Health Tech Entrepreneurs Can Plug In</p><p>The Startup Angles Nobody Is Pitching Yet</p><p>The Fly in the Ointment: Execution Risk in Federal Programs</p><p>Where This Is All Going</p><p>-----</p><h2>Abstract</h2><p>- DOL announced a $145M Pay-for-Performance Incentive Payments Program on February 13, 2026, with applications due April 3, 2026</p><p>- Program issues up to 5 cooperative agreements over 4 years; individual agreements range from $10M to $40M</p><p>- Healthcare is a named priority sector alongside AI/semiconductor, shipbuilding/defense, IT, transportation, and telecom</p><p>- U.S. faces projected shortage of 3.2 million healthcare workers by 2026, 141,160 physicians by 2038, and 73,000+ nursing assistants by 2028</p><p>- Unlike traditional grants, this is a pay-for-performance model where money flows per enrolled apprentice, not as upfront allocation</p><p>- This structure creates at least 4 distinct categories of commercial opportunity for health tech entrepreneurs</p><p>- Eligible applicants include state agencies, national industry associations, labor-management orgs, workforce intermediaries, consulting orgs, and consortia</p><p>- Required to include at least one national or multi-regional industry association as a partner</p><p>- Deadline is April 3, 2026, which is tight but workable for a well-networked team</p><p>-----</p><h2>The Setup: Why the Healthcare Workforce Is a Burning Platform</h2><p>There is a version of this story that has been told so many times that investors and operators have sort of tuned it out. Healthcare workforce shortage, aging population, burnout, nursing schools rejecting 92,000 qualified applicants in 2021 alone because there are not enough faculty and classroom seats, hospitals spending more than 50% of their total operating budget on labor. Right, we know, everyone has a deck about it. But it is worth spending a moment actually confronting the magnitude of what is happening before getting to the commercial opportunity, because the numbers have gotten genuinely alarming in a way that shifts what solutions the market will actually pay for.</p><p>One analysis of EMSI data projects a critical shortage of 3.2 million healthcare workers by 2026, and that is not a point far in the future anymore, that is now. HRSA projects an overall physician shortage of 141,160 FTEs by 2038 and already estimates that the physician workforce in 2026 will only meet 90% of demand nationally, a figure that gets substantially worse in rural areas where some specialties face shortages approaching 46%. The American College of Physicians is projecting a shortage of 85,000 physicians by 2036. Nursing is its own disaster: more than 100,000 nurses left the workforce in recent years, about 35% of the physician workforce is projected to hit retirement age within five years, and per a Harris Poll survey conducted mid-2025, 55% of healthcare employees say they intend to search for, interview for, or switch jobs in 2026. That last number is almost too large to be believed, but the underlying drivers are real: 84% report feeling underappreciated, and only 1 in 5 feel their employer is invested in their long-term career growth.</p><p>What this creates at a macro level is a market where health systems are both financially desperate (labor is north of 50% of operating costs and rising) and operationally desperate (there simply are not enough trained bodies to fill open roles). That combination, expensive and scarce, is exactly the condition that makes a market receptive to structural innovation rather than marginal improvement. Systems are not looking for a 10% efficiency gain on top of a broken model anymore. They are looking for ways to fundamentally change how they build, retain, and deploy clinical and clinical-adjacent staff. That is the context into which the DOL just dropped $145 million and told the market to figure out how to use it.</p><p>-----</p><h2>What DOL Actually Announced (and Why the Details Are Weird in a Good Way)</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-145m-federal-subsidy-nobody-in">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Chatbot in the Courtroom: What U.S. v. Heppner Means for Health Tech Founders Who Use AI to Think Through Legal Problems]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-chatbot-in-the-courtroom-what</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-chatbot-in-the-courtroom-what</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Sun, 15 Feb 2026 14:17:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wMrd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F473f159e-c4c9-491c-a436-7672f1715394_1290x756.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>The February 10, 2026 bench ruling in U.S. v. Heppner (No. 25-cr-00503-JSR, S.D.N.Y.) is the first major federal decision holding that a client&#8217;s use of a consumer AI tool to prepare defense strategy documents waived attorney-client privilege. Judge Jed Rakoff&#8217;s reasoning rested on three pillars: (1) AI tools are not attorneys, (2) consumer ToS explicitly disclaim confidentiality and permit training on inputs and disclosure to gov&#8217;t authorities, and (3) pre-existing non-privileged documents don&#8217;t become privileged just by being sent to a lawyer after the fact. This matters disproportionately to health tech founders and executives who routinely use consumer AI to draft, summarize, or analyze content that touches on FDA regulatory submissions, HIPAA compliance memos, investor deal terms, IP licensing strategy, employment disputes, board communications, and reimbursement appeals. The ruling does not automatically doom enterprise AI use &#8211; courts are likely to treat tools that contractually prohibit training on inputs and maintain data confidentiality differently &#8211; but the line between &#8220;safe&#8221; and &#8220;not safe&#8221; is less obvious than most people think.</p><h3>Key takeaways:</h3><p>- Consumer AI (ChatGPT free tier, [Claude.ai](http://Claude.ai) free/Pro) = no privilege protection per Heppner</p><p>- Enterprise/API tiers with zero-data-retention = likely safer, not yet definitively tested in court</p><p>- AI note-takers on calls with counsel = almost certainly problematic</p><p>- Work product doctrine is a separate shield, and it also failed in Heppner</p><p>- The three-part test for privilege: attorney, confidentiality, legal-advice purpose &#8211; consumer AI fails all three</p><p>- Heppner had Quinn Emanuel and still lost &#8211; this isn&#8217;t just a DIY legal-research problem</p><p>- Health tech-specific exposure: HIPAA enforcement, FDA 483s, SEC/FTC investigations, M&amp;A due diligence, cap table disputes</p><h2>Table of Contents</h2><p>What Actually Happened in Heppner</p><p>The Three Ways Privilege Dies</p><p>Work Product Doctrine: The Second Line of Defense That Also Collapsed</p><p>Why Health Tech Is More Exposed Than Most Sectors</p><p>The Enterprise Carve-Out: Real Safety or Just a Better Story to Tell Your Board</p><p>The AI Note-Taker Problem Nobody Is Talking About</p><p>Practical Protocol for Founders and Executives Right Now</p><h2>What Actually Happened in Heppner</h2><p>Before getting into implications, it helps to actually understand the facts, because a lot of the commentary floating around gets them a little wrong in ways that matter.</p><p>Bradley Heppner was the former CEO of Beneficient, an alternative finance company in Dallas. He got indicted in October 2025 on securities fraud, wire fraud, conspiracy, and making false statements to auditors in connection with an alleged scheme that prosecutors say cost GWG Holdings investors roughly $1 billion &#8211; GWG filed for bankruptcy, and prosecutors allege Heppner extracted over $150 million for himself before that happened, including $40 million to renovate his Dallas mansion. He was represented by Quinn Emanuel, which is as serious a white-collar defense firm as you can hire.</p><p>Here&#8217;s the key sequence. After Heppner knew he was a law enforcement target and had retained counsel, he used the consumer version of a commercial AI tool to run queries related to the government&#8217;s investigation. He fed information he&#8217;d learned from his Quinn Emanuel attorneys into the AI, generated 31 documents of prompts and AI responses, and then transmitted those documents to his lawyers. He apparently thought this would help him organize his thinking and prepare for the case. When the FBI executed a search warrant on his home, they seized devices containing those documents. His legal team asserted both attorney-client privilege and work product protection. The government moved to compel production. Judge Rakoff ruled from the bench on Feb. 10, 2026 that neither doctrine protected the documents, saying &#8220;I&#8217;m not seeing remotely any basis for any claim of attorney-client privilege.&#8221;</p><p>A few things worth flagging that people tend to glide past. First, Heppner wasn&#8217;t some naive founder who didn&#8217;t know he was in legal jeopardy &#8211; he was under active investigation, knew it, and had already engaged counsel. Second, he had actually received substantive legal strategy input from Quinn Emanuel attorneys and then incorporated that information into his AI queries. Third, his lawyers argued strenuously that the documents should be protected, and they didn&#8217;t win. If you&#8217;re imagining that &#8220;well, I was just using AI to prep some notes before calling my lawyer&#8221; puts you in a different category than Heppner, think that through more carefully.</p><h2>The Three Ways Privilege Dies</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-chatbot-in-the-courtroom-what">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[World models walk into a hospital: why this time it actually matters ￼]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/world-models-walk-into-a-hospital</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/world-models-walk-into-a-hospital</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Mon, 09 Feb 2026 12:53:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ZSzB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0db29db0-20a7-4150-a0cc-47e6ecaa6635_1290x733.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>This essay examines how world models fundamentally alter healthcare software architecture beyond surface-level demos. Written for investors and operators who have already navigated rules engines, feature stores, deep learning cycles, and previous AI revolutions, the analysis focuses on why world models represent a genuine inflection point rather than another overhyped technology wave.</p><h3>Key themes:</h3><p>- Core mechanics of world models stripped of marketing terminology</p><p>- Why healthcare represents both worst case environment and highest value opportunity</p><p>- How representation learning, latent state prediction, and planning transform clinical, operational, and financial workflows</p><p>- Where genuine venture scale opportunities exist versus dead ends</p><p>- Why current healthcare AI architectures will age poorly in world model driven future</p><p>Primary technical framing draws from recent world modeling research and workshops, including work presented at the World Model Workshop at MILA in 2026.</p><h2>Table of Contents</h2><p>Introduction: Healthcare Is the Messy Real World</p><p>What a World Model Actually Is and Is Not</p><p>Why Healthcare Breaks Pattern Recognition</p><p>From Prediction to Counterfactuals</p><p>Clinical Care as Partially Observable Control</p><p>Operational Healthcare Is the Sleeper Use Case</p><p>Why Generative Models Alone Stall Out</p><p>Joint Embedding, Latent State, and Why Abstraction Wins</p><p>Memory, Time, and the Curse of Long Horizons</p><p>Safety, Guardrails, and Why Healthcare Forces the Issue</p><p>Implications for Founders</p><p>Implications for Investors</p><p>Where This All Probably Breaks</p><p>Closing Thoughts</p><h2>Introduction: Healthcare Is the Messy Real World</h2><p>Healthcare has always been where software optimism goes to die. Data arrives missing, mislabeled, delayed, or legally cordoned off. Outcomes are fundamentally stochastic. Interventions interact in ways nobody fully understands. Feedback loops stretch across months or years, and running controlled experiments often crosses ethical boundaries. The environment remains partially observable, non-stationary, subtly adversarial, and regulated by organizations that still rely on fax machines. Healthcare looks nothing like the benchmark datasets that made the last decade of machine learning feel tractable.</p><p>This is precisely why world models matter more in healthcare than anywhere else. Not because they magically solve healthcare&#8217;s problems, but because they&#8217;re explicitly designed for environments where perception remains incomplete, dynamics matter, and actions reshape future observations. Pattern recognition systems assume the world is static and fully visible. Healthcare violates both assumptions. A hospital resembles a real time strategy game under fog of war far more than it resembles ImageNet.</p><p>Most health tech AI today still operates in System 1 territory. Take an input, map it to an output, maybe calibrate the probability, hope the environment doesn&#8217;t drift too quickly. World models push the stack into System 2 thinking. They force systems to internalize how the world evolves, not just what it looks like right now. This distinction matters more in healthcare than in almost any other domain because healthcare decisions inherently require reasoning about sequences, not snapshots.</p><h2>What a World Model Actually Is and Is Not</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/world-models-walk-into-a-hospital">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Labor Reallocation Problem: Why Healthcare Productivity Is a Structural GDP Issue and How Task Decomposition Plus Robotics Could Actually Fix It]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-labor-reallocation-problem-why</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-labor-reallocation-problem-why</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Sat, 07 Feb 2026 11:56:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Wr7p!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7280dcad-05ec-4956-97c3-9faecb031e7a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>Healthcare consistently underperforms every other sector in productivity growth, creating a structural drag on GDP that compounds annually. This essay examines why healthcare labor efficiency matters for macroeconomic performance and explores two complementary technological paths that could reverse decades of stagnation: cognitive task decomposition through AI and physical labor substitution via hospital robotics. The argument centers on labor as the dominant GDP driver, healthcare&#8217;s disproportionate and growing share of both employment and economic output, and the specific mechanisms by which current clinical workflows waste expensive human capital. Evidence suggests nursing labor alone represents 25 to 45 percent of hospital operating budgets while physicians spend 35 to 49 percent of working hours on administrative tasks that generate zero clinical value. The essay evaluates emerging robotics platforms for hospital logistics, patient handling, and environmental services alongside AI systems for documentation, triage, and clinical decision support, arguing that meaningful productivity gains require simultaneous decomposition of both cognitive and physical nursing work rather than incremental automation of isolated tasks.</p><h2>Table of Contents</h2><p>The GDP Framework Nobody Actually Uses But Should</p><p>Why Labor Quality and Quantity Dominate Everything Else</p><p>Healthcare as a Labor Sink That Shows Up in National Accounts</p><p>The Clinical Labor Waste Taxonomy</p><p>Cognitive Task Decomposition as Productivity Infrastructure</p><p>Physical Nursing Labor and the Robot Question</p><p>Hospital Robotics: Current State and Economic Viability</p><p>Why This Matters Beyond Hospital Margins</p><p>Implementation Barriers That Actually Matter</p><p>Second Order Effects on Labor Markets and Training Systems</p><h2>The GDP Framework Nobody Actually Uses But Should</h2><p>Most discussions about economic growth devolve into hand waving about innovation or appeals to vague cultural factors. The productive framework starts with the identity that GDP equals productive capacity times utilization times prices, then works backward to isolate variables that actually move the needle. Energy infrastructure matters for obvious reasons. Transportation networks determine whether goods reach markets. Communication systems enable coordination at scale. But these are enablers, not prime movers. The structural variables that determine output per unit time are narrower and more mechanical than most people want to admit.</p><p>The algebra gets interesting when you disaggregate by sector and realize that not all labor hours contribute equally to measured output. Healthcare represents roughly 18 percent of US GDP and employs about 16 million people, making it the largest employment sector. Yet healthcare productivity growth has been essentially flat or negative by most measures over the past 40 years, even as other service sectors posted consistent gains. This creates a compositional drag where an increasing share of the labor force moves into a sector with stagnant output per worker, pulling down aggregate productivity growth regardless of what happens in manufacturing or tech.</p><p>The Baumol cost disease explanation holds that sectors with low productivity growth experience rising relative prices if wages are set in a competitive labor market. Healthcare exhibits this perfectly. Real spending per capita has grown at 4 to 5 percent annually while measurable health outcomes improved far more slowly. Some of this reflects genuine quality improvements that evade measurement, but much of it stems from structural inefficiency in how clinical labor gets allocated.</p><h2>Why Labor Quality and Quantity Dominate Everything Else</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-labor-reallocation-problem-why">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Deepgram’s healthcare gambit: when the best voice AI isn’t built for healthcare]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/deepgrams-healthcare-gambit-when</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/deepgrams-healthcare-gambit-when</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Wed, 04 Feb 2026 18:05:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WJ5U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F410041f0-4b9a-4f5f-988e-1bdda727f082_1289x1625.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>This essay examines Deepgram&#8217;s $143.2 million Series C raise at a $1.2 billion pre-money valuation and analyzes their competitive positioning in healthcare voice AI. The company faces an apparent paradox: attempting to win in healthcare while maintaining a horizontal platform approach across multiple industries. This analysis explores what differentiates Deepgram&#8217;s technical architecture from healthcare-specific competitors, identifies their primary use cases and competitive landscape, and evaluates whether a platform-first strategy can succeed in a vertical known for punishing generalists. The healthcare voice AI market has become extraordinarily crowded, with dozens of well-funded startups promising to solve clinical documentation burden, yet Deepgram&#8217;s approach suggests a different thesis about where durable value accrues in this market.</p><h2>Table of Contents</h2><p>- The Crowded Healthcare Voice AI Landscape</p><p>- Deepgram&#8217;s Technical Architecture and Differentiation</p><p>- Primary Healthcare Use Cases and Customer Profiles</p><p>- Competitive Analysis and Market Positioning</p><p>- The Platform Versus Vertical Debate</p><p>- Economics and Unit Economics Considerations</p><p>- Risk Factors and Path to Dominance</p><h2>The Crowded Healthcare Voice AI Landscape</h2><p>The healthcare voice AI market has reached a saturation point that would make any rational investor queasy. Between 2022 and 2024, venture funding poured into companies promising to eliminate clinical documentation burden, with nearly every pitch deck featuring the same statistic about physicians spending two hours on EHR documentation for every hour of patient care. The result has been a Cambrian explosion of startups, each claiming their particular approach to speech recognition, natural language processing, or ambient clinical documentation represents a meaningful breakthrough.</p><p>Current estimates suggest there are somewhere between 40 and 60 companies actively building voice AI solutions targeting healthcare, depending on how broadly one defines the category. On the clinical documentation side alone, companies like Nuance (acquired by Microsoft for $19.7 billion), Suki, Abridge, Nabla, DeepScribe, Freed, Augmedix, and Notable have all raised substantial capital. Then there are the contact center plays like PolyAI, Parlance, and various others trying to automate patient scheduling and triage. Add in the EHR vendors building native ambient documentation features, the revenue cycle management companies bolting on voice capabilities, and the large language model providers like OpenAI positioning Whisper as healthcare-ready, and the market starts to look less like an opportunity and more like a bloodbath waiting to happen.</p><p>What makes this crowding particularly problematic is that many of these solutions appear functionally similar to the average buyer. A health system CIO evaluating ambient documentation tools will see demos from five vendors that all show a physician having a natural conversation with a patient, with the system magically generating a structured SOAP note that populates directly into Epic or Cerner. The value proposition sounds identical, the workflows look comparable, and the accuracy metrics all claim to exceed 95 percent. Differentiation becomes a game of minor feature differences, integration depth, and price competition rather than fundamental technological superiority.</p><p>This commoditization risk is real and already manifesting in pricing pressure. Early ambient documentation deals in 2021 and 2022 were commanding $200 to $400 per provider per month. By 2024, prices had compressed to $100 to $150 per provider per month for many vendors, with some aggressive new entrants offering pilot programs at near-cost to gain reference customers. The gross margins that looked attractive at $300 per month start looking significantly less compelling at $120 per month, particularly when factoring in the compute costs for inference, the customer success expenses for a notoriously high-touch healthcare market, and the integration maintenance burden across dozens of EHR instances.</p><p>The crowding also creates distribution challenges that favor incumbents with existing healthcare relationships. Nuance, now part of Microsoft, has decades-long relationships with nearly every major health system in the United States. Their Dragon Medical suite has been the standard for physician dictation since before most current healthcare AI founders were born. When Microsoft decided to integrate Nuance&#8217;s DAX technology directly into Teams and position it as the default ambient documentation solution for their massive installed base of healthcare customers, they effectively created a moat that would be extraordinarily expensive for any startup to overcome. Similarly, Epic&#8217;s decision to build native ambient documentation features using a combination of their own models and partnerships with multiple AI vendors means that health systems can access voice AI capabilities without adding a new vendor, negotiating a new contract, or managing another integration.</p><p>Deepgram enters this market with $143.2 million in fresh capital and a $1.2 billion valuation, yet their approach appears fundamentally different from the healthcare-specific competitors. Rather than building a vertical solution focused exclusively on clinical documentation or patient engagement, Deepgram has positioned itself as a horizontal speech AI platform serving multiple industries. Their customer base includes companies in financial services, media, customer service, and sales alongside whatever healthcare presence they have built. This creates an interesting strategic question: can a company that is not exclusively focused on healthcare possibly be focused enough to win against vertical specialists who eat, sleep, and breathe HIPAA compliance and clinical workflows?</p><h2>Deepgram&#8217;s Technical Architecture and Differentiation</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/deepgrams-healthcare-gambit-when">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Interface Wars: Why Apple Spent Two Billion Dollars on Mind Reading Technology and What It Means for Healthcare AI]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-interface-wars-why-apple-spent</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-interface-wars-why-apple-spent</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Sun, 01 Feb 2026 12:50:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Wr7p!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7280dcad-05ec-4956-97c3-9faecb031e7a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>Apple&#8217;s acquisition of Q.ai for approximately two billion dollars represents more than another big tech purchase - it signals the next major interface revolution in computing. Q.ai&#8217;s technology reads facial micro-movements to detect silent speech, enabling communication with AI systems without vocalization. This follows the rapid adoption of ambient voice documentation in healthcare, where companies like Abridge, Nuance, and others have fundamentally changed clinical workflows. The pattern is clear: the companies winning in AI aren&#8217;t necessarily building better models but rather solving the interface problem. This essay examines how interface innovation drives adoption in healthcare technology, why voice was just the beginning, what comes next in the evolution of human-computer interaction for clinical settings, and why the ultimate interface breakthrough won&#8217;t be about language at all but about capturing the full sensory and emotional bandwidth of human experience. The stakes are massive - interface shifts create winner-take-all markets, and healthcare represents the most complex, highest-value use case for the next generation of AI interaction paradigms.</p><h2>Table of Contents</h2><p>The Two Billion Dollar Bet on Reading Your Face</p><p>Why Healthcare Became the Proving Ground for Voice AI</p><p>The Ambient Documentation Market Explosion</p><p>Interface Physics: Why Voice Beat Typing and What Beats Voice</p><p>Beyond Voice: The Next Wave of Clinical Interaction Models</p><p>The Language Trap: Why Words Are Just the Beginning</p><p>The Brain Interface as Deep Tech Holy Grail</p><p>Why Sensory Bandwidth Matters More Than Linguistic Precision</p><p>The Apple Healthcare Strategy Nobody Talks About</p><p>What This Means for Healthcare AI Investing</p><h2>The Two Billion Dollar Bet on Reading Your Face</h2><p>Apple just put down roughly two billion dollars for Q.ai, an Israeli company most people have never heard of. The tech sounds like science fiction - they can read tiny facial movements to detect what you&#8217;re trying to say without you actually speaking. Silent speech recognition through computer vision. This is not vaporware or some distant R and D project. The technology works now, and Apple clearly believes it works well enough to write a check that makes this their second-largest acquisition in history.</p><p>Context matters here. Q.ai&#8217;s founder Aviad Maizels previously sold PrimeSense to Apple back in 2013. That became Face ID, the technology millions of people use dozens of times per day without thinking about it. Apple doesn&#8217;t buy companies at random. They acquire specific technical capabilities to solve specific product roadmap problems, then spend years integrating and shipping. The PrimeSense acquisition took several years to ship as Face ID. This pattern suggests Q.ai&#8217;s tech probably won&#8217;t show up in products next quarter, but when it does ship, it will be polished and integrated into something people actually want to use.</p><p>The timing is fascinating. OpenAI is reportedly building AirPods competitors. Google has been iterating on Pixel Buds with better AI integration. Meta continues dumping money into Reality Labs building interfaces for a metaverse that may or may not materialize. Amazon built Alexa into everything with a speaker. Every major tech company is racing to own the primary interface between humans and AI systems, because they understand something critical - the intelligence itself is increasingly commoditized, but the interface creates lock-in and determines who captures value.</p><p>Think about what happened with smartphones. The intelligence moved to the cloud pretty quickly. What mattered was who controlled the interface layer - iOS and Android. Everything else became middleware. The same dynamic is playing out with AI, except the interface war is happening faster and with higher stakes because AI creates more value per interaction than mobile apps ever did.</p><p>Healthcare has been the early proving ground for this interface revolution, specifically voice-to-clinical-documentation. The results have been dramatic enough that they offer a roadmap for what happens next across all of computing.</p><h2>Why Healthcare Became the Proving Ground for Voice AI</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-interface-wars-why-apple-spent">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The API is the Scalpel: A Business Plan for a Multimodal Health Data Layer]]></title><description><![CDATA[Abstract]]></description><link>https://www.onhealthcare.tech/p/the-api-is-the-scalpel-a-business</link><guid isPermaLink="false">https://www.onhealthcare.tech/p/the-api-is-the-scalpel-a-business</guid><dc:creator><![CDATA[Special Interest Media]]></dc:creator><pubDate>Fri, 30 Jan 2026 12:56:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!L3ed!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e7480e5-1e79-43e9-bbe6-d8c43432b5a3_935x426.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>This document outlines the business plan for a new venture: an API-first healthcare data infrastructure company. The company will provide a developer-centric platform to solve the pervasive problem of multimodal data integration in healthcare. By offering a suite of APIs, we will enable health tech companies, research institutions, and providers to seamlessly ingest, harmonize, and fuse disparate data types including imaging, clinical notes, time-series data, and tabular records. Our core technology leverages state-of-the-art machine learning techniques for data pre-processing, feature extraction, and fusion, abstracting away the immense complexity and computational cost that currently stifles innovation. The business model is a usage-based API subscription, creating a scalable, recurring revenue stream. This plan details the market opportunity, the technical solution, go-to-market strategy, and financial projections, making a case for investment in what we believe will become the foundational data layer for the next generation of healthcare innovation.</p><h3>Key elements of the plan:</h3><p>- Market opportunity: 50B+ healthcare analytics market, with multimodal integration as a foundational requirement</p><p>- Product: Developer-first API platform for ingesting, processing, and fusing healthcare data across modalities</p><p>- Business model: Usage-based pricing with free, standard, and enterprise tiers</p><p>- Go-to-market: Phased approach targeting startups first, then academia, then enterprise</p><p>- Competitive advantage: Domain-specific, API-first approach vs generic cloud tools or closed platforms</p><p>- Financial model: 70-80 percent gross margins at scale, path to profitability in 3-4 years</p><h2>Table of Contents</h2><p>1. Introduction: The Great Data Traffic Jam</p><p>2. The Core Problem: Multimodal Mayhem</p><p>3. The Solution: An API-First Data Fusion Engine</p><p>4. How It Works: A Peek Under the Hood</p><p>5. Go-to-Market: Who Needs This Yesterday</p><p>6. The Business Model: It's All About the API Calls</p><p>7. The Competitive Landscape: Why We Win</p><p>8. Risk Factors and Mitigation</p><p>9. Conclusion: The Future is Fused</p><h2>Introduction: The Great Data Traffic Jam</h2><p>Anyone who has spent more than a week in health tech knows the grand paradox of our industry. We are swimming, practically drowning, in a tsunami of data. Electronic health records, genomic sequences, DICOM images, continuous streams from wearables, and gigabytes of clinical notes are being generated at a pace that makes Moore's Law look quaint. Yet, for all this raw data, the industry remains information-starved. It is a colossal traffic jam where everyone has a car, but no one has a paved road to drive on. The promise of AI and personalized medicine feels perpetually just around the corner, perpetually held back by the mundane, brutal reality of data fragmentation. Every ambitious startup, every innovative hospital research wing, every pharmaceutical company trying to accelerate clinical trials slams into the same wall. Their data is a mess. It lives in a dozen different formats, in a hundred different silos, each speaking a unique and belligerent dialect. The result is a tragic waste of resources, as brilliant engineers and data scientists spend the vast majority of their time not on building breakthrough models, but on the digital equivalent of janitorial work: cleaning, mapping, and attempting to stitch together data that was never designed to coexist. This is not a problem of a single missing application or a single bad actor. It is a fundamental, infrastructural deficit. The industry lacks the foundational plumbing required to make its own data useful. And in that deficit lies an enormous opportunity.</p><p>The numbers tell the story. England alone performed over 43 million X-rays in 2022. Each one of those images is a data point, but without the accompanying clinical context from text notes, lab values, and patient history, it is just pixels on a screen. The landscape is littered with examples of this fragmentation. Studies on Alzheimer's disease prediction, for instance, typically work with datasets ranging from just a few dozen to maybe a couple thousand patients, not because larger cohorts do not exist, but because assembling and harmonizing multimodal data across institutions is so prohibitively difficult. Cancer prediction studies fare slightly better, with some datasets reaching over 10,000 patients, but even these represent years of painstaking manual data curation. The opportunity cost is staggering. How many breakthrough diagnostic tools have not been built because the team could not get past the data integration hurdle? How many clinical trials have been delayed or abandoned because the data infrastructure could not keep up? This is the problem we are solving, and it is a problem that touches every corner of the healthcare industry.</p><h2>The Core Problem: Multimodal Mayhem</h2>
      <p>
          <a href="https://www.onhealthcare.tech/p/the-api-is-the-scalpel-a-business">
              Read more
          </a>
      </p>
   ]]></content:encoded></item></channel></rss>