Thoughts on Healthcare Markets and Technology

Thoughts on Healthcare Markets and Technology

GPT-Rosalind Lands: What OpenAI’s First Domain-Specific Life Sciences Model, the Codex Life Sciences Plugin & the Trusted Access Program Actually Mean

Apr 18, 2026
∙ Paid

Table of Contents

What actually shipped on April 16

Benchmarks, with the appropriate skepticism

The plugin is the real story

Trusted access, biosecurity, and why the gate matters

Where this lands inside the pharma stack

What gets compressed, what gets created, what gets killed

Read-throughs for tools, data, and services companies

Angel and seed-stage implications

The quiet part about moats

Caveats, open questions, and the boring stuff nobody wants to talk about

So what now

Abstract

- OpenAI shipped GPT-Rosalind on April 16, 2026, its first purpose-built domain model, aimed at biochemistry, genomics, and protein engineering

- Access is gated via a trusted-access program; launch partners include Amgen, Moderna, Thermo Fisher Scientific, Allen Institute, plus a Los Alamos collab on protein and catalyst design

- Claimed benchmark results: 0.751 pass rate on BixBench, beats GPT-5.4 on 6 of 11 LABBench2 tasks, and in a Dyno Therapeutics eval on unpublished RNA sequences, best-of-10 submissions cleared the 95th percentile of human experts on sequence-to-function prediction and roughly 84th percentile on sequence generation

- A Life Sciences research plugin for Codex connects the model to 50+ scientific tools and public bio databases, which is arguably more commercially important than the model weights themselves

- Preview phase does not consume tokens or credits for approved orgs, meaning the effective price is zero for the enterprise tier, which will distort willingness-to-pay data across the entire biotech software market for roughly 6 to 12 months

- Read-through for founders: data-access wrappers, lit-review tools, and protocol-design copilots with no proprietary data are now at existential risk; differentiated wet-lab data, closed-loop experimentation, regulated workflows, and vertical systems of record are relatively safer

- Read-through for angels: pause any check into a pure RAG-over-PubMed startup, underwrite biotech software against a post-Rosalind baseline rather than a GPT-4-era baseline, and lean into companies producing new, non-public scientific data

- Caveats: dual-use risk is non-trivial, no fully AI-discovered drug has cleared phase 3, and OpenAI’s benchmark numbers are self-reported against evals where OpenAI had training-time knowledge of the tasks

What actually shipped on April 16

OpenAI pushed out three things in a single announcement, and the health tech crowd keeps conflating them. Separating the three is how the analysis gets interesting.

The first is GPT-Rosalind itself, a frontier reasoning model in a new Life Sciences series. It is designed to support evidence synthesis, hypothesis generation, experimental planning, and multi-step scientific workflows across biochemistry, genomics, and protein engineering.  Named after Rosalind Franklin, which is a nice bit of historical housekeeping given the Nobel committee’s 1962 miss. The model is available in ChatGPT, Codex, and the API, but you cannot just sign up for it.

The second is the gating layer, which OpenAI calls the trusted access program. Eligibility is restricted to qualified enterprise customers in the US working on health-relevant research, with governance and safety oversight controls in place. Launch partners named publicly are Amgen, Moderna, Thermo Fisher Scientific, and the Allen Institute, plus an existing collaboration with Los Alamos National Laboratory on AI-guided protein and catalyst design. During the preview, usage does not consume existing credits or tokens for approved orgs, subject to abuse guardrails. The pricing part is worth staring at for a minute. OpenAI is effectively giving the model away to pharma at the moment, which is a fairly aggressive land grab and is going to wreck price discovery for every startup trying to sell AI-for-biotech software to the same buyers.

The third is the Life Sciences research plugin for Codex, published to GitHub. The plugin connects models to over 50 scientific tools and data sources,  including human genetics, functional genomics, protein structure, and clinical evidence data. Quietly, OpenAI said it is also making the connectors and the plugin more broadly available for use with mainline models, not just Rosalind. That matters more than the model itself. More on that below.

Benchmarks, with the appropriate skepticism

The benchmark numbers are noteworthy but worth reading carefully, because every model vendor publishes whatever makes their thing look good.

User's avatar

Continue reading this post for free, courtesy of Special Interest Media.

Or purchase a paid subscription.
© 2026 Thoughts on Healthcare · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture