skip to content
Stephen Van Tran
Table of Contents

The model that trades tokens for molecules

OpenAI announced GPT-Rosalind on April 16, 2026 — a frontier reasoning model built exclusively for life sciences research — and in doing so made the most consequential product decision in the company’s history. Not because the model is bigger, faster, or more expensive than GPT-5.4. Because it is smaller, more focused, and deliberately restricted. GPT-Rosalind is OpenAI’s first domain-specific model series, fine-tuned for biochemistry, genomics, and protein engineering, and it is available only through a trusted-access program for vetted enterprise customers including Amgen, Moderna, Thermo Fisher Scientific, and the Allen Institute. OpenAI is also working with Los Alamos National Laboratory on AI-guided protein and catalyst design. During the research preview phase, usage does not consume existing API credits. This is not a product launch. It is a strategic pivot from selling general intelligence to selling specialized expertise — and the life sciences market it targets is worth tens of billions of dollars in annual R&D spending that has historically produced fewer than 50 new drugs per year.

The naming is deliberate and pointed. Rosalind Franklin was the British chemist and X-ray crystallographer whose Photo 51 — the famous diffraction image of DNA — was instrumental in revealing the double helix structure. Watson, Crick, and Wilkins received the 1962 Nobel Prize for the discovery. Franklin, who died of ovarian cancer in 1958, did not. By naming its first scientific model after Franklin, OpenAI is making a specific claim about what this model represents: foundational work that enables breakthroughs others will build upon. The claim is ambitious. The early benchmarks suggest it might be justified.

On BixBench, a bioinformatics benchmark developed by Edison Scientific that evaluates models on real-world computational biology tasks, GPT-Rosalind achieved a 0.751 pass rate — the highest score among models with published results. On LABBench2, a broader research task benchmark, the model outperformed GPT-5.4 on six of eleven tasks, with its most significant advantage on CloningQA, a task requiring the end-to-end design of reagents for molecular cloning protocols. The most striking validation came from a third-party evaluation with Dyno Therapeutics, a gene therapy company specializing in AAV capsid protein design. Using unpublished, previously unseen RNA sequences to guard against benchmark contamination, GPT-Rosalind’s best-of-ten submissions ranked above the 95th percentile of human experts on sequence-to-function prediction and around the 84th percentile on sequence generation. A model that can predict how RNA sequences will function with expert-level accuracy is not a chatbot with a biology textbook bolted on. It is a research instrument.

OpenAI is framing GPT-Rosalind as an accelerant for the 10-to-15-year timeline that typically separates a drug target hypothesis from FDA approval in the United States. The model can query specialized databases, parse scientific literature, interact with computational tools, and suggest new experimental pathways within a single interface. Alongside the model itself, OpenAI is releasing a free Life Sciences research plugin for Codex that connects to more than 50 scientific tools and data sources across human genetics, functional genomics, protein structure, biochemistry, clinical evidence, and public study discovery. The plugin gives researchers programmatic access to biological databases and computational pipelines — the equivalent of handing a scientist a single interface to query PubMed, UniProt, the Protein Data Bank, and dozens of specialized resources simultaneously while an AI reasons about the results. The combined system transforms the traditional workflow of read-hypothesize-design-test from a serial human process into a parallel human-AI collaboration where the model handles evidence synthesis and the scientist handles judgment.

The financial logic for OpenAI is straightforward. The company crossed $25 billion in annualized revenue in Q1 2026 and is preparing for an IPO that could value it near $1 trillion. General-purpose AI is a commodity race where margins compress as competitors match capabilities. Specialized AI for regulated industries — pharma, finance, defense — is a premium market where domain expertise, trusted access programs, and regulatory compliance create barriers to entry that protect margins. OpenAI is not building GPT-Rosalind because the life sciences need a chatbot. It is building GPT-Rosalind because the life sciences represent one of the few markets where AI companies can charge premium prices for demonstrably superior capabilities in a domain where the cost of being wrong is measured in failed clinical trials and lost years of patient access to treatments.

Follow the data into the valley of death

The pharmaceutical industry’s central problem is not a lack of money. Global pharma R&D spending exceeded $265 billion in 2025. The problem is that the money produces remarkably little output relative to its scale. The average cost to bring a single drug from discovery to market is approximately $2.6 billion. The average timeline from target identification to FDA approval is 12 to 15 years. The overall probability of a drug entering Phase I clinical trials and eventually receiving approval is roughly 7.9 percent. For every drug that reaches patients, more than a dozen candidates fail at various stages of development, and their costs are absorbed into the industry-wide average. This is Eroom’s Law — Moore’s Law spelled backward — the empirical observation that drug discovery has become exponentially more expensive per approved drug over the past six decades despite enormous advances in biological understanding and computational power.

AI is attacking the earliest and most wasteful stages of this pipeline with measurable results. In February 2026, Insilico Medicine announced that INS018_055, the first fully AI-designed drug for idiopathic pulmonary fibrosis, had completed Phase IIa clinical trials with statistically significant efficacy. The drug was conceived, designed, and optimized using AI in 18 months at a computational and discovery cost of approximately $6 million. For context, the traditional path to the same milestone typically costs $100 to $200 million and takes six to eight years. More than 173 AI-originated drug programs are now in clinical development globally, up from roughly 24 in late 2023. Phase I success rates for AI-discovered compounds run between 80 and 90 percent, compared to the historical average of approximately 52 percent. The AI drug discovery market was valued at approximately $1.9 billion in 2025 and is projected to reach $2.6 billion in 2026, growing at a 27 percent compound annual rate.

GPT-Rosalind enters this landscape as a different kind of tool than the generative chemistry platforms that produced INS018_055. Insilico’s pipeline uses variational autoencoders and generative adversarial networks to design novel molecules from scratch — a bottom-up approach that starts with target identification and works through molecular design, synthesis, and testing. GPT-Rosalind operates at a higher layer of abstraction: evidence synthesis, hypothesis generation, experimental planning, and multi-step research workflows. The model is not designing molecules. It is helping researchers think faster and more comprehensively about which molecules to design, which experiments to run, which targets to pursue, and which data points in the literature contradict or support a given hypothesis. In a pipeline that takes 12 years and costs $2.6 billion, compressing the first two years of target identification and validation by even 30 to 50 percent saves hundreds of millions of dollars and, more importantly, gets potential treatments to patients years earlier.

The competitive landscape for AI in drug discovery is already crowded and well-capitalized. Google’s Isomorphic Labs, a DeepMind spinout, released its proprietary drug discovery engine IsoDDE in February 2026, which doubled AlphaFold 3’s accuracy for certain drug design predictions — achieving 50 percent accuracy on the most difficult protein-ligand structure prediction cases where AlphaFold 3 managed 23.3 percent. Isomorphic Labs has secured partnerships with Eli Lilly and Novartis valued at nearly $3 billion in potential milestone payments and is keeping IsoDDE proprietary, marking a departure from the open-source approach that characterized AlphaFold’s earlier iterations. Recursion Pharmaceuticals, which merged with Exscientia in 2024, operates one of the most comprehensive AI drug discovery platforms in the industry with multiple programs in Phase I and Phase II clinical trials. Schrödinger’s physics-based platform produced zasocitinib, currently the most clinically advanced AI-assisted drug in the world, now in Phase III clinical trials through Takeda for inflammatory and autoimmune diseases.

What distinguishes GPT-Rosalind from these competitors is not its capabilities in molecular design — it does not do molecular design — but its positioning as the reasoning layer that sits above the entire drug discovery stack. OpenAI is not competing with Isomorphic Labs on protein structure prediction or with Recursion on cellular imaging. It is building the interface through which researchers interact with all of these tools and data sources simultaneously. The Codex Life Sciences plugin, which connects GPT-Rosalind to more than 50 scientific tools and databases, is the strategic play. If GPT-Rosalind becomes the default reasoning interface for drug discovery researchers, OpenAI captures value from every downstream tool and platform without needing to compete in any specific technical domain. It is the AWS strategy applied to scientific research: own the platform layer, let others build the specialized tools, and collect rent from the entire ecosystem.

The dual-use dilemma that keeps biosecurity experts awake

GPT-Rosalind’s restricted access model is not a marketing strategy. It is a direct response to the most serious risk category in AI development: the potential for AI systems trained on biological data to be misused for the design of dangerous pathogens. OpenAI’s decision to gate access exclusively through a vetted trusted-access program — requiring organizations to demonstrate they are conducting legitimate scientific research with clear public health benefits and maintaining strong security and governance controls — reflects a calculation that the reputational and existential risk of unrestricted access outweighs the revenue opportunity of broad deployment.

The concern is not theoretical. In 2023, MIT researchers demonstrated that large language models could provide step-by-step guidance for synthesizing potential pandemic pathogens when given appropriately crafted prompts. A 2024 RAND Corporation study found that AI systems could meaningfully lower the barrier to acquiring knowledge relevant to biological weapons development, even for non-experts. As AI models become more capable in biochemistry, genomics, and protein engineering — precisely the domains GPT-Rosalind targets — the dual-use risk intensifies proportionally. A model that can predict how RNA sequences will function with 95th-percentile expert accuracy can potentially predict how engineered sequences might cause harm with comparable accuracy. The capability that makes GPT-Rosalind valuable for drug discovery is the same capability that makes it dangerous in the wrong hands.

OpenAI’s response is to treat GPT-Rosalind more like a controlled substance than a software product. Access requires institutional vetting, demonstrated research legitimacy, and governance infrastructure. This approach mirrors how the U.S. government regulates access to Select Agents — dangerous pathogens like anthrax and Ebola — through the Federal Select Agent Program, which requires facility inspections, personnel security risk assessments, and ongoing compliance monitoring. The analogy is imperfect because GPT-Rosalind does not contain dangerous biological material. But the precedent is significant: OpenAI is voluntarily imposing restrictions on its most capable scientific model that go well beyond anything currently required by regulation, and in doing so is establishing a template for how AI companies might handle dual-use capabilities in the future.

The FDA has been adapting its regulatory framework to accommodate AI-discovered drugs, issuing draft guidance in January 2026 for evaluating AI-discovered drugs and launching an Accelerated AI Pathway Pilot that allows AI-discovered drugs with strong computational evidence to enter Phase I trials with streamlined IND applications. The RAISE Act, which took effect on March 19, 2026, imposes transparency, compliance, safety, and reporting requirements on developers of large frontier AI models. These regulatory developments create a framework within which GPT-Rosalind can operate — but they also impose obligations on OpenAI to demonstrate that its restricted access model is genuine and not merely performative. The test will come when competitors offer similar capabilities with fewer restrictions. If researchers migrate to less restrictive platforms because OpenAI’s vetting process is too slow or too burdensome, the responsible access model collapses. OpenAI’s challenge is to make the restricted access program fast enough to be practical while thorough enough to be meaningful — a balance that no institution, government or private, has fully achieved in the biosecurity domain.

The broader implication extends beyond OpenAI. Every AI company developing biological reasoning capabilities faces the same dual-use calculus. Google’s Isomorphic Labs chose to make IsoDDE fully proprietary. Anthropic has invested heavily in safety research and triggered its ASL-4 safety protocol for Claude Mythos 5, the first model to cross the 10-trillion-parameter threshold. The industry is converging on a consensus that the most capable AI systems require the most restrictive access controls — a principle that sounds obvious in the abstract but creates profound tension with the commercial incentive to maximize revenue by maximizing access. GPT-Rosalind’s restricted launch is OpenAI’s attempt to resolve that tension. Whether it succeeds will depend on whether the model delivers enough value to its restricted user base to justify the revenue it forecloses from unrestricted deployment.

The prescription for operators and the odds that matter

GPT-Rosalind’s launch reshapes the strategic landscape for every participant in the AI drug discovery ecosystem — from frontier labs building foundation models to pharmaceutical companies allocating R&D budgets to individual researchers choosing which tools to integrate into their workflows. The model’s significance lies not in any single capability but in the architectural decision it represents: the unbundling of general-purpose AI into domain-specific reasoning systems optimized for regulated industries where domain expertise is the primary source of value.

For pharmaceutical companies, the immediate question is whether to integrate GPT-Rosalind into existing discovery workflows or continue investing in proprietary AI platforms. The answer depends on the specific stage of the pipeline. Companies like Pfizer, Roche, and AstraZeneca — which have each committed $500 million or more to internal AI capabilities — are unlikely to outsource core drug design to OpenAI. But GPT-Rosalind’s value proposition is not drug design. It is the evidence synthesis and experimental planning layer that precedes design — the stage where researchers spend months reading literature, querying databases, and formulating hypotheses before a single molecule is drawn. If GPT-Rosalind can compress that phase meaningfully, even companies with mature internal AI programs will find it difficult to justify not using the tool alongside their proprietary platforms. The Codex Life Sciences plugin, with its connections to more than 50 scientific tools and data sources, makes the integration case even stronger by reducing the switching cost from existing workflows.

For AI drug discovery startups, GPT-Rosalind represents both a threat and an opportunity. The threat is platform displacement: if GPT-Rosalind becomes the default reasoning interface for life sciences research, startups that offer competing evidence synthesis or literature analysis tools face margin compression as their capabilities become a commodity layer beneath OpenAI’s platform. The opportunity is integration: startups that build specialized tools — molecular design engines, cellular imaging platforms, high-throughput experimental systems — can integrate with the Codex ecosystem and reach GPT-Rosalind’s user base without building their own reasoning layer. The companies most at risk are those in the middle: general-purpose AI platforms for life sciences that lack the deep specialization of a Recursion or Isomorphic Labs and the broad reasoning capabilities of GPT-Rosalind. The market is bifurcating into specialized computational tools and general reasoning platforms, and companies caught in the middle will be squeezed from both directions.

For investors, the key metric to watch is not GPT-Rosalind’s benchmark performance but its adoption velocity within the trusted-access program. OpenAI’s launch partners — Amgen, Moderna, Thermo Fisher Scientific, the Allen Institute, and Los Alamos National Laboratory — represent a cross-section of the pharmaceutical, scientific instruments, and research institution markets. If these partners report measurable productivity improvements within six months — faster literature reviews, higher-quality hypotheses, more efficient experimental designs — the case for GPT-Rosalind as a platform play becomes compelling. If adoption stalls because the restricted access model is too cumbersome or the model’s capabilities do not translate from benchmarks to real-world research productivity, the domain-specific AI thesis weakens for the entire industry.

The stakes are quantifiable. The pharmaceutical industry spends $265 billion per year on R&D and produces fewer than 50 new drugs annually. AI-discovered compounds now show Phase I success rates of 80 to 90 percent versus the historical average of 52 percent. The first fully AI-designed drug has demonstrated efficacy in Phase IIa for $6 million in discovery costs versus the traditional $100 to $200 million. Multiple analysts project a 60 percent probability that the first AI-designed drug receives regulatory approval by 2027. GPT-Rosalind does not need to revolutionize drug discovery to be valuable. It needs to shave months off the hypothesis-to-experiment timeline for a handful of programs at companies that spend billions on R&D annually. At that scale, even a 10 percent efficiency improvement is worth hundreds of millions of dollars per year across the pharmaceutical industry.

The question that GPT-Rosalind raises is larger than any single model or company. It is whether the future of AI is general-purpose systems that do everything adequately or domain-specific systems that do a few things exceptionally well. OpenAI has placed a $25-billion-revenue bet that the answer is both — general-purpose models for the mass market and specialized models for industries where expertise commands premium pricing. GPT-Rosalind is the first expression of that strategy. It will not be the last. If the model demonstrates that domain-specific AI can outperform general-purpose models in regulated scientific domains, expect every frontier lab to launch its own specialized series within 12 months. The race to build the best general-purpose model may soon be overshadowed by the race to build the best models for finance, law, engineering, and every other profession where getting the answer right matters more than getting it fast. For the pharmaceutical industry, that shift cannot come soon enough. Every month that a promising drug sits in the discovery phase instead of entering clinical trials is a month that patients who need it do not have access to it. If GPT-Rosalind can compress even a fraction of that timeline, its impact will be measured not in revenue or benchmarks but in years of life.

In other news

Anthropic releases Claude Opus 4.7 with benchmark-leading performance — Anthropic launched Claude Opus 4.7, its most capable commercially available model, matching the price of Opus 4.6 at $5/$25 per MTok. Opus 4.7 introduces improved agentic coding, high-resolution vision, and a new xhigh effort level, beating GPT-5.4 and Gemini 3.1 Pro across coding and reasoning benchmarks — though Anthropic conceded the model does not match the withheld Mythos.

Google brings AI Mode to Chrome with split-screen browsing — Google announced a redesigned AI Mode for Chrome on desktop, Android, and iOS that opens web links side-by-side with the AI chat interface, letting users compare details and ask follow-up questions without losing search context. A new cross-tab search feature also lets users pull context from multiple open tabs into a single AI conversation.

Anti-AI violence escalates with attacks on Sam Altman’s San Francisco home — OpenAI CEO Sam Altman’s residence was attacked twice in three days — first with a Molotov cocktail, then with gunfire — amid a broader wave of anti-AI sentiment that has included shootings at a pro-data-center councilman’s home in Indianapolis. The suspect was carrying a manifesto detailing anti-AI beliefs and a list of AI executive names, highlighting a growing and increasingly violent public backlash against the industry.

Physical Intelligence unveils π0.7 robot brain that learns untaught tasks — Robotics startup Physical Intelligence released π0.7, an updated foundation model that can plan, execute, and generalize across manipulation tasks it was never explicitly trained on. The model uses a combination of language conditioning and learned world models to transfer skills between different robot embodiments and environments.

EU proposes forcing Google to share search data with rival AI firms — European regulators are drafting a proposal that would require Google to share search index data with competing AI companies, arguing that Google’s dominance in web search gives it an unfair advantage in training AI models. The proposal, part of the ongoing Digital Markets Act enforcement, could reshape how AI companies access training data across the continent.