skip to content
Stephen Van Tran
Table of Contents

The most closely watched annual temperature check on the state of artificial intelligence landed on April 13, and the numbers inside it deserve more attention than the week’s model announcements and funding rounds combined. Stanford HAI’s ninth annual AI Index is a 400-page audit covering capability benchmarks, investment flows, geopolitical competition, employment effects, environmental costs, and public sentiment — a breadth that no single press release can replicate. The headline finding, repeated across multiple chapters, is also the most uncomfortable: AI capability is accelerating at a pace that the systems built to govern, evaluate, and understand it cannot match. That gap between what these models can do and what anyone can reliably say about their risks is not a future problem. It is the problem that organizations deploying AI at scale are living with today.

The ninth edition arrives in a year when frontier models have cleared benchmarks that their predecessors couldn’t touch, when investment has more than doubled, when two billion people are within arm’s reach of a capable AI system, and when the institutions responsible for regulating that reach are measurably less prepared than they were twelve months ago. The Stanford Index is valuable precisely because it refuses to flatten those contradictions into a simple optimism or pessimism thesis. The evidence it assembles points, in nearly every chapter, to a technology running faster than the track it was built to run on.

The Machine Outran the Manual

A year ago, “AI at human level” was a marketing claim dressed up as a benchmark score. This year, the Stanford Index documents something closer to a genuine leap. According to the IEEE Spectrum analysis of the report, AI models now meet or exceed human baselines on PhD-level science questions, competition-level mathematics, and multimodal reasoning — domains where human expertise was considered a safe ceiling as recently as 2024. On SWE-bench Verified, the most widely cited proxy for real-world software engineering, scores climbed from roughly 60 percent to nearly 100 percent of human baseline in a single year — a trajectory that tracks directly with this month’s Claude Opus 4.7 release, which scored 87.6% on that same benchmark. Google’s Gemini Deep Think won gold at the International Mathematical Olympiad. Cybersecurity agents jumped from 15 percent problem-solving success in 2024 to 93 percent in 2026, a number that has implications not only for defense but for offense — and that directly informed why OpenAI and Anthropic moved to restrict access to their most powerful cyber AI capabilities earlier this week. And as The Next Web reports, AI agents on Terminal-Bench — which measures autonomous command-line task completion in live shell environments — surged from 20 percent last year to 77.3 percent today. For infrastructure teams thinking about where autonomous agents fit in their pipelines, that single datapoint is more consequential than any model release announcement.

The report pairs this performance surge with what the authors call the “jagged frontier” — a paradox that becomes more pronounced as the models get stronger. The same systems that win international mathematics competitions read analog clocks correctly only 50.1 percent of the time. Robots still succeed at just 12 percent of common household tasks like laundry sorting or dishwashing. Frontier models that score above 50 percent on Humanity’s Last Exam — a benchmark designed to be unsolvable by conventional AI — cannot reliably sequence multi-step plans that a reasonably organized person would accomplish before lunch. Unite AI’s deep-read of the report describes this pattern precisely: the jagged frontier is not a bug waiting to be patched. It is a structural feature of how these systems are trained, and it will persist until the training paradigm itself changes.

The jaggedness matters enormously for enterprise deployment. Organizations that see a 93 percent cybersecurity success rate and deploy accordingly, without accounting for the 50 percent clock-reading failure — a stand-in for any task requiring common-sense physical-world reasoning — will build brittle systems that perform brilliantly until they don’t, with failure modes that were baked in from the start. This is not a reason to avoid deployment; the capabilities are real and the productivity gains documented in the economics chapter are equally real, with AI showing 14 to 26 percent gains in customer support productivity and up to 72 percent gains in marketing output generation. It is a reason to map the jagged frontier for your specific workloads before the production incident maps it for you.

One figure crystallizes the capability acceleration more starkly than any benchmark table. The report finds that global AI compute capacity has grown 3.3 times annually since 2022, representing a 30-fold increase since 2021. Digital Information World’s breakdown of the findings puts data center power consumption at 29.6 gigawatts — roughly equivalent to powering the entire city of New York at peak demand. Nvidia GPUs account for over 60 percent of that capacity globally. The infrastructure building that underlies the capability curve is not a trailing indicator; it is the leading signal that the acceleration documented in this year’s Index will continue into next year’s.

What is not keeping pace is the measurement infrastructure for understanding what can go wrong. Documented AI incidents — tracked through the AI Incident Database and similar repositories — rose from 233 in 2024 to 362 in 2025. That 55 percent increase happens in an environment where frontier model developers overwhelmingly report capability benchmarks and have progressively less to say about safety and responsible AI metrics. The Foundation Model Transparency Index, which measures how much the 10 most capable models disclose about their training data, evaluation procedures, and known failure modes, dropped from a score of 58 to 40 between last year’s Index and this one. The models got stronger. The window into how they work got smaller.

China Closed the Gap for Pennies

The geopolitical chapter of the Stanford Index contains the single most jarring quantified finding in the entire report, and it has received a fraction of the attention it warrants. As of March 2026, the performance gap between the leading American AI model and the leading Chinese AI model had narrowed to 2.7 percentage points — down from a gap of 17.5 to 31.6 percentage points in May 2023. The Next Web’s analysis makes the investment context explicit: the United States deployed $285.9 billion in private AI investment in 2025, while China deployed $12.4 billion. That is a 23-to-1 spending ratio for a performance gap that is now functionally negligible on the benchmarks the field uses to measure capability.

The immediate reading is that the US is vastly outspending China for very little competitive performance advantage. The more precise reading is more complicated. The US still produces more top-tier models — 50 notable models in 2025 versus China’s 30, up from 15 the prior year — and dominates high-impact patent production. China leads in publications (23.2 percent of global output), citation volume (20.6 percent), and industrial robotics installations (295,000 in 2024, against Japan’s 44,500 and the US’s 34,200). These are not symmetric advantages. The US model count and patent leadership reflect investment in frontier capabilities; China’s publication dominance and robotics deployment reflect a broader industrial application strategy. The competition is not between two entities doing the same thing at different price points. It is between two entities with different theories of how AI power compounds.

The infrastructure concentration data is the chapter’s other critical finding. The United States hosts 5,427 data centers — more than ten times any other nation. California alone absorbed $218 billion of the US total, or 76 percent. Taiwan’s TSMC fabricates virtually every leading AI chip. That single-point dependency for the world’s AI semiconductor supply chain is a supply-chain risk that no amount of domestic model production can fully hedge. WebProNews flagged the same asymmetry: the US leads in model deployment infrastructure but is exposed in the manufacturing layer underneath it, while China is building the manufacturing and energy infrastructure to support a scaling trajectory that its current model performance does not yet justify.

The talent data may be the most structurally alarming finding in the geopolitics chapter. AI researchers and developers migrating to the United States dropped 89 percent since 2017 — with 80 percent of that entire decline occurring in the past year alone. Switzerland now ranks first globally in AI talent density, at 110.5 AI researchers and developers per 100,000 inhabitants. The US produces the most capable models and attracts the most capital, but it is losing the human pipeline that has historically driven frontier research. If that trend continues for another cycle, the 23-to-1 investment ratio will begin to look different on model performance tables than it does today. One proprietary takeaway synthesized across both the investment and talent data: the US currently buys capability advantages through compute spending rather than researcher concentration. That is a strategy that works until compute scaling laws stop compounding — and the Index’s own benchmark data suggests some of those laws are approaching inflection points on narrow tasks even as general capability continues to climb.

The Bill Nobody Budgeted For

The Stanford Index’s governance and responsible AI chapters represent its most underreported findings, which is also the most revealing thing about how the industry reads this document. Techopedia’s coverage frames the problem cleanly: almost every frontier model developer publishes results on capability benchmarks, while reporting on responsible AI benchmarks remains sparse. The implication is not that these developers are indifferent to safety. It is that the field has not yet converged on responsible AI metrics that are comparably rigorous, comparable across models, and accepted as authoritative by buyers and regulators. Capability has Elo scores and benchmark tables. Responsibility has guidance documents and disclosure pledges.

The Foundation Model Transparency Index score dropping from 58 to 40 is the quantified version of that trend. The most powerful models are increasingly developed by the largest organizations, which have the most competitive incentives to limit disclosure about training data provenance, fine-tuning procedures, and known failure domains. The irony documented across the report is precise: as models become more capable and consequently more consequential, the information available to independent researchers, enterprise buyers, and regulators about how they actually work is diminishing, not growing. Artificial Intelligence News summarizes the core finding: “AI safety benchmarks are falling behind,” and the infrastructure for evaluating responsible AI does not yet exist at the quality required to provide meaningful assurance to anyone making high-stakes deployment decisions.

The employment data in the economics chapter contains a similar gap between perception and measurement. The aggregate employment numbers remain relatively stable — unemployment did not spike, and workers with the highest AI exposure saw more job stability than those with the least. But the age-cohort data tells a different story: software developers aged 22 to 25 experienced a nearly 20 percent employment decline since 2024. Entry-level positions in customer support have also contracted. The pattern is consistent with a transition where AI absorbs the most structured, highest-volume portions of junior work before restructuring the surrounding roles rather than eliminating them entirely. The 14 to 26 percent productivity gains documented in the report accrue first to organizations with existing professional staff; the cost of that accrual falls first on entry-level workers. That distribution problem is not visible in aggregate employment statistics.

The environmental accounting chapter documents a cost that is even less visible in current procurement decisions. Training Grok 4 generated approximately 72,816 tons of carbon-equivalent emissions — compared to 5,184 tons for GPT-4, a 14-fold increase as model scale increased. Carbon emissions from inefficient model inference run more than ten times higher than from optimized alternatives. Water consumption for model inference at scale may exceed the drinking water needs of 12 million people annually, based on estimates the report cites for large-scale inference deployments. None of these costs currently appear on model pricing cards. They do appear in enterprise sustainability reporting requirements, and they will appear in regulatory compliance frameworks as those frameworks become more specific. Organizations embedding frontier models into high-volume workflows today are incurring environmental liabilities that their current cost models do not capture.

The trust data brings the governance picture to its sharpest focus. Only 31 percent of US respondents trust the government to regulate AI effectively — the lowest figure among all surveyed nations and a significant drop from prior editions of the Index. The global average sits at 54 percent. In Singapore, the figure is 81 percent. The US regulatory environment is not failing simply because it lacks the technical capacity to keep pace with AI development. It is failing partly because the public that it nominally serves has already concluded that it will. That perception, once entrenched, becomes self-fulfilling: low regulatory credibility reduces compliance incentives, which produces the incidents that confirm the original skepticism.

What to Do Before Next Year’s Report

The Stanford Index is not a policy brief, and Stanford HAI is explicit that it does not advocate for specific regulatory approaches. But the data it assembles points to clear action priorities for enterprise operators who cannot wait for the governance gap to close before deploying AI in production. The following checklist is derived from the report’s own findings and the competitive dynamics the data implies.

  • Map your jagged frontier before deployment. The Index documents that no frontier model excels uniformly across domains. Before committing a model to a high-stakes workflow, run a representative benchmark test across the specific tasks in scope — not the headline benchmarks the model vendor highlights. A 93 percent cybersecurity success rate is not the number that governs your deployment; the success rate on your specific threat intelligence pipeline is. Maintain a living map of where each deployed model fails and rotate workloads accordingly.

  • Demand transparency metrics from vendors, not just capability scores. The Foundation Model Transparency Index score of 40 means that the average frontier model discloses less than half the information a rigorous risk assessment would require. When procuring models for regulated use cases — healthcare, finance, legal — require vendor disclosure on training data provenance, fine-tuning scope, and known failure modes as a contractual term, not a courtesy ask. If a vendor cannot provide it, that is a risk disclosure in itself.

  • Reassess your geographic vendor concentration. The 2.7 percent US-China performance gap means that Chinese-origin models are now benchmarked equivalently to US-origin models on many standard evaluations. For organizations with workloads that are legally or operationally constrained by data residency, sovereignty, or export control requirements, that performance parity matters: the optimization space for compliant deployment has expanded. For organizations without those constraints, the competition should drive harder pricing negotiations with existing US-origin vendors.

  • Build entry-level talent pipelines deliberately. The 20 percent decline in junior software developer employment since 2024 is a leading indicator, not an outcome to celebrate. Organizations that eliminate entry-level roles to capture short-term AI efficiency gains are also eliminating the pipeline for mid-career and senior practitioners five years from now. The Index’s talent migration data — 89 percent fewer AI researchers moving to the US in the last year alone — compounds this risk. Invest in structured AI training pathways for junior staff rather than replacing them; the medium-term cost of not doing so will arrive precisely when the next capability curve demands expertise your bench does not have.

  • Price environmental costs into your AI infrastructure budget now. The 14-fold increase in training emissions from GPT-4 to Grok 4 will not reverse as models continue to scale. Inference optimization is the highest-leverage near-term intervention: the Index documents more than a 10-to-1 variance in inference emissions between efficient and inefficient deployment configurations. Audit your inference stack for utilization rates, batching efficiency, and model routing — the organizations that optimize this layer in 2026 will have a meaningful cost and compliance advantage when environmental reporting requirements tighten.

  • Treat the AI incident count as a lagging indicator. Documented incidents rose 55 percent to 362 in 2025. That figure reflects only incidents that were identified, attributed, and logged — an unknown fraction of actual events. Establish internal AI incident tracking before regulators require it; internal data gives you the signal before the regulatory framework defines what counts. Organizations with their own incident baseline will have a much cleaner compliance story than those scrambling to reconstruct it under external audit.

The Stanford AI Index will return next April with a tenth edition, and the trajectory its data implies suggests the gaps it documents this year will be wider, not narrower, unless the organizations deploying these systems begin treating governance as an investment rather than a cost center. Capability will compound regardless. The question is whether governance compounds with it.

In other news

Meta debuts Muse Spark, its first closed-source AI model — Meta launched Muse Spark on April 8, the first model released by its new Superintelligence Labs led by former Scale AI CEO Alexandr Wang. Built over nine months and powered by $115–135 billion in 2026 capital expenditures, Muse Spark marks a sharp pivot away from Meta’s Llama open-source strategy. The model is competitive on multimodal and health benchmarks but lags frontier models on coding, and will roll out to WhatsApp, Instagram, and Facebook in the coming weeks (CNBC).

Anthropic and OpenAI fight over who bills you for agentic AI — Anthropic launched Managed Agents in public beta on April 8 at $0.08 per session hour plus standard Claude API rates, with Notion, Rakuten, Sentry, Asana, and Atlassian among its launch customers. One week later, OpenAI shipped an updated open-source Agents SDK with a model-native runtime at no additional charge beyond existing API pricing. Google and Microsoft are metering the equivalent layer by consumption component in their cloud platforms. The strategic split matters: Anthropic is building recurring runtime revenue, OpenAI is using the free layer to defend API market share, and the difference will shape developer platform loyalty for the next three years.

OpenAI crosses $25 billion in annualized revenue — OpenAI reached $25 billion in annualized revenue at the end of February, up from $21.4 billion at year-end 2025 and roughly $6 billion at the close of 2024. Winbuzzer reports that Anthropic has closed to $19 billion ARR — roughly 14 times its revenue from a year earlier — compressing the gap between the two companies to $6 billion and raising the competitive stakes for OpenAI’s reported H2 2026 IPO filing target, which would value the company at up to $1 trillion.