Photo by Alina Grubnyak on Unsplash
Nvidia's $40B AI equity bet: chipmaker as banker
/ 18 min read
Table of Contents
The chipmaker became the bank
Four months into 2026, Nvidia is no longer just selling AI infrastructure — it is underwriting the customers that buy it. On May 9, CNBC reported that the company had crossed $40 billion in equity commitments across the AI ecosystem this year alone, anchored by a $30 billion stake in OpenAI announced in late February and spanning seven publicly traded firms plus roughly two dozen private startup rounds, per a tally compiled by Benzinga from Nvidia’s filings. That number is more than the entire annual venture deployment of any single US fund family, and it has been assembled in less than 130 trading days. The pace is so brisk that the strategic partnership Nvidia announced on May 7 with data-center operator IREN, which includes the right to acquire up to $2.1 billion in IREN equity, was the chipmaker’s second multi-billion equity transaction of the same calendar week.
Nvidia is doing more than printing chips. It is buying its way upstream and downstream in the same motion: an $3.2 billion investment in 175-year-old optical fiber maker Corning to underwrite three new US factories — a deal CNBC framed as a bet that AI factories will need ten times more optical interconnect than today’s data centers — sits beside a $2 billion January investment in CoreWeave, a $2 billion stake in AI cloud company Nebius, $2 billion deployments in silicon-photonics specialists Marvell, Lumentum, and Coherent, and the $30 billion OpenAI piece that towers over everything. The non-marketable equity line on Nvidia’s balance sheet has grown from $3.39 billion in January 2025 to $22.25 billion as of January 2026, a 556 percent jump that the company disclosed in its fiscal 2026 annual report. The shape of the strategy is unmistakable: Nvidia is financing the entire AI supply chain on the bet that all of it runs on Nvidia silicon.
That logic has produced both the bull case and the bear case in the same week. Bulls argue that Nvidia is rationally pre-funding the demand its own products need — power, fiber, racks, model labs — and locking in a competitive moat that no rival can replicate without writing comparable checks. Bears point to a pattern that looks unsettlingly familiar: vendor financing as a tool to manufacture the appearance of growth, last seen at scale in the dot-com era when Cisco, Lucent, and Nortel underwrote their own customers’ purchases right up to the collapse. Both readings depend on the same fact set. What distinguishes them is whether the demand inside the loop is real, durable, and external to Nvidia’s own capital allocation — and that is now the dominant question on every Nvidia analyst call.
The stakes climb when you stack the equity bets against the parallel customer commitments. OpenAI is the obvious example: the $30 billion stake sits alongside a series of capacity commitments under which OpenAI has pledged to deploy Nvidia hardware as the backbone of its compute roadmap, even after Jensen Huang walked back a separate $100 billion September 2025 pledge as “never a commitment” and the deal pivoted to a smaller $20 billion equity vehicle. CoreWeave’s prior round was capitalized in part to buy Nvidia GPUs. IREN’s five-year warrant comes alongside a commitment to deploy up to five gigawatts of Nvidia’s DSX-branded AI infrastructure across its global pipeline, focused initially on a two-gigawatt Sweetwater, Texas campus. Money goes out, hardware orders come back, revenue books, then the cycle restarts. Whether that loop is a moat or a mirror depends on the answer to a question Wall Street will not be able to settle until the demand side of the equation either confirms or refuses the assumption.
Follow the wires, follow the watts
Read the deal pattern as a supply-chain map and Nvidia’s strategy snaps into focus. The investments cluster around three categories: compute consumers (OpenAI, CoreWeave, Nebius), physical-layer enablers (Corning for fiber, Marvell/Lumentum/Coherent for silicon photonics), and power/data-center capacity (IREN, plus partnerships with Constellation, AES, NextEra, and Vistra under the Emerald AI grid-flexibility consortium Nvidia helped seed). Each tranche addresses a known bottleneck. OpenAI consumes GPUs faster than any cloud can provision them, so the capital flows where the demand sits. Corning’s optical mfg capacity needs a tenfold increase if rack-scale systems are going to shift from copper to fiber — that is what the $3.2 billion deal underwrites, along with three new US facilities Corning is committed to building in North Carolina and Texas. IREN’s value is its access to five gigawatts of contracted power in markets where new substation capacity now has a five-to-seven-year queue at most utilities.
The Corning piece is the cleanest illustration of why Nvidia is willing to buy upstream rather than just sign procurement contracts. Optical interconnect is no longer a finishing component; it is becoming the dominant cost line in a rack-scale system. Nvidia’s own DSX architecture pushes the bandwidth requirement per GPU past what copper can credibly support at distances over a few meters, which means every multi-rack AI factory of the next two years has to be wired in glass. Corning’s expanded fiber and connectivity output is the difference between Nvidia delivering the racks on time and apologizing to OpenAI, Anthropic, and Google for slippage. The $3.2 billion equity right is cheap insurance against the bottleneck that would otherwise gate Nvidia’s own top-line growth. Simply Wall St’s breakdown of the announcement tallied 3,000 new American manufacturing jobs and a fiber-capacity expansion of more than 50 percent, but the operational story is simpler: Nvidia paid for the optical fiber it needs before the next supply crunch could occur.
IREN demonstrates the second pattern. The five-gigawatt commitment, paired with the Sweetwater flagship campus, is the kind of capacity that would have been considered absurd two years ago and is barely sufficient today. Five gigawatts is roughly the peak summer load of a US state the size of New Mexico, dedicated entirely to AI training and inference. Goldman Sachs Research, in its updated 2030 data-center power-demand note, now projects roughly 122 gigawatts of global data-center capacity online by the end of the decade, with US capacity alone climbing toward 95 gigawatts. IREN’s contracted footprint covers a meaningful share of that gap. Nvidia’s $2.1 billion warrant — exercisable across five years at $70 per share, for 30 million shares — gives the chipmaker an option on the upside of a power-arbitrage business it cannot itself operate, while ensuring that the gigawatts get filled with Nvidia hardware rather than custom silicon from Google, Amazon, or Microsoft. That is a remarkably efficient capital deployment: a few billion dollars buys a hedge against the most expensive bottleneck in the AI buildout.
The pattern’s most interesting wrinkle is the “neocloud” tier — CoreWeave, IREN, Nebius, and a half dozen smaller operators that Nvidia has stitched into something like a captive distribution network. These firms buy GPUs at scale, rent them to hyperscalers under specialized contracts, and provide a relief valve for capacity that AWS, Azure, and Google Cloud cannot or will not provision on Nvidia’s preferred timelines. Mizuho’s Jordan Klein flagged this dynamic explicitly, telling CNBC that neocloud investments “feel more questionable to me and likely investors” because “it smells like you are pre-funding the purchase of your own GPUs and products.” The criticism is sharp, but the structural rationale is also clear: Nvidia needs alternative buyers because the hyperscalers are simultaneously its largest customers and its most aggressive competitors, as the $40 billion Google-Anthropic deal earlier this year demonstrated when Google paired Anthropic with TPU capacity. A neocloud network that the chipmaker partially owns is a hedge against the day a hyperscaler-grade rival decides Nvidia is no longer required.
A proprietary takeaway worth surfacing here: stitching together the disclosed equity commitments against the announced gigawatt deployments and optical capacity expansions, Nvidia has effectively pre-purchased roughly 35 to 45 percent of the 2027-2028 hyperscale AI buildout it needs to maintain its current revenue trajectory. The OpenAI stake plus capacity commitments alone cover an estimated 12-15 gigawatts of compute. The IREN, Nebius, and CoreWeave investments cover another 8-10 gigawatts. The Corning, Marvell, Lumentum, and Coherent investments lock down the optical interconnect for that compute footprint. The Emerald AI consortium starts to address the grid-flexibility problem that would otherwise gate the rest. The picture is not vendor financing for the sake of revenue; it is forward integration into every constraint that could otherwise cap Nvidia’s growth.
When the loop becomes the leverage
The danger in any forward-integration story is that the loop becomes load-bearing in ways that hide demand fragility. Vendor financing has a uniquely bad history in technology. Cisco extended billions in customer financing during 1999-2000, took back a $2.25 billion inventory writedown in 2001, and watched its stock fall more than 80 percent before stabilizing. Lucent’s Lucent Capital Funding underwrote competitive carrier purchases in the same era and ultimately contributed to a $16 billion goodwill writedown in 2001. Nortel’s vendor finance vehicle wrote off $1.3 billion in 2001 alone. The NPR retrospective on those collapses is required reading for anyone tempted to dismiss the parallel out of hand: in each case, the vendor financing inflated the appearance of organic demand right up until the moment the underlying customers could not service their own borrowings, at which point the entire revenue ladder collapsed at once. The hardware-supplier bias toward overestimating downstream demand has been a structural feature of every major technology cycle since the railroads.
Ben Bajarin at Creative Strategies named the risk plainly in his CNBC commentary: “if the cycle turns, the market starts questioning how much of the demand was organic versus supported by Nvidia’s own balance sheet.” That sentence is the entire bear case in one line. Today, Nvidia trades at a multiple that assumes its current revenue is durable, organic, and external. Even a partial revelation that some material share of that revenue is being recycled through Nvidia’s own equity outlays would compress the multiple sharply. The 24/7 Wall St. analysis published in March argued the circular nature of these investments is potentially fueling its own AI bubble, and the more aggressive estimates put the total circular financing across the major chip and cloud players at over $800 billion in 2025-2026, per a BlockEden analysis that walked through the OpenAI-Microsoft-Nvidia-Oracle web in detail. At that scale, a meaningful demand shock would not just compress Nvidia’s multiple; it would force restatement of revenue across the AI supply chain.
Industry analyses have mapped the dependencies with disturbing clarity. The Beam.ai walkthrough of the AI infrastructure race traces Nvidia, Microsoft, OpenAI, Oracle, AMD, and CoreWeave passing capital and contracts to each other in a tight, recursive structure where the same dollars appear to count as revenue, capex, and equity contribution at different nodes of the network. The graphic is not an accusation. But it does illustrate why the question of organic versus engineered demand is genuinely hard to answer from the outside. Each leg of each deal is documented, audited, and disclosed. What is not disclosed is what the equilibrium would look like if every participant simultaneously had to source funding externally rather than from each other. Vendor financing is not fraud; it is, structurally, an information problem. The cycle turns when the information problem becomes a confidence problem.
The bull rebuttal is that the comparison to Cisco and Lucent breaks down on the demand side. AI workloads, the argument goes, are not telecom-bandwidth bets being made by undercapitalized competitive local exchange carriers; they are foundational compute being procured by the world’s most profitable corporations to deliver products with already-confirmed user demand. ChatGPT has 800 million weekly active users. Claude is run by 60 percent of the Fortune 500. Gemini is integrated across Google’s first-party surfaces. Microsoft’s Copilot revenue, while opaque, sits inside an Office 365 base that has paid for its own software upgrades for thirty years. Noah Smith made this argument forcefully in March, pointing out that the circular deals between hyperscalers and frontier labs are largely “intracompany” capital flows between cash-generative giants rather than vendor financing to fragile startups; the structural risk profile, he argued, is not comparable to dot-com vendor finance. The argument has real force. Microsoft writing a check that flows through OpenAI to Nvidia is different from Lucent writing a check that flowed through a CLEC into Lucent equipment, because Microsoft can pay the bill regardless.
But the argument has limits. The neocloud tier is structurally closer to the dot-com pattern than the hyperscaler tier. CoreWeave and IREN are highly leveraged, mid-cap operators whose business models depend on continued GPU price stability and continued strong demand from a small concentrated set of customers. If Anthropic or OpenAI renegotiates pricing — or, more dangerously, if a credible open-source frontier model collapses inference costs by 5x — the cash flow assumptions inside the neocloud tier degrade quickly. Nvidia’s equity stakes in those firms would mark down in step. The hyperscaler tier may be insulated; the neocloud tier is not. The risk is concentrated precisely where Nvidia’s own financing pattern is most aggressive. Klein’s “questionable” framing is reading the marginal balance, not the average one.
Regulators are watching, though without yet acting. The SEC has not opened a formal inquiry into circular financing in AI, but staff guidance issued in March quietly tightened revenue-recognition reminders for transactions where the buyer is also a material equity investee. The EU’s AI Office has not commented. The FTC, under Chair Andrew Ferguson, has signaled AI competition oversight more broadly, and the kind of supply-chain consolidation Nvidia is engineering through its equity bets would historically have drawn antitrust scrutiny if performed through outright acquisitions. The fact that Nvidia is buying minority stakes rather than control positions is what keeps the deals below the Hart-Scott-Rodino reporting thresholds in most cases. That is a regulatory arbitrage with its own brittleness: the equity structure works as long as Nvidia avoids effective control, but the practical leverage embedded in five-gigawatt deployment commitments arguably gives the chipmaker effective control without legal control. That is exactly the kind of distinction enforcement agencies have, historically, eventually decided to collapse.
What the next quarter has to prove
The next two quarters will settle two questions Wall Street has been unable to answer from the outside. The first is whether the demand inside the loop holds up when external macro pressure rises. Credit conditions are tightening, as FXEmpire noted in its 2026 outlook on the AI market, and any compression in venture funding or hyperscaler capex would pass through to the neocloud tier first. The second is whether Nvidia’s revenue mix begins to show the tell that vendor-financed customers always show: a slow rise in receivables, lengthening days-sales-outstanding, and a quiet shift in the customer concentration disclosure toward the firms in which Nvidia holds equity. Cisco’s revenue mix in 1999 showed all three signals six quarters before the writedowns hit. Nvidia’s fiscal Q1 2027 print, due in late August, will be the first opportunity to look for any of them.
The competitive dimension matters too. Nvidia’s moat thesis depends on the absence of a credible second-source supplier at the rack-scale tier, but AMD’s MI400 architecture roadmap and the continued maturation of custom silicon at Google, Amazon, and Microsoft chip away at that premise on a quarterly basis. If a hyperscaler-grade alternative reaches Nvidia parity on the workloads that matter — frontier training plus inference at the new reasoning-model scales — the circular-financing apparatus loses much of its purpose. The neoclouds would not collapse, but they would be repriced. The OpenAI stake would still be valuable on its own merits, but the implicit “every dollar of AI capex eventually books through Nvidia” assumption would no longer hold. That risk does not look acute in 2026, but it does look more acute in 2027 and 2028 than it has at any point in the prior cycle.
There is also a stranger question on the horizon: what happens to Nvidia’s equity portfolio when its largest investee, OpenAI, completes its planned public offering. OpenAI’s IPO timeline, as my analysis of the Wall Street deployment ventures discussed, could put a roadshow into market as early as Q4 2026 with revenue at $25-billion-plus annualized. A successful IPO would mark Nvidia’s $30 billion stake to public-market pricing in real time. A failed one — or one that prices below the most recent $852 billion private valuation — would force a writedown that ripples through Nvidia’s own non-marketable equity disclosures. Either outcome would crystallize a question that today exists only as a theoretical concern: how much of Nvidia’s reported earnings power is durable cash flow versus mark-to-market gain on equity bets the company assembled with the proceeds of selling GPUs to those same firms. That accounting boundary has been blurry for two years. It will not stay blurry forever.
Operator checklist for engineering leaders, finance teams, and platform investors watching the loop unfold:
- Audit your AI vendor’s funding stack. If a critical AI tool is built on a neocloud whose equity round was anchored by Nvidia, your service reliability depends on a financing chain you do not control. Diversify across at least two compute providers whose capital structures do not overlap, and ensure SLAs cap your downside if the upstream renegotiates terms.
- Pressure-test pricing assumptions. Inference costs that look stable today are partially supported by GPU pricing that the circular financing apparatus has helped maintain. Model a 30-40 percent inference-cost shock and confirm your unit economics survive it; if they do not, restructure the workload now while you still have time.
- Watch the Q1 and Q2 Nvidia 10-Q footnotes. The non-marketable equity disclosure, the customer concentration note, and the related-party transactions section will tell you more about the durability of the AI capex cycle than any keynote will. Look specifically for changes in DSO, growth in receivables tied to investee customers, and any new related-party disclosures.
- Track the OpenAI IPO pricing. A range below $800 billion in the formal S-1 would mark down Nvidia’s largest equity position and could trigger broader portfolio remeasurement. A range above $1 trillion would, conversely, validate the entire forward-integration thesis. The S-1 filing, when it comes, will move more than just OpenAI.
- Map your own customer concentration honestly. If you are an AI startup whose business model depends on credits, partnerships, or co-marketing from one of the major hyperscalers or chip vendors, you are inside the loop whether you have raised primary capital from them or not. The exit door narrows as the loop tightens — secure independent runway now.
In other news
Anthropic ships “dreaming” for self-improving agents — Anthropic introduced a scheduled background process called “dreaming” that lets Claude Managed Agents review their own session transcripts, extract patterns, and write learnings into a curated memory layer before deployment, with legal AI firm Harvey reporting roughly 6x improvement in task completion rates after enabling the feature (VentureBeat). The release was paired with Multi-agent orchestration and Outcomes additions at the Code w/ Claude 2026 conference.
Snap and Perplexity end $400M deal with no revenue booked — Snap disclosed in its Q1 2026 earnings that the $400 million cash-and-equity partnership with Perplexity, announced last November, was “amicably ended” in Q1 after the integration into Snapchat’s Chat interface never moved beyond limited testing, with Snap’s forward guidance now explicitly assuming zero Perplexity contribution (TechCrunch). The collapse is the largest publicly disclosed AI partnership failure of the cycle and a useful reminder that headline deal values are not the same as contracted revenue.
Mistral launches 128B flagship with async coding sessions — Mistral introduced a 128-billion-parameter flagship model bundled with async cloud coding sessions and a new Work agentic mode inside Le Chat, per industry coverage of the May rollout. The release reframes Mistral as a competitor in agentic developer tooling rather than purely a foundation-model provider, and tracks the broader market shift toward managed coding agents.
Microsoft Agent 365 reaches GA — Microsoft Agent 365, the company’s enterprise governance layer for AI agents that runs identity, security, and policy across third-party agents alongside Microsoft’s own, became generally available in May, per release notes summaries from the period. The GA milestone matters because it gives compliance teams a single management surface for the multi-agent enterprise deployments that have proliferated since late 2025.
OpenAI rolls GPT-5.5 Instant as default — OpenAI replaced GPT-5.3 Instant with GPT-5.5 Instant as the default model for ChatGPT and the chat-latest API alias on May 5, with the new model scoring 81.2 on AIME 2025 (vs 65.4) and 76 on MMMU-Pro (vs 69.2), per TechCrunch’s coverage of the launch. Plus and Pro web users also picked up the ability for GPT-5.5 Instant to reference past conversations, files, and Gmail when answering.
White House preps an FDA-style AI vetting executive order — National Economic Council Director Kevin Hassett confirmed on May 7 that the White House is drafting an executive order to formally vet new frontier AI models before release, with Hassett explicitly comparing the framework to the FDA’s drug-approval process and citing Anthropic’s Mythos vulnerability-discovery model as the trigger (CNN Business covered the policy context in adjacent reporting on AI’s labor-market impact). The order would build on the CAISI pre-deployment testing regime that all five US frontier labs agreed to in April.