skip to content
Stephen Van Tran
Table of Contents

The numbers stopped making sense about three earnings calls ago. Amazon plans to spend $200 billion in capital expenditures in 2026. Alphabet is guiding $175 to $185 billion. Meta has committed $115 to $135 billion. Microsoft rounds out the quartet at $110 to $120 billion. Add Oracle’s aggressive data center buildout and the combined hyperscaler capex bill for 2026 approaches $690 billion — a 60 percent increase over 2025 and roughly three times what the United States spent building the entire Interstate Highway System, adjusted for inflation. Worldwide AI spending, per Gartner’s January forecast, will total $2.52 trillion this year, a 44 percent jump that makes 2025’s already eye-popping $1.76 trillion look quaint.

What makes 2026 different from previous capex cycles is not the magnitude of the checks being written. It is the simultaneous collapse of the financial cushion that used to absorb them. Morgan Stanley projects Amazon will post negative free cash flow of $17 billion this year — the first time the company has burned cash since its pre-AWS era. Pivotal Research sees Alphabet’s free cash flow plummeting nearly 90 percent, from $73.3 billion in 2025 to $8.2 billion. Barclays expects a similar 90 percent drop at Meta and a 28 percent decline at Microsoft. These are not struggling companies. They are the most profitable corporations on Earth, and they are voluntarily converting fortress balance sheets into construction budgets because every one of them believes that whoever builds the most GPU clusters fastest wins the next decade of computing. The wager is existential — not because any of them will go bankrupt, but because the scale of the bet leaves almost no margin for the thesis to be wrong.

The infrastructure arms race has already triggered a cascade of second-order effects: a trillion-dollar bond binge that is reshaping capital markets, a chip startup funding surge that poured $1.1 billion into Nvidia challengers in a single week, and a Broadcom projection of $100 billion in AI chip revenue by 2027 that signals the end of Nvidia’s monopoly on the hardware layer. The capex tsunami arrives in the same week that OpenAI shipped GPT-5.4 with native computer use and the Pentagon blacklisted Anthropic over safety guardrails — two events that underscore how desperately every player in this market needs more compute, more infrastructure, and more leverage over the hardware supply chain. This is the story of the biggest capital deployment in technology history — and the uncomfortable question of whether it will pay off before the money runs out.

Follow the concrete and the copper

To understand where $690 billion actually goes, start with the physical reality. AI data centers are not just bigger versions of cloud data centers. They consume orders of magnitude more power, generate far more heat, and require specialized cooling, electrical distribution, and networking that traditional facilities were never designed to support. A single Nvidia GB200 NVL72 rack draws roughly 120 kilowatts — enough to power about 40 American homes. A thousand-rack training cluster, which is what frontier model companies now consider baseline, needs the electrical output of a small power plant. When Amazon says it is spending $200 billion, the majority of that budget is not going to chip vendors. It is going to land acquisition, power contracts, construction labor, liquid cooling systems, and the miles of copper and fiber that connect thousands of racks into a coherent training fabric.

The geographic footprint is staggering. TechCrunch reported in late February that OpenAI, Oracle, Nvidia, Microsoft, Google, Meta, and a constellation of smaller players have signed more than $100 billion in data center construction deals in the last six months alone. Oracle committed to raising $45 to $50 billion in debt during the 2026 calendar year to fund its buildout. The company that once seemed content as a database legacy play has transformed itself into the preferred colocation partner for frontier AI labs, with customers including OpenAI and xAI that need capacity faster than they can build it themselves.

The energy dimension is becoming politically explosive. President Trump faces a growing dilemma as AI data centers strain the electrical grid ahead of the 2026 midterms, and major tech companies — including all the hyperscalers — have pledged to cover energy and infrastructure costs so consumers do not see higher utility bills. The scale of the power demand is hard to overstate: a modern AI training campus requires 500 megawatts to a gigawatt of continuous power, equivalent to a mid-size city, and the dozens of campuses now under construction will collectively draw more electricity than many small countries consume in a year. Utilities in Virginia, Texas, and the Pacific Northwest have begun warning that grid capacity is approaching its limits in key data center corridors, forcing some developers to invest in on-site natural gas generation or nuclear partnerships. That pledge to shield consumers is easy to make when revenue is growing 30 percent year-over-year. It becomes harder to honor if the AI revenue thesis stumbles and the data centers still need their gigawatts.

Gartner’s breakdown illuminates the spending hierarchy: AI infrastructure remains the largest category at $1.37 trillion globally in 2026, up from $965 billion in 2025. AI-optimized servers alone will see a 49 percent spending increase, representing 17 percent of total AI spending. AI services clock in at nearly $589 billion, and AI software at $452 billion. The raw infrastructure layer — the concrete, copper, silicon, and electricity — still dwarfs everything else. We are building the physical substrate of a new computing paradigm, and the construction bill is arriving before the tenants have moved in.

Here is a proprietary calculation that illustrates the scale risk: if you divide the projected $690 billion in hyperscaler capex by the combined AI-attributable revenue these companies are expected to generate in 2026 — approximately $180 billion across AWS, Azure, Google Cloud, and Meta’s advertising AI — the implied capex-to-AI-revenue ratio is roughly 3.8x. For every dollar of AI revenue, the hyperscalers are spending nearly four dollars on infrastructure to support it. That ratio needs to compress below 2x within three years for the investment to generate acceptable returns. History offers one precedent: the fiber-optic buildout of the late 1990s, where the capex-to-revenue ratio peaked above 5x before collapsing — along with dozens of telecom companies. The hyperscalers are better capitalized than WorldCom ever was. But the math is the math.

The trillion-dollar IOU

The most consequential shift in 2026 is not how much money the hyperscalers are spending. It is where the money is coming from. For the first time in the modern tech era, the industry’s richest companies are turning to debt markets at scale to fund their AI ambitions, shattering what CNBC described as an “unspoken contract” between big tech and its investors. That contract was simple: mega-cap tech companies would maintain fortress balance sheets, fund growth from free cash flow, and return excess capital to shareholders through buybacks and dividends. The AI capex surge has ripped that contract in half.

By the end of February 2026, the big five hyperscalers had already issued $45 billion in U.S. bonds — nearly half as much as they did in all of 2025. Alphabet led the charge with a $20 billion bond offering that included a rare 100-year sterling-denominated tranche — a century bond from a company that has existed for less than three decades. Fortune reported in early March that Google, Meta, and Oracle are collectively on a $1 trillion borrowing trajectory, with investors buying 30-, 40-, and even 100-year maturities. The appetite for the debt, ironically, signals how much faith the bond market has in these companies even as equity analysts grow nervous. One bond fund manager told Fortune that investors are comfortable because the risk-reward remains balanced. A century bond from Alphabet is not the same as a century bond from a startup. But it is a century bond from a company betting nearly 50 percent of its revenue on an infrastructure thesis that may or may not play out.

Amazon’s filing with the SEC indicating it may seek to raise additional equity and debt marks perhaps the most telling signal. The company that closed a $110 billion funding round for OpenAI with a $50 billion anchor check is now signaling that it may need to raise capital itself. Amazon has not needed external capital since its early days. The fact that it is now contemplating equity dilution — even as a contingency — suggests that internal financial models see the possibility of sustained negative free cash flow extending beyond 2026 into 2027 or even 2028. This is not a liquidity crisis. Amazon, Alphabet, and Meta all retain investment-grade credit ratings that make borrowing cheap. But cheap debt is still debt, and the cumulative leverage building across the sector is creating a new category of risk that did not exist two years ago.

The Federal Reserve is watching. Fortune reported in February that Fed staff are monitoring the intersection of elevated tech valuations and rising tech debt, though they currently judge stock price risk as more concerning than the debt load itself. That assessment may shift if a recession compresses cloud growth while the capex bills keep arriving. The hyperscalers have locked in long-duration debt at relatively low rates, which insulates them from interest rate spikes. But their shareholders have not locked in anything. Every dollar of capex that does not translate into AI revenue within three to five years comes directly out of the buyback budget, the dividend policy, or the stock price. For investors who bought these stocks as cash-generation machines, the pivot to infrastructure companies with negative free cash flow is a fundamentally different value proposition.

The cracks beneath the concrete

The bull case for $690 billion in AI capex is straightforward: AI will transform every industry, demand for inference will grow exponentially, and the companies that own the most compute will capture the most value. The bear case is more nuanced, and it splits into two distinct failure modes that most analysts are conflating. Both deserve serious scrutiny, because the hyperscalers are not merely making a financial bet — they are making an architectural bet on the shape of computing for the next two decades, and the penalty for being wrong on architecture is far steeper than the penalty for being wrong on timing.

The first failure mode is demand disappointment — the possibility that enterprise AI adoption proceeds more slowly than the infrastructure buildout assumes. The hyperscalers are building capacity for a world where every Fortune 500 company runs thousands of AI agents, every consumer application integrates real-time inference, and every workflow generates continuous demand for GPU cycles. That world may arrive. But it has not arrived yet. TechCrunch’s year-ahead analysis argued that 2026 will be the year AI shifts from hype to pragmatism, with enterprises focusing on smaller models and reliable deployment rather than frontier capabilities. If the pragmatism thesis is correct — if most companies find that fine-tuned seven-billion-parameter models handle 80 percent of their use cases — then the demand for the massive GPU clusters the hyperscalers are building may not materialize on the timeline their capex budgets assume. The data centers will not sit empty. Cloud demand is growing regardless of AI. But the premium AI compute that justifies $200 billion in Amazon capex, specifically, needs AI-specific workloads to fill it at AI-specific prices.

The second failure mode is architectural obsolescence — the risk that the hardware being installed today becomes stranded by advances in chip design, model architecture, or inference optimization. This is not a hypothetical concern. Broadcom just projected $100 billion in AI chip revenue by 2027, driven entirely by custom silicon designed for specific workloads rather than general-purpose GPUs. Google is deepening its investment in Tensor Processing Units. Anthropic has committed to 1 gigawatt of Broadcom-designed TPUs in 2026, scaling to 3 gigawatts in 2027. OpenAI is building its first custom chip with Broadcom for delivery in 2027. Meta’s custom accelerator MTIA is progressing through its roadmap. If custom silicon delivers the efficiency gains its proponents claim — 3x to 10x better performance per watt on specific workloads — then the general-purpose Nvidia GPUs dominating today’s data centers could become the least efficient option within two years.

Meanwhile, the Nvidia challengers are flush with cash. MatX raised $500 million in a Series B led by Jane Street and Situational Awareness, the investment fund created by former OpenAI researcher Leopold Aschenbrenner, who has been arguing since 2024 that whoever controls the chip supply controls the future of AI. MatX, founded by two former Google hardware engineers, is building training chips that claim 10x the performance of Nvidia GPUs for LLM workloads, with volume production targeted for 2027 through TSMC. Dutch startup Axelera AI closed $250 million for edge inference chips the same week, backed by BlackRock and Samsung Catalyst Fund. The Register calculated that AI chip startups absorbed $1.1 billion in a single week — capital that is explicitly betting against the durability of Nvidia’s architecture advantage.

None of this means the hyperscaler buildout is foolish. It means it is a bet with a specific time horizon and specific assumptions about the shape of demand. If inference costs drop 10x over three years due to better chips — which is within the range of expert projections — then the data centers being built today may be generating only a tenth of the revenue per rack that current financial models assume. The hyperscalers can absorb that compression. Their shareholders may not feel as patient.

The AI infrastructure buildout of 2026 will be remembered either as the moment the industry laid the foundation for a multi-decade computing transformation or as the most expensive construction project since the fiber-optic bubble. The honest answer is that nobody — not Satya Nadella, not Sundar Pichai, not Jensen Huang — knows which narrative will prevail. What they know is that the cost of being wrong by building too much is survivable, while the cost of being wrong by building too little could be fatal. That asymmetry explains why every hyperscaler is racing to pour concrete even as their CFOs wince at the free cash flow projections.

For operators, investors, and technologists trying to navigate this landscape, the following signals matter most. First, watch the capex-to-AI-revenue ratio quarterly — if it compresses from the current 3.8x toward 2x by late 2027, the thesis is working, and if it stalls or expands, the overcapacity narrative gains teeth. Second, track the custom silicon timeline, because Broadcom’s projection of $100 billion in AI chip revenue by 2027 is the single most important data point for assessing whether Nvidia’s GPU monopoly holds or fragments, and every percentage point of market share that moves to custom silicon changes the economics of existing data centers. Third, monitor the bond market, where the hyperscalers have been absorbing debt at historically favorable rates, but the sheer volume — $45 billion in two months, with a trillion-dollar trajectory through 2027 — means that any widening of credit spreads will amplify the cost of the buildout in real time. Fourth, pay attention to enterprise adoption metrics across AWS, Azure, and Google Cloud, all of which report AI-specific revenue growth; if that growth rate decelerates for two consecutive quarters while capex continues rising, the market will reprice the entire sector.

The precedent that matters most is not the dot-com bubble. It is the buildout of the electrical grid in the early twentieth century. That infrastructure was overbuilt, wildly expensive, and took decades to generate adequate returns — but it ultimately transformed every industry and created more value than any other infrastructure investment in history. The hyperscalers are betting that AI compute is the electricity of the twenty-first century. They may be right. But the operators of that original electrical grid went through several cycles of bankruptcy, consolidation, and forced restructuring before the economics worked. The question is not whether AI infrastructure is valuable. It is whether the companies building it today will be the ones collecting the returns in 2035 — or whether a new generation of efficient, custom-silicon operators will inherit the demand that $690 billion in concrete and copper was meant to capture.

In other news

DeepSeek prepares V4, its most ambitious model yet — China’s DeepSeek is finalizing V4, a natively multimodal trillion-parameter model optimized for Huawei Ascend chips with a one-million-token context window. The release is timed to coincide with China’s annual Two Sessions parliamentary meetings and internal benchmarks suggest it could outperform Claude and ChatGPT on long-context coding tasks (PYMNTS).

Samsung targets 800 million Gemini-equipped devices — Samsung Electronics announced plans to double its footprint of mobile devices running Google’s Gemini AI, aiming for 800 million units by the end of 2026 by extending generative AI features to mid-tier and budget devices. The partnership deepens Google’s distribution advantage in the consumer AI race.

UK regulators demand answers from Grok — The UK’s Information Commissioner’s Office and Ofcom issued a formal information demand to Elon Musk’s xAI regarding its Grok model, signaling escalating regulatory scrutiny of AI systems operating in European markets.

Andrew Ng says AGI is decades away — AI pioneer Andrew Ng argued last week that the industry is decades from artificial general intelligence, defining AGI as AI matching the full range of human intellectual capabilities. The statement pushes back against timeline projections from labs that have AGI on their near-term roadmaps.

Anthropic finds 22 Firefox vulnerabilities — In a security partnership with Mozilla, Anthropic’s AI systems discovered 22 separate vulnerabilities in Firefox, 14 classified as high-severity, demonstrating the growing utility of AI models in automated security auditing (Anthropic).

Caitlin Kalinowski exits OpenAI over Pentagon deal — Hardware executive Caitlin Kalinowski resigned from OpenAI’s robotics team in protest of the company’s controversial agreement with the Department of Defense, becoming the highest-profile departure linked to the Pentagon AI deals that have roiled the industry since late February.