TBPN
← Back to Blog

Why Big Tech Is Spending Like the AI Bubble Will Never End

Big Tech is pouring over $300B into AI infrastructure annually. Data centers, chips, and energy: is this rational investment or bubble behavior? Full analysis inside.

Why Big Tech Is Spending Like the AI Bubble Will Never End

In the first quarter of 2026, Microsoft, Google, Amazon, and Meta collectively reported capital expenditure guidance exceeding $320 billion for the year. Three hundred and twenty billion dollars. That is more than the GDP of Finland. It is roughly the entire market capitalization of Walmart. And nearly all of it is going to one thing: artificial intelligence infrastructure — data centers, GPU clusters, custom silicon, networking equipment, and the electricity to power it all.

These numbers have crossed a threshold where they are no longer just corporate finance stories. They are macroeconomic events. They are reshaping the energy grid, the global semiconductor supply chain, and the construction industry. And they are raising a question that we discuss on the Technology Brothers Podcast Network almost daily: is this the smartest capital allocation in the history of technology, or is this the biggest bubble since the dot-com era?

The honest answer is that it might be both.

The Numbers: What Big Tech Is Actually Spending

The 2026 Capex Supercycle

Let us start with the specific numbers, because they are genuinely staggering:

  • Microsoft: Guided $85-90B in capital expenditure for fiscal year 2026 (ending June 2026), the majority allocated to Azure data centers and AI infrastructure. This is up from approximately $55B in FY2025 and $44B in FY2024.
  • Google (Alphabet): Guided $75B+ for calendar year 2026, focused on TPU v6 production, data center expansion, and cloud infrastructure. CFO Ruth Porat explicitly called this "the most significant infrastructure investment in Alphabet's history."
  • Amazon (AWS): Guided $80B+ for 2026, with AWS data center expansion across North America, Europe, and Asia. Amazon is building more data center capacity in 2026 than it built in all years prior to 2023 combined.
  • Meta: Guided $55-65B for 2026, focused on AI training clusters for Llama model development and inference infrastructure for AI features across Facebook, Instagram, WhatsApp, and the metaverse platform.

Combined, that is approximately $300-320 billion in a single year. For context, the entire United States spent approximately $30 billion on data center construction in 2020. Big Tech is now spending ten times that amount annually.

Where the Money Is Going

The capital expenditure breaks down into four major categories:

  1. Data center construction ($100-120B): Physical buildings, cooling systems, networking infrastructure, land acquisition, and permitting. Companies are building campuses with 500MW+ power capacity — enough electricity to power small cities.
  2. Compute hardware ($120-140B): Primarily NVIDIA GPUs (H200, B200, and the new Blackwell Ultra), along with custom silicon (Google TPU v6, Amazon Trainium3, Microsoft Maia 2, Meta MTIA v3). This category alone represents more revenue than NVIDIA's entire company generated in 2023.
  3. Networking and interconnects ($30-40B): High-bandwidth networking to connect GPU clusters within and across data centers. InfiniBand, Ethernet at 800Gbps+, and custom optical interconnects.
  4. Energy infrastructure ($20-30B): Power generation, transmission, and storage. This includes investments in nuclear power (small modular reactors), natural gas plants, solar farms, and battery storage. Several Big Tech companies are now among the largest energy buyers in the United States.

The Chip Wars: NVIDIA, AMD, and Custom Silicon

NVIDIA's Dominance and Its Challengers

NVIDIA remains the undisputed king of AI compute. The company's data center revenue exceeded $130B in fiscal year 2026, driven by the H200 and B200 GPU generations. NVIDIA's competitive moat — the CUDA software ecosystem — remains formidable. Virtually every AI framework, library, and training pipeline is optimized for CUDA. Switching away from NVIDIA is not a hardware decision; it is a software migration that most organizations cannot afford.

But the challengers are making progress:

  • AMD: The MI350 accelerator, shipping in mid-2026, represents AMD's most competitive AI chip yet. AMD has been gaining traction with cost-sensitive buyers and in inference workloads where CUDA dependency is less critical. AMD's AI data center revenue is projected to exceed $15B in 2026.
  • Google TPU v6: Google's sixth-generation Tensor Processing Unit is competitive with NVIDIA's B200 on transformer training workloads and offers significantly better cost-per-FLOP for Google Cloud customers. Google uses TPU v6 internally for Gemini training and offers it to external customers through Cloud TPU.
  • Amazon Trainium3: Amazon's custom AI chip for AWS, designed to offer better price-performance than NVIDIA GPUs for common training and inference workloads. Adoption has been slower than Amazon hoped, but Trainium3's improved performance and Anthropic's use of Trainium for Claude training provide credibility.
  • Custom silicon from xAI: Elon Musk's xAI is reportedly developing its own AI accelerator, leveraging expertise from Tesla's Dojo program. Details are scarce, but the Memphis data center's planned 200,000+ GPU cluster suggests xAI is serious about becoming a compute infrastructure player.

The Supply Chain Pressure

The aggregate demand for AI chips is straining the global semiconductor supply chain. TSMC, which manufactures chips for NVIDIA, AMD, Apple, and most other major chip designers, is running its advanced nodes (3nm, 2nm) at near-maximum capacity. Lead times for AI accelerators remain at 6-12 months, and companies that failed to place orders early are finding themselves at the back of a very long queue.

This supply constraint creates a self-reinforcing spending dynamic: companies order more chips than they immediately need because the alternative — not having compute when they need it — is worse. This "hoarding" behavior inflates demand signals, causes TSMC to invest in more capacity, and creates a cycle that looks sustainable as long as AI demand keeps growing and looks like a bubble if demand plateaus.

The Energy Crisis No One Is Talking About

AI's Electricity Appetite

The most underreported story in the AI boom is its energy consumption. Training a single frontier AI model — GPT-5.5, Claude 4.5 Sonnet, Gemini 2.5 — consumes an estimated 50-100 gigawatt-hours of electricity. For reference, that is the annual electricity consumption of approximately 5,000-10,000 American homes. And these models are being trained multiple times per year as companies iterate on architectures and datasets.

Inference — actually running the models to serve user requests — consumes even more electricity in aggregate. OpenAI reportedly uses over 1 gigawatt of continuous power just for inference serving. At current growth rates, the AI industry's total electricity consumption could reach 3-5% of total US electricity generation by 2028.

The Nuclear Renaissance

Big Tech's response to the energy challenge has been, characteristically, to throw money at the problem. And the solution they are converging on is nuclear power:

  • Microsoft: Signed a deal to restart the Three Mile Island Unit 1 reactor and has invested in multiple small modular reactor (SMR) companies.
  • Google: Announced partnerships with Kairos Power for SMRs and has signed the largest corporate nuclear power purchase agreement in history.
  • Amazon: Investing in X-energy and other SMR developers, with plans to deploy nuclear-powered data centers by 2030.
  • xAI: The Memphis data center is reportedly exploring on-site nuclear power for its massive GPU cluster.

The nuclear investments are long-term bets — SMRs will not produce meaningful power before 2029-2030 at the earliest. In the meantime, Big Tech is consuming available grid power, building natural gas plants as bridge solutions, and purchasing renewable energy credits at a pace that has reshaped the clean energy market.

The Bull Case: Why This Spending Is Rational

AI Revenue Is Real and Growing

The strongest argument for the current spending levels is that AI revenue is not hypothetical — it is real, growing fast, and showing no signs of plateauing:

  • Microsoft's AI revenue: Azure AI services reportedly generate over $20B annually, with growth exceeding 60% year-over-year. GitHub Copilot alone generates approximately $2B annually.
  • Google's AI revenue: Gemini API, Cloud AI services, and AI-enhanced advertising contribute an estimated $25B+ annually. Google's core search advertising business also benefits from AI through improved ad targeting and generation.
  • Amazon's AI revenue: AWS AI services, including Bedrock (which hosts Claude, Llama, and other models), contribute an estimated $15B+ annually.
  • Meta's AI revenue: AI-powered ad targeting and content recommendations drive improvements across Meta's $160B+ annual advertising revenue. Internal estimates suggest AI improvements contribute $20-30B in incremental ad revenue annually.

These revenue numbers provide genuine economic justification for the infrastructure spend. If Microsoft spends $85B on AI infrastructure and generates $20B+ in annual AI revenue growing 60%+, the investment will generate positive returns within 3-5 years — well within the useful life of the data centers and equipment.

The Compute Demand Curve

The bull case also rests on the observation that AI compute demand has consistently outpaced supply. Every prediction about AI compute demand over the past three years has been too conservative. Companies that invested aggressively in compute have been rewarded — those that did not have lost market share and developer mindshare. The lesson the market has internalized is that under-investing in AI compute is more dangerous than over-investing.

The Bear Case: Why This Could Be a Bubble

Revenue Growth vs. Capex Growth

The bear case begins with a simple observation: capital expenditure is growing faster than revenue. Microsoft's AI capex roughly quadrupled from $22B in FY2023 to $85B+ in FY2026. Its AI revenue, while impressive, has not quadrupled in the same period. The gap between investment and returns is widening, not narrowing.

This gap is sustainable as long as revenue growth accelerates. If it does not — if AI revenue growth decelerates from 60%+ to 30% or less — the companies will be sitting on massive infrastructure assets generating insufficient returns. History is littered with examples of this dynamic: telecoms in the late 1990s, fiber optic companies in 2001, crypto mining farms in 2022.

The Model Efficiency Question

A potentially disruptive counter-narrative is that AI models are getting more efficient faster than most projections assume. If a model 50% as expensive to run delivers 90% of the capability of a frontier model — which is approximately what GPT-5.5 Mini and Claude Haiku achieve today — then the demand for massive compute clusters may plateau sooner than the hyperscalers expect.

The DeepSeek episode in early 2025, when a Chinese AI lab demonstrated training a competitive model at a fraction of the expected cost, briefly spooked AI infrastructure investors for exactly this reason. If compute efficiency improves faster than demand grows, the tens of billions spent on GPU clusters become stranded assets.

The Customer Concentration Risk

A significant portion of AI infrastructure revenue comes from a small number of very large customers. Anthropic alone reportedly spends $2-3B annually on cloud compute. OpenAI spends a similar amount. If any of these major AI companies reduces spending — due to a funding crunch, a strategic pivot, or improved training efficiency — the impact on cloud revenue would be material.

The xAI and SpaceX Wild Card

Elon Musk's Compute Empire

xAI's approach to the AI infrastructure race is characteristically Musk: move fast, build big, and worry about demand later. The Memphis data center, reportedly housing 200,000+ NVIDIA H100 GPUs with plans to expand to next-generation B200 chips, represents one of the largest single-site compute clusters in the world. xAI has also reportedly secured additional data center capacity in Austin, Texas, and is exploring sites in the Middle East.

SpaceX's role in this story is often overlooked. Starlink provides xAI with a potential distribution channel for AI inference at the edge — running AI models on Starlink ground stations to provide low-latency AI services globally. While this is speculative, it illustrates how Musk's companies create synergies that competitors cannot easily replicate.

What This Means for the Broader Economy

The Infrastructure Multiplier

Big Tech's AI infrastructure spending does not exist in a vacuum. Every data center creates construction jobs, generates demand for steel and concrete, requires electricians and HVAC technicians, and increases electricity demand for local utilities. The economic multiplier effect of $300B+ in annual AI infrastructure investment is significant — potentially adding 0.3-0.5% to US GDP growth annually.

This multiplier effect is one reason why policymakers have been relatively supportive of the AI buildout despite concerns about market concentration. AI infrastructure investment creates broadly distributed economic benefits even if the profits are concentrated among a handful of companies.

The Real Estate and Energy Implications

Data center demand is reshaping commercial real estate markets in Northern Virginia, Oregon, Texas, and other data center hubs. Land prices near major power substations have increased 3-5x in the past two years. Local governments are racing to attract data center investment with tax incentives, expedited permitting, and infrastructure subsidies.

The energy implications are even more profound. Utilities in data center-heavy regions are facing unprecedented demand growth after decades of flat or declining electricity consumption. Some utilities have paused new data center connections because their grids cannot support additional load without billions in transmission upgrades. The AI boom is forcing a reckoning with America's aging energy infrastructure that no climate policy or EV mandate was able to catalyze.

TBPN's Framework: How to Think About AI Spending

On the Technology Brothers Podcast Network, we have developed a framework for evaluating the AI capex supercycle that avoids both uncritical optimism and reflexive bubble-calling. We call it the "3R" framework: Revenue, Returns, and Resilience.

  • Revenue: Is AI generating real, measurable revenue? Yes — at least $80-100B annually across the major companies, growing 40-60%+ per year. This is not speculative demand.
  • Returns: Will the infrastructure investments generate positive returns over their useful life (7-10 years for data centers, 3-5 years for GPUs)? Probably, if revenue growth continues at even half the current rate. Possibly not, if efficiency improvements dramatically reduce compute demand.
  • Resilience: Can the companies absorb a downturn if AI revenue growth decelerates? Yes — Microsoft, Google, Amazon, and Meta all have diversified revenue streams and healthy balance sheets. A slowdown in AI returns would hurt stock prices but would not threaten corporate solvency.

The 3R framework suggests that the current spending is rational but risky — more analogous to the infrastructure buildout that preceded the mobile internet boom (which ultimately justified the investment) than the fiber optic buildout that preceded the dot-com bust (which did not). The key variable is whether AI applications continue to generate growing revenue, and on that question, the evidence today is strongly positive.

Get your TBPN tumbler, throw on a TBPN t-shirt, and tune in to our daily analysis of the AI spending race. The numbers get bigger every quarter, and we are here to help you make sense of them.

Frequently Asked Questions

Is the AI spending boom a bubble?

The spending boom has characteristics of both rational investment and speculative excess. On the rational side, AI is generating real revenue ($80-100B+ annually across major tech companies), growing rapidly (40-60%+ annually), and showing clear product-market fit in enterprise, developer tools, and consumer applications. On the speculative side, capital expenditure is growing faster than revenue, companies are building capacity ahead of demand, and there is a herd mentality where no company wants to be seen as under-investing. Our assessment is that the current spending is more likely justified than not, but the margin for error is thin. If AI revenue growth decelerates significantly (below 20% annually), the overinvestment will become apparent. If growth continues at 40%+, the current spending will look prescient.

Why is energy such a big concern for AI?

AI model training and inference are extraordinarily energy-intensive. A single large data center can consume 500MW+ of continuous power — equivalent to a medium-sized city. The aggregate power demand from AI data centers is projected to reach 3-5% of total US electricity generation by 2028. This creates three problems: (1) many electrical grids lack the capacity to serve new data center loads, (2) the carbon footprint of AI at scale is significant, and (3) competition for power between data centers and other users is driving up electricity prices in some regions. Big Tech is investing heavily in nuclear power, renewables, and energy storage to address these challenges, but the solutions are years away from deployment at scale.

What happens if NVIDIA loses its AI chip monopoly?

NVIDIA's dominance rests on two pillars: hardware performance and the CUDA software ecosystem. If competitors (AMD, Google TPU, Amazon Trainium, custom silicon from xAI and others) match NVIDIA's hardware performance, CUDA remains a significant moat — most AI software is written for CUDA, and migration costs are substantial. However, the industry is investing heavily in hardware-agnostic AI frameworks (PyTorch 3.0, JAX, Triton) that could reduce CUDA dependency over time. A realistic scenario is that NVIDIA's market share gradually declines from 80%+ to 50-60% over the next 3-5 years as alternatives mature, while the overall market grows large enough that NVIDIA's absolute revenue continues increasing. A sudden collapse of NVIDIA's position is unlikely absent a major architectural disruption.

Should investors be worried about Big Tech capex levels?

The answer depends on your time horizon. Short-term (1-2 years), high capex levels depress free cash flow and create earnings pressure, which can weigh on stock prices. Long-term (5-10 years), the infrastructure being built today will generate revenue for decades — data centers have useful lives of 15-20+ years. The historical analogy is Amazon's early years, when massive capital investment in logistics infrastructure depressed earnings for years but ultimately created an unassailable competitive moat. The risk is that AI infrastructure does not generate the same long-term returns as logistics infrastructure. Investors should monitor the ratio of AI revenue to AI capex — as long as revenue growth outpaces capex growth, the investments are accretive. If that ratio inverts, caution is warranted.