A bubble occurs when asset prices detach from underlying value and then collapse. The challenge is that we can only confirm a bubble in hindsight—at the time, elevated prices might reflect genuine value or speculative excess, and distinguishing between them is the challenge.
The standard approach is to examine demand and supply separately. On the demand side: is there real economic value being created, or are people buying assets purely because they expect prices to rise? On the supply side: is capacity being financed in ways that can survive if assumptions change, or is the capital structure fragile?
Bubbles can form from either side. Tulip mania and Crypto 2021 were demand-side bubbles—speculative buying untethered from any use value. The 2008 housing crisis was partly a supply-side phenomenon—real demand for housing existed, but the financing structures couldn’t survive a modest price decline. The dotcom bust had elements of both: real demand for internet connectivity, but overbuilt capacity financed on revenue projections that never materialized.
For AI infrastructure, the demand side is the easy part. The evidence for real economic value is substantial and growing. The interesting question is on the supply side: whether the capital structure financing the buildout can survive the assumptions embedded in it.
Demand: Where the Evidence Points
The case for genuine demand rests on observable business impact, not projected future value.
What these numbers represent is companies disclosing AI-driven financial impact in earnings calls—statements made under securities law, not marketing materials. The tripling of companies reporting AI benefits since early 2023 reflects actual deployment reaching the income statement. When Walmart says GenAI replaced work that would have required 100x the headcount, or Estée Lauder reports 31% ROI improvement on media campaigns, these are claims with consequences for executives making them. The investor relations teams and other internal teams vet and validate what executives say publicly.
Consumer adoption data reinforces this. Over half of US adults have used GenAI, with workplace usage nearly doubling within a year. At 900+ million weekly active users, ChatGPT alone has reached a scale where “will people use this?” is no longer a relevant question.
One way to contextualize where AI sits developmentally: think of it like human capital formation. Humans absorb roughly $350,000 in rearing and education costs over 20 years, then produce approximately $2.8 million in lifetime earnings over a 40+ year career. AI spent 2022-2025 in its equivalent “education phase”—massive investment in training foundation models. We’re now entering the “working phase” where deployment generates returns. The important difference: humans retire and their knowledge doesn’t transfer perfectly. AI systems can operate indefinitely and improve continuously.
This framing helps explain why the traditional bubble indicators aren’t flashing red. Economists look at five macro signals that preceded the dot-com crash: extreme valuations, declining corporate profits, elevated credit risk, high cash burn, and rising leverage. These tend to move in sequence—valuations run hot first, then fundamentals deteriorate, then credit markets crack.
The current picture shows valuations elevated but not extreme (95th percentile vs. the 99th percentile at the March 2000 peak), with the other four indicators in healthy territory. This is what you’d expect if demand is real rather than speculative: companies can support elevated valuations because underlying profits are strong, not because everyone is buying in anticipation of selling to a greater fool.
The demand side of the AI market doesn’t look like a bubble. It looks like a genuine technological transition with measurable economic impact.
Supply: What’s Being Built
Meeting this demand requires building across the entire AI value chain—from semiconductor fabrication through data center infrastructure to the software layers that deliver AI to end users.
The chain spans multiple industries with different economics. Semiconductor fabrication is concentrated (TSMC, Samsung) and capital-intensive. Data center buildout requires coordinating power systems, servers, networking, and physical facilities. Cloud providers and operators sit in between, purchasing compute capacity and selling it to software companies and enterprises. Each layer depends on the others. Constraints anywhere ripple through the system.
The buildout currently faces real bottlenecks. In a survey of 149 senior data center professionals, 92% cited utility capacity as a barrier, with 44% facing 4+ year wait times for grid connections. Permitting issues affect 86%, fiber availability 85%, chip access 81%.
The industry response has been to route around constraints rather than wait for them to clear. 62% now plan on-site power generation as primary contingency—a shift from treating utilities as default to treating them as backup. When Elon Musk built Colossus 1, he brought in 35 semi-trailer-sized gas turbines. For Colossus 2, he bought a power plant overseas, disassembled it, and shipped it to the US for reassembly. These reflect how tight the energy market has become.
These constraints explain why shortage pricing exists today. The question for financing is whether they persist long enough to support the debt structures being built around them.
Supply: The Financing Structure
The AI buildout requires approximately $2.9 trillion in data center capital expenditure through 2028. The hyperscalers are spending aggressively and so is OpenAI. CEO Sam Altman references $1.4 trillion in compute commitments over eight years. Dario Amodei describes ambitions for hundred-billion-dollar training clusters. Google’s infrastructure leadership talks about doubling compute every six months.
Hyperscalers’ operating cash flows can fund roughly half of this. The remainder—about $1.5 trillion—requires external capital. This isn’t unusual. Large infrastructure buildouts routinely rely on outside financing. What matters is whether that financing is structured appropriately for the assets it’s funding.
The capital stack breaks down as follows: $200 billion from corporate debt issuance (companies borrowing against their own balance sheets), $150 billion from securitized assets (loans packaged and sold to investors), $800 billion from private bilateral credit (direct loans from non-bank lenders), and $350 billion from other sources including private equity.
The $150 billion in securitized assets and $800 billion in private credit deserve particular attention because of their scale and characteristics. Private credit consists of loans extended by asset managers, pension funds, and specialty lenders rather than traditional banks. Private credit has grown dramatically over the past decade, partly because post-2008 banking regulations pushed riskier lending outside the regulated system. The loans are typically held to maturity rather than traded, which means problems don’t surface until borrowers actually struggle to pay. Securitized assets carry a different risk: loans packaged and sold to dispersed investors who may not fully understand the underlying collateral dynamics—a structure that should sound familiar from 2008.
The nearly $1 trillion in external debt expected to finance data centers through 2028 carries a specific assumption: that the underlying assets—primarily GPUs and related infrastructure—will retain enough value over a 5-6 year loan term to support the debt. This assumption is standard for infrastructure lending, where assets like buildings, fiber optic cables, or aircraft depreciate slowly and predictably over decades.
GPUs don’t work that way.
The Depreciation Problem
When lenders finance infrastructure, they model whether cash flows provide enough buffer to service the debt. In a worst-case default scenario, they assess how much they can recover by liquidating the assets. Ideally, loan-to-asset value stays below 100% so lenders can recoup their principal.
For most infrastructure, this is straightforward. Commercial buildings depreciate over 30-40 years. Fiber optic cable lasts 25-30 years and, because improvements happen in endpoint equipment rather than the cable itself, the same physical glass can carry more data over time as technology advances. Aircraft operate for 20-25 years. These timelines give lenders comfortable cushion against typical 5-year loan terms.
GPUs follow a different pattern. NVIDIA releases a new architecture roughly every 12-18 months, and each generation delivers 2-4x better performance per watt than its predecessor. Unlike fiber, where new technology makes existing infrastructure more valuable, new GPU technology makes existing chips obsolete. Operators must replace the GPUs themselves.
The practical result: GPUs have competitive lives of 2-4 years before they’re outperformed enough to lose pricing power. An H200 purchased today will compete against B200 chips delivering twice the performance. Within 2-3 years, it will face Rubin delivering nearly 9x the compute—an efficiency gap that makes older hardware economically uncompetitive.
This matters for financing because the loans assume 5-year asset lives while the assets themselves become economically obsolete in 2-4 years. The only way the math works is if shortage pricing persists—if demand so exceeds supply that even older GPUs command premium prices throughout the loan term.
How the Financing Breaks
To see where the stress point lies, consider the economics of a GPU-based data center at different pricing scenarios.
The table models a single-GPU data center using a financing structure similar to CoreWeave’s: 85% debt, 10% interest rate, 5-year amortization. Three price points matter:
Market price is what operators currently charge customers for compute, set by supply and demand in today’s constrained environment.
Breakeven price is the minimum rate needed to cover debt service plus operating expenses (primarily electricity). Below this, the operator loses money and defaults on debt.
Equilibrium price is what pricing would look like if supply caught up with demand and GPUs competed purely on performance per dollar. At equilibrium, older chips must price at parity with newer chips on a $/TFLOPS basis, which means steep discounts to their original cost.
At today’s market prices, all GPU generations show healthy margins. This shortage premium is what makes the leveraged financing structure work.
At equilibrium pricing, the picture fractures. H200s and newer chips still cover their costs. But A100s fall deep into negative margins—well below breakeven. Operators holding A100 debt would be unable to service their loans from operating income.
The key question is how long shortage pricing persists. Risky loans in the private credit and securitized asset stack is a bet that equilibrium arrives after the debt matures. If equilibrium arrives earlier, operators face a choice between default and refinancing at unfavorable terms—and refinancing requires finding new lenders willing to take the same bet on even older equipment or invest more money to purchase the latest GPUs.
NVIDIA’s excellence is, paradoxically, the problem. The company’s relentless execution—B200 ramping now, Rubin shipping in late 2026, Rubin Ultra in 2027—means product cycles have compressed to roughly 18 months for a new generation. Each generation delivers dramatic efficiency improvements. NVIDIA’s engineering prowess might be driving the obsolescence cycle faster than lenders modeled.
The hyperscalers themselves are starting to acknowledge this. In January 2025, Amazon reversed course—shortening server depreciation from six years back to five. The company’s 10-K was explicit about why:
In 2024, we also determined, primarily in the fourth quarter, to retire early certain of our servers and networking equipment. We recorded approximately $920 million of accelerated depreciation and related charges for the quarter ended December 31, 2024 related to these decisions. The accelerated depreciation will continue into 2025 and decrease operating income by approximately $0.6 billion in 2025. These two changes above are due to an increased pace of technology development, particularly in the area of artificial intelligence and machine learning.
This is notable because it goes against the trend of other hyperscalers: Meta, Alphabet, and Microsoft have extended depreciation schedules from 3 years to 5 to 6 years since 2020, Amazon looked at the same data and moved the opposite direction.
Two Forces Accelerating Supply Equilibrium
Several developments are accelerating the path toward compute equilibrium, each working through a different mechanism but producing the same effect: more compute capacity available per dollar, which erodes shortage pricing.
Algorithmic efficiency is compounding faster than most models assumed. DeepSeek provides the clearest data points. Their V3 model, released December 2024, trained with 2.6 million GPU hours compared to 30.8 million for Meta’s Llama 3 405B—roughly 12x more efficient for comparable benchmark performance. The headline “$5.5 million training cost” is debatable, but the efficiency differential was real and reflected genuine architectural innovations.
By December 2025, their V3.2 release achieved results that benchmark against GPT-5: gold medal performance at the International Mathematical Olympiad, top-10 at the International Olympiad in Informatics, second place at ICPC World Finals (programming). Cost to run the Artificial Analysis benchmark is $54 vs Grok 4’s $1,900—35x cheaper for better results. On January 2, 2026, DeepSeek published a paper introducing “Manifold-Constrained Hyper-Connections,” a training architecture that addresses scaling instability—one of the key obstacles to training ever-larger models efficiently. Expect this technique to appear in their next major release.
The implication for infrastructure financing: efficiency gains reduce compute demand per unit of AI capability. If capability-per-dollar improves 10x annually, the compute needed to serve a given market shrinks proportionally. That’s positive for AI adoption and negative for GPU pricing power. Shortage premiums compress not because demand falls, but because less supply is needed to meet it.
Energy constraints may ease faster than expected. Power availability is the most commonly cited bottleneck for data center expansion, but the nuclear pipeline is materializing on an aggressive timeline. In the past 18 months, hyperscalers signed agreements for over 10GW of new nuclear capacity: Microsoft’s $16 billion deal to restart Three Mile Island by 2028, Google’s 500MW Kairos Power small modular reactor (SMR) fleet targeting 2030, Amazon’s $20+ billion Susquehanna campus conversion plus 5GW X-energy SMR pipeline. Regulatory timelines compressed after May 2025 presidential executive orders mandated 18-month maximum Nuclear Regulatory Commission (NRC) reviews for new reactors, down from historical 7-year processes.
If power becomes abundant sooner than expected, the primary constraint on supply expansion disappears. More data centers can be built and operated, which increases available compute and pressures pricing toward equilibrium. The power scarcity premium embedded in current rates erodes.
Both forces point the same direction: toward compute abundance arriving sooner than lenders modeled. Efficiency shrinks demand per unit of capability. Energy expansion grows potential supply. For AI adoption, these are positive developments. For financing structures betting on persistent scarcity, they compress the timeline.
Evidence of Stress in Specific Companies
Broad credit markets remain healthy—spreads are tight, capital is available, no systemic distress is visible. But within the specific segment of data center and GPU financing, stress signals have emerged.
CoreWeave’s stock has declined approximately 60% from peak levels. The company exemplifies leveraged GPU economics. Financials from their Q3 2025 report:
+ $997 million gross profit
- $631 million depreciation
- $310 million interest expense
- $1,412 million quarterly debt repayment
+ $3,129 million in new debt issuance
I believe the stock price reflects investor reassessment of whether shortage pricing will persist long enough for the model to work.
Oracle’s situation is different but related. The company maintains investment-grade credit ratings, but its bonds due 2035 trade at yields closer to BB-rated (junk) than BBB-rated (investment grade) paper. CDS spreads have reached levels last seen during the 2009 financial crisis. The market is pricing specific risk around Oracle’s $248 billion in off-balance-sheet data center commitments—obligations that don’t appear on the balance sheet but represent real financial exposure.
To be clear: the hyperscalers will survive this. When Alphabet extended depreciation to six years in 2023, it boosted pre-tax income by $3.9 billion. When the writedowns eventually come, Alphabet absorbs an accounting hit—painful for the stock, but not existential. Alphabet has $96 billion in cash. The same applies to Microsoft, Meta, and Amazon. These companies have fortress balance sheets. They can afford to be wrong about depreciation schedules.
The risk sits one layer down: in the operators who financed GPU acquisitions with debt rather than cash flow. When CoreWeave faces the same depreciation math, the $310 million in quarterly interest payments don’t pause for accounting adjustments. The hyperscalers take writedowns. The leveraged operators can’t service their debt.
These aren’t signs of a broad bubble. Credit conditions overall remain accommodative. What they indicate is targeted concern about the gap between GPU economic lives and the financing structures built around them. The stress is appearing exactly where the depreciation thesis predicts—in leveraged operators holding older equipment against long-dated debt.
What This Is and Isn’t
The AI infrastructure situation doesn’t fit the pattern of a classic demand-side bubble. Demand is real, growing, and generating measurable economic returns. Traditional macro indicators show one warning sign (elevated valuations) against four healthy readings (profits, credit spreads, cash burn, leverage). The comparison to tulip mania or dot-com speculation mischaracterizes what’s happening.
The risk is narrower and more technical: financing structures built on the assumption that GPUs will retain economic value—the ability to generate cashflow—for 5-years, when competitive obsolescence arrives in 2-4 years. This works only if shortage pricing persists—and several forces are working to end shortage pricing sooner than the financing models anticipated.
If the risk materializes, the result wouldn’t look like a market-wide crash. It would look like credit impairments concentrated in GPU-backed lending, distress among leveraged data center operators, and margin compression for providers running older hardware. The stress would likely surface in credit markets first—in private loan defaults, securitized asset writedowns, and distressed refinancings—before rippling into equity markets as the hyperscalers eventually take the writedowns. This isn’t a prediction that stock prices won’t fall. It’s a view on sequencing: credit cracks first, equity follows. CoreWeave’s stock and Oracle’s bonds suggest markets are beginning to price this scenario.
The investors exposed to this risk are not the ones betting on AI. Its the ones betting on scarcity.
A Note on Michael Burry’s AI Thesis
Michael Burry —the investor who identified the subprime mortgage crisis before it unfolded and the person whom I’ve watched & read about a lot when I started my finance career—has been publishing detailed analysis of the AI buildout on his Substack, Cassandra Unchained. Having now gone through his writing carefully, we’re seeing the same thing—where we differ are more about emphasis than fundamental disagreement.
Burry’s analysis rests on several pillars:
Capital Cycle Theory. Burry presents historical data showing that stock market peaks tend to occur mid-buildout, often before capital expenditures peak. His chart of S&P 500 net capital investment divided by GDP shows this pattern across the dot-com boom, housing bubble, and shale revolution. The telecom bust wasn’t about demand being fake—internet demand was real. The problem was supply massively overshooting: by 2002, less than 5% of the fiber infrastructure built during the bubble was actually lit.
Depreciation mismatch. Hyperscalers have extended depreciation schedules from 3 years (2020) to 5-6 years today, while NVIDIA’s product cycles have compressed to roughly one year. This inflates earnings and overstates asset values. Burry documents this from 10-K filings and estimates double-digit billions of earnings overstatement for hyperscalers over 2026-2028. When Alphabet extended to 6 years in 2023, it boosted pre-tax income by $3.9 billion.
Economic obsolescence. Burry distinguishes physical utilization from economic value creation. NVIDIA’s CFO noted that A100s from six years ago are “fully utilized.” Burry’s response: “Just because a widget is used does not mean the widget is profitable to a degree that it is worth more than residual value.” The A100 uses 2-3x more power per FLOP than the H100. Blackwell is reportedly 25x more energy efficient than H100. An older chip can be running at full capacity and still be economically unviable.
Demand relative to CAPEX. Burry cites Menlo Ventures data showing application-layer GenAI revenue at $37 billion this year versus $400 billion in chip spending. ChatGPT has 900 million weekly users, but only 5% pay. His concern isn’t that AI has no demand—it’s that supply is being built far in excess of what current revenue can support. As he puts it: “One can believe AI is transformational and also believe that some public and private AI stocks are very overvalued.”
Credit risk. This is where our analyses converge most directly. In “Unicorns and Cockroaches,” Burry writes: “The big target for private credit now is AI data centers... there is a duration mismatch of catastrophic proportions between the asset and the loan.” In his Q&A piece, he adds: “There is a lot of leverage behind the data center buildout, and I believe bank and non-bank entities will be at risk if that buildout busts.”
Burry is clear about what he’s not saying: “I do not predict that Nvidia, Meta, Microsoft, Amazon, Alphabet are doomed. They all should survive. I believe some will survive like 2000 Microsoft did, some will survive like 2000 Cisco did.”
Where Burry and I align:
Where we differ in emphasis:
On demand, we weight the evidence differently. Burry emphasizes consumer app revenue and conversion rates. I weight enterprise deployment—the S&P 500 earnings disclosures are claims made under securities law, which I find harder to dismiss. But if he’s right that demand is weaker than I assess, that makes the credit thesis worse, not better. Less demand means faster arrival at equilibrium pricing, which means faster erosion of the shortage premium that makes the financing structures work.
One area Burry has developed that I haven’t fully explored: circular financing. He describes how tech giants invest in AI startups who then buy from them, and how “off balance sheet special-purpose vehicles” and “badly structured but easily obtained financing” are goosing GPU demand. To the extent this is true, it amplifies credit risk—loans underwritten against demand that’s partially self-referential are even more exposed when capital stops flowing.
One final thought. Burry made his name by poring through individual mortgage loan files—finding the misaligned incentives and mispriced risk that aggregate data obscured. He saw what the rating agencies and institutional investors missed because he did the granular work.
The same opportunity exists here. The nearly $1 trillion in private credit and securitized assets projected to finance GPU infrastructure contains specific assumptions about depreciation schedules, collateral valuations, residual values, and covenant structures. Someone with access to those documents—the actual loan files, not the marketing materials—could identify which pools are most exposed and how the risk is distributed.
Burry notes he plans to address financing, circular deals, and “a curious liability hiding in the financial statements” in his Part 3. I’ll be reading carefully.
That work remains to be done. I’m curious what he finds.













Thanks for the cogent reporting on this. I think my biggest question is about the depreciation mismatch portion: that is, H200s purchased today will compete against Rubin chips delivering ~9x the compute within 2-3 years
In an inference and energy constrained system, does this really matter? Even if you're competing against more efficient chips, aren't you still making money if there's an ongoing demand for your product?
Seems like the problem only appears if supply catches up with demand. If demand exceeds total available compute, then doesn’t even less efficient chips command premium pricing? Seems like for the financing to break, equilibrium pricing must arrive before the debt matures.
Or am I missing something?
Missing the point entirely. AI investment is mostly about who gets to provide compute to the Western world . Only 3 companies have the legitimate power to furnish tokens to the western world. It’s about tokens . And the volume of tokens has exploded. Msft , goog and AWS , are the only capable providers of compute. The rest is small share and therefore noise .. think about a world where only 3 providers have the heft to provide most compute. Oligopoly confirmed. Oligopolistic rents are confirmed.
All this talk about financing is silly . This is the biggest economic transformation ever .. token volumes will explode as we make entire movies from a single context window..