Formerly known as Wikibon

298 | Breaking Analysis | Resetting GPU Depreciation — Why AI Factories Bend, But Don’t Break, Useful Life Assumptions

Much attention has been focused in the news on the useful life of GPUs. While the pervasive narrative suggests GPUs have a short lifespan, and operators are “cooking the books,” our research suggests that GPUs, like CPUs before, have a significantly longer useful life than many claim. In this Breaking Analysis we use the infographic below to explain our thinking in more detail. 

Premise

In January 2020, Amazon changed the depreciation schedule for its server assets, from three years to four years. This accounting move was implemented because Amazon found that it was able to extend the useful life of its servers beyond three years. Moore’s Law was waning and at Amazon’s scale, it was able to serve a diverse set of use cases, thereby generating revenue out of its EC2 assets for a longer period of time. Other hyperscalers followed suit and today, the big three all assume six year depreciation schedules for server assets. 

The question is, in the AI factory era, will this same dynamic hold or will the useful life of servers compress? It is our view that the dynamic will hold in that today’s most demanding training infrastructure will serve a variety of future use cases, thereby extending the useful life (i.e. revenue per token) of GPUs beyond their initial purpose. At the same time, we believe the more rapid innovation cycles being pushed by NVIDIA will somewhat compress depreciation cycles from their current six years to a more conservative five year timeframe. 

The Great Recalculation

AI, Server Depreciation, and the 6-Year CPU/GPU Lifespan

The “6-Year Shift”

In the first half of the 2020s, a major financial shift occurred. The world’s largest cloud providers – Amazon, Alphabet, and Microsoft – all extended the official “useful life” of their servers from a 3- or 4-year standard to 6 years. This is an accounting change that noticeably impacts the income statement. The question is does the extended lifecycle apply for the new economics of Artificial Intelligence?

Provider 2017 2018 2019 2020 2021 2022 2023 2024 2025
Amazon (AWS) 3 3 3 4 4 5 5 6 6
Alphabet (Google Cloud) 4 4 4 4 4 4 6 6 6
Microsoft (Azure) 4 4 4 4 4 4 6 6 6

Fig. 1: Server useful life (depreciation schedule) for the “Big 3” hyperscalers. The coordinated shift from 3/4 years to 6 years is easy to see across time.

The New Players Emerge

This trend isn’t limited to the “Big 3.” A new class of AI-focused “Neoclouds” like CoreWeave have adopted the 6-year schedule from the start, while others like Nebius and Lambda Labs are setting their own policies on a more conservative basis. This visual compares the 2025 landscape, showing how 6 years has become the new standard for AI infrastructure, but gives us a glimpse from the pure play GPU clouds how the cycle may compress.

Amazon (AWS)
6 years
Alphabet (Google)
6 years
Microsoft (Azure)
6 years
CoreWeave*
6 years
Nebius*
4 years
Lambda Labs*
5 years

Fig. 2: 2025 server useful life comparison. The 6-year schedule dominates, while Nebius and Lambda Labs keep shorter, more conservative policies.

Will 6 Years Hold for AI? The “GPU Value Cascade”

How can $50,000+ GPUs that go into servers that cost millions, in a field that innovates every 1-2 years, still be “useful” in year 5 or 6? The answer is the value cascade. A GPU’s economic life extends far beyond its primary use. It “cascades” down from the most demanding tasks (training) to less demanding, but still highly profitable, tasks (inference).

Years 1-2

Primary Economic Life

The asset (e.g., NVIDIA GB200) is deployed for its most valuable, cutting-edge task.

Primary Workload Foundational Model Training
Years 3-4

Secondary Economic Life

A new generation GPU arrives. The old GPU (GB200 / H100) “cascades” to high-end inference.

Secondary Workload Premium Real-Time Inference
Years 5-6

Tertiary Economic Life

The asset (H100 / A100) is fully depreciated but still generates revenue from less-demanding tasks.

Tertiary Workload Batch Inference & Analytics

Figure 3: The idea that GPUs will have a useful life of 2-3 years is incorrect in our view. Today’s training infra will support future workloads and applications such as inference .

The Financial Impact

The benefit of such a change to operating profit is meaningful, but perhaps not as dramatic as many investors fear. For example, Amazon increased the estimated useful life of its servers in three stages: Jan ’20 (3 to 4), Jan ’22 (4 to 5) and the most recent change from 5 years to 6 years in early 2024. The most recent change added roughly $3.2 billion to its operating income in 2024. Over time this could grow but so will the company’s operating profits. So we’re talking about perhaps a 10-12% tailwind. Not game changing but noticeable.

6

Years

The new de facto standard depreciation schedule for AI-focused cloud infrastructure.

2-3x

Extended Life

The “value cascade” gives GPUs an economic life 2-3 times longer than their primary training role.

~$7B

Annual Impact

Extending asset life directly boosts operating income by billions of dollars for hyperscalers.

Figure 4: Extending server and GPU lifecycles creates measurable, multi-billion-dollar financial tailwinds. But at the scale of large cloud providers, the overall operating income impact is around 10% and not as dramatic as many believe.

This infographic is a single-page experience built with HTML, Tailwind-style classes, and simple CSS-based visuals. No SVG, Mermaid JS, or post-level JavaScript is required.

Data sourced from public SEC filings and industry reports.

The “6-Year Shift”

What the infographic shows:
Figure 1 shows the depreciation schedules for AWS, Google Cloud, and Azure from 2017–2025. The chart highlights the coordinated progression from 3- and 4-year schedules to a uniform six-year useful life assumption beginning in 2023–2024.

Key takeaways:

  • AWS was first to extend asset life, triggering a fast-follow response.
  • By 2023, all three hyperscalers normalized on six years.
  • The expansion represents a noticeable change in operating income by spreading depreciation over a longer horizon.
  • The shift occurred before the massive wave of AI CAPEX, meaning AI infrastructure inherits but may eventually challenge this standard.

Our research suggests that this shift reflects the hyperscalers’ confidence in workload diversification. Even as hardware aged, demand for general compute, analytics, web services, and long-tail workloads sustained revenue-generation across an asset’s lifecycle. The open question is whether AI GPU estates, which are more expensive, power constrained, and evolving faster, behave the same way.

The New Players Emerge

What the infographic shows:
Figure 2 shows a horizontal bar comparison of depreciation schedules across hyperscalers and AI-native neoclouds (CoreWeave, Nebius, Lambda Labs). The big clouds have settled firmly on six year depreciation, while neoclouds take more conservative positions (5 years for Lambda Labs, 4 years for Nebius).

Key takeaways:

  • CoreWeave is the outlier, adopting an aggressive six-year posture despite an exclusively AI-intensive focus;
  • Nebius and Lambda Labs use shorter cycles, reflecting faster modernization cycles and possibly less heterogeneous workload mix;
  • Neocloud strategies give an indicator of where AI-pure-play economics may differ from general cloud economics.

In our view, Figure 2 is a harbinger for slight compression. AI-first clouds cannot afford stagnant infrastructure; performance/Watt gains in successive GPU generations directly determine competitiveness. As neoclouds scale, their 4-5-year cycles will likely influence hyperscaler modeling —especially as accelerated computing consumes a larger share of CAPEX.

Will 6 Years Hold for AI? The GPU Value Cascade

What the infographic shows:
Figure 3 depicts a three-stage lifecycle framework:

  • Years 1–2: Primary economic life to support foundational model training
  • Years 3–4: Secondary life to support high-value real-time inference
  • Years 5–6: Tertiary life to support batch inference & analytics workloads

Key takeaways:

  • The value cascade structurally extends GPU usefulness, even as generations turn over rapidly;
  • Training requires peak performance; inference tolerates lower latency constraints; batch/analytics operate at the long tail;
  • This is analogous to server repurposing that justified past depreciation extensions.

We believe this framework supports longer useful life assumptions than the 2-3 years many have stated. Silicon doesn’t die at end-of-training usefulness, rather it transitions to less demanding, but still revenue-generating tasks. This is the essence of token revenue maximization in the AI factory era.

The Financial Impact

What the infographic shows:
Figure 4 shows three tiles summarizing impact:

  • 6 years as emerging standard
  • 2–3x extended life from cascading
  • $7B annual impact to operating income (example using approximate figures from hyperscaler).

Key takeaways:

  • Extending useful life does lift GAAP profitability for hyperscalers;
  • GPUs however may actually have longer economic tails than CPUs because inference and internal workloads are insatiable. But perhaps not as long as pre-GenAI CPU cycles;
  • The financial leverage from longer depreciation schedules will become even more pronounced as GPU CAPEX scales into the trillions.

Our analysis suggests that AI factories magnify the issue but not to the extent many in the media have projected. A one-year change in useful life assumptions for trillion-dollar GPU estates can swing operating income by tens of billions; but in the grand scheme of operating profits for hyperscalers it’s not game changing. This dynamic will become a recurring narrative in earnings calls and investor guidance. As such, investors should look at operating and free cash flows to get a better sense of business performance.

The following section summarizes our thinking on this issue:

Premise 1: GPUs Have Resilient Lifecycles

Key takeaways:

  • Training is a short window (1–2 years) where only the newest silicon is competitive;
  • Inference demand is exploding and absorbs “last-gen” GPUs for years 3–4;
  • Utility workloads (internal processing, fine-tuning, retrieval augmentation) extend asset life to year 5 and potentialy beyond.

We believe the proliferation of agentic applications, retrieval workflows, automation assistants, and fine-tuning pipelines will expand the useful life of GPUs beyond the training phase. This reinforces the durability of extended economic cycles.

Premise 2: Neoclouds Hint at Slightly Shorter Lifespans for Hyperscalers

Key takeaways:

  • AI-native clouds optimize for modernization speed over accounting benefits;
  • A 5-year midpoint appears to be the emerging equilibrium for AI-centric providers;
  • Accelerated computing is changing the shape of budgets – shorter depreciation aligns better with faster GPU innovation cycles.

Our research indicates that this five-year midpoint is where hyperscalers will ultimately converge as accelerated infrastructure dominates CAPEX. While general compute has historically survived at six years, AI factories operate under different competitive pressures. Moreover, the utility of legacy CPU technologies may compress as they become less useful. But on balance, we don’t see a dramatic alteration of the income statement as a result of AI.

Premise 3: Cash Flow Becomes the More Meaningful KPI

Key takeaways:

  • High depreciation masks the true economic performance of AI factories;
  • OCF adds back depreciation and reveals that GPU estates generate strong cash even when earnings appear compressed;
  • As CAPEX rises, earnings divergence from cash flow will widen.

In our opinion, investors must shift their valuation frameworks. Operating cash flow will become an increasingly important KPI relative to non-GAAP and GAAP income. OCF will be a primary indicator of AI factory health, sustainability, and ROI timelines. This becomes crucial as depreciation cycles compress.

Synthesis of the Premise

Key takeaways:

  • GPUs are durable economic assets, not short-lived commodities;
  • A five-year cycle is emerging as the balance point between modernization and economic efficiency;
  • Investors must evaluate AI factories through a cash-flow lens rather than earnings alone.

The above graphic frames the strategic direction of the industry, which will likely extended economic utility with slightly shorter accounting life – reflecting both technological realities and revenue/token expectations.

Conclusion

Our research indicates that the useful life of GPUs will continue to benefit from the “value cascade,” allowing assets to generate revenue well beyond their initial training window. At the same time, the arrival of new architectures every 12–18 months will put pressure on formal depreciation schedules. We believe hyperscalers will ultimately converge on a five-year cycle – shorter than today’s six-year model but still supported by extended economic usefulness.

In the AI factory era, depreciation becomes a lever. The winners will be operators who maximize the utility curve of GPUs across training, inference, and internal workloads while maintaining access to compute, land, water, power and the skills to build AI factories. As CAPEX scales into the trillions, cash flow will increasingly be the metric investors will watch.

Article Categories

Join our community on YouTube

Join the community that includes more than 15,000 #CubeAlumni experts, including Amazon.com CEO Andy Jassy, Dell Technologies founder and CEO Michael Dell, Intel CEO Pat Gelsinger, and many more luminaries and experts.
"Your vote of support is important to us and it helps us keep the content FREE. One click below supports our mission to provide free, deep, and relevant content. "
John Furrier
Co-Founder of theCUBE Research's parent company, SiliconANGLE Media

“TheCUBE is an important partner to the industry. You guys really are a part of our events and we really appreciate you coming and I know people appreciate the content you create as well”

Book A Briefing

Fill out the form , and our team will be in touch shortly.
Skip to content