Leading AI companies have hundreds of thousands of cutting-edge AI chips

The world's leading tech companies—Google, Microsoft, Meta, and Amazon—own AI computing power equivalent to hundreds of thousands of NVIDIA H100s. This compute is used both for their in-house AI development and for cloud customers, including many top AI labs such as OpenAI and Anthropic. Google may have access to the equivalent of over one million H100s, mostly from their TPUs. Microsoft likely has the single largest stock of NVIDIA accelerators, with around 500k H100-equivalents.

A large share of AI computing power is collectively held by groups other than these four, including other cloud companies such as Oracle and CoreWeave, compute users such as Tesla and xAI, and national governments. We highlight Google, Microsoft, Meta, and Amazon as they are likely to have the most compute, and there is little public data for others.

Published

October 09, 2024

Last updated

February 22, 2025

Epoch’s work is free to use, distribute, and reproduce provided the source and authors are credited under the Creative Commons BY license.

Learn more

Overview

We estimate how many NVIDIA accelerators and TPUs each company owns, then express these in terms of H100-equivalent processing power, based on their tensor-FP16 FLOP/s performance. For NVIDIA, we do this by combining data on overall sales and the allocation of these sales across different customers. For TPUs, we rely on industry analysis about Google’s installed TPU capacity. Note that our “Google” estimate includes all of Alphabet.

A large portion of this compute is rented out to other parties. For example, OpenAI rents its compute from Microsoft, and Anthropic rents compute from Amazon and Google. On the flip side, these four companies may have access to compute owned by other companies; e.g. Microsoft rents at least some compute from Oracle and from smaller cloud providers.

Analysis

We estimate NVIDIA AI chip sales from their reported revenue. NVIDIA’s total data center revenue from the ten quarters from beginning of 2022 to mid-2024, was $111 billion. We estimate that around 80-85% of NVIDIA’s data center revenue comes from AI accelerators, with the rest coming from networking equipment, meaning NVIDIA has sold around $90 billion in AI accelerators in this time period.

To find the number of chips sold, we divide this by the price per chip. For simplicity, we assume that all the chips sold in 2022 were A100s and all chips sold from 2023 onwards were H100; in reality, there are other chip models and the exact split is uncertain (see Assumptions). We then estimate the average price of A100s in 2022 and H100s in 2023-2024 (likely $10-15k and $20-40k respectively).

This leads to an estimate of roughly 3 million H100 chips (or the equivalent thereof) sold between the beginning of 2022 and mid-2024. Subsequently, we estimate the per-company allocation of H100-equivalents by matching them to their estimated fraction of NVIDIA’s revenue.

We estimate confidence intervals based on differing reports of companies’ shares of NVIDIA’s revenue, and differing reports of H100 prices. We model these as log-normally distributed variables, and create a simple Monte Carlo model to estimate chips owned by each company.

For Google’s TPU chips, we rely on reports from two semiconductor research firms, TechInsights and SemiAnalysis. TechInsights estimated TPU manufacturing volume based on revenue from Broadcom and memory sellers (graph available here). We matched these numbers with TPU version release dates and estimated how many TPUs of each model were manufactured over time. SemiAnalysis reported that as of September 2024, Google has deployed “millions” of TPUs in data centers that draw over 1 GW in power. This is the equivalent of ~650,000 H100s, assuming that the TPU fleet overall has the same energy efficiency as the TPUv4. We again use a simple Monte Carlo model to derive our overall estimates from these sources.

Data

This insight uses several external data sources:

  • NVIDIA’s overall sales from their earnings reports. We consider the period from early 2022 to the quarter ending July 2024 (most recently reported). We chose this time frame because a large boom in AI investments began in 2022, so most of these sales are for the AI industry, and these sales represent a large majority of NVIDIA’s AI chip sales to date. Note that these chips must be actually delivered for the revenue to be reported, under generally accepted accounting principles.
  • NVIDIA’s disclosures in their earnings reports and their most recent regulatory filings. NVIDIA’s 10-Q form reports the portion of total revenue that came from five customers that made up around 10% of total revenue in the last two quarters. However, the breakdown is anonymized and may not reflect final ownership (as opposed to sales to chip resellers), so we also use other sources to help interpret these results.
  • Large tech companies’ capital expenditure on NVIDIA products from recent Bloomberg data. Cross-referencing this against NVIDIA’s regulatory filings allows us to estimate how much revenue came from different leading AI labs. This is corroborated by recent statements from NVIDIA that 45% of data center revenue came from “large cloud providers” in Q2 FY2025. These likely include Microsoft, Amazon, Alphabet, and Oracle.
  • Google’s TPU deployment from recent reports on TPU manufacturing and Google’s cluster scaling. We can express these in terms of H100-equivalents using details from the reports and TPU specs.

These data are tabulated below. We estimate uncertainties based on differing reports of NVIDIA’s revenue shares for different customers, and differing reported prices for H100s.

Chip allocation by customer
Datacenter revenue Feb 2023-July 2024
5% CI 95% CI
$96,400,000,000 $96,400,000,000
These figures come from summing NVIDIA quarterly revenue (source). Nvidia sells various products; most of its revenue comes from its data center division, and most of those sales are chips.
Chip portion of datacenter revenue in 2023-2024
5% CI 95% CI
0.81 0.86
Based on NVIDIA reports. We don't have complete data for the last six quarters but all reported were 14-20% (and trending down): source.
Average price of H100
5% CI 95% CI
$20,000 $40,000
Media reports vary from 25k to 40k. Average price is very unlikely to be >= 40k due to bulk discounts. 20k chosen for the lower end because of possible recent discounts due to upcoming Blackwells, which will cost ~$40k while being ~2x as powerful.
Datacenter revenue Feb 2022-Jan 2023
5% CI 95% CI
$15,000,000,000 $15,000,000,000
See "Nvidia Revenue" for additional details.
Chip portion of datacenter revenue in 2022
5% CI 95% CI
0.7 0.83
No data for 2022, but this has trended downwards from ~20% to 14% since early 2023: source.
A100 average price
5% CI 95% CI
$10,000 $15,000
13k per this report from 2022: "Today, an Nvidia A100 80GB card can be purchased for $13,224" (source). 10k in early 2023: source.
A100 TFLOP/s (bf16)
312
H100 TFLOP/s (bf16)
989
(Divided official spec by 2, assuming sparsity isn't used for AI training) source
Microsoft share of total revenue
5% CI 95% CI
0.12 0.25
NVIDIA's Form 10-Q reports that "Customer A" had a 14% share of total revenue in the last two quarters; this is almost certainly Microsoft. Note that this is last two quarters, not last six, so there is uncertainty over their share of last six quarters. Bloomberg estimates Microsoft is 19% of Nvidia's total revenue. Resales may also affect amounts owned by each company. Add an uncertainty buffer in both directions for 95% CI.
Meta share of total revenue
5% CI 95% CI
0.075 0.15
Bloomberg estimates Meta is 10% of total revenue. In Form 10-Q, Customers B through E made up around 10% of total revenue in either the last quarter or last two quarters. The form also notes that "two indirect customers which primarily purchase our products through system integrators and distributors, including through Customer B and Customer E, are estimated to each represent 10% or more of total revenue attributable to the Compute & Networking segment." This suggests that Customers B and E are primarily resellers (such as Dell or Supermicro) and two other customers purchase ~10% of NVIDIA chips through these resellers; in any case, this suggests there are five customers who own around 10% of NVIDIA chips, or more.
Alphabet share of total revenue
5% CI 95% CI
0.06 0.12
Bloomberg estimates Alphabet is 8% of total revenue. Alphabet is likely one of the customers that make up close to 10% of revenue.
Amazon share of total revenue
5% CI 95% CI
0.05 0.12
Bloomberg estimates Amazon is 6% of total revenue. Amazon is likely one of the customers that make up close to 10% of revenue.
Additional constraint: "Large cloud providers" were 45% of *data center* revenue. Meta is not a large cloud provider, so this probably means that Microsoft, Amazon, and Google should add up to around 35-40%. The lower-end chip share estimates for Microsoft, Alphabet, and Amazon add up to 34%, and the higher-end estimates add up to 47%, so the lower-end shares may be more accurate.
Nvidia Revenue
Quarter Datacenter revenue (billions USD) Total revenue (billions USD)
Q2 FY25 (ending July 2024) 26.3 30
Q1 FY25 22.6 26
Q4 FY25 18.4 22.1
Q3 FY24 14.51 18.12
Q2 FY24 10.32 13.51
Q1 FY24 4.28 7.19
Q4 FY23 3.62 6.05
Q3 FY23 3.83 5.93
Q2 FY23 3.81 6.7
Q1 FY23 (beginning February 2022) 3.75 8.29
Total 111.42 143.89
TPUs
2022 TPU count
20% CI 80% CI
1,250,000 1,800,000
TechInsights estimates 1.5 million (we add a +/- 20% uncertainty interval), but we don’t know breakdown by models.
2022 TPUv3 fraction
20% CI 80% CI
0.2 0.6
Based on release dates, we assume ~half were TPUv3, but with high uncertainty.
2022 TPUv4i fraction of all TPUv4
20% CI 80% CI
0.1 0.9
We're very uncertain about how many were v4i. We believe most usage is inference, but regular chips can be used for inference.
2023 TPU count
20% CI 80% CI
1,670,000 2,400,000
TechInsights estimates 2 million (we add a +/- 20% uncertainty interval), but we don’t know breakdown by models.
2023 v4 fraction
20% CI 80% CI
0.5 0.8
We believe all produced in 2023 were v4 or v5, based on release dates. Assume they were mostly v5, but with high uncertainty.
2023 v5e fractionof v5
20% CI 80% CI
0.5 0.9
Given release dates, we expect most were v5e, but with high uncertainty.
2024 H1 TPUs
20% CI 80% CI
1,000,000 1,500,000
Extrapolate from TechInsights quantities in previous years.

Assumptions

Non-NVIDIA/Google chips: We only estimate quantities for NVIDIA AI accelerators and Google TPUs, because we assume that they make up a very large majority of AI hardware. This seems consistent with available information on other hardware manufacturers. NVIDIA is estimated to have a large majority of AI chip market share. AMD’s AI chip quarterly revenue only reached ~4% of NVIDIA’s in 2024 ($1 billion). Intel has an even smaller share, predicted to sell $500 million of its Gaudi 3 AI chips in 2024. Chinese companies such as Huawei are ramping up their AI chip efforts, but analysts generally still consider NVIDIA to be the market leader in China, despite US export controls preventing the sale of the highest-quality chips to China.

Several large tech companies, including Amazon, Meta, and Microsoft, are developing their own in-house AI chips. Amazon has said it will offer clusters of up to 100,000 of its “Trainium” chips. There currently isn’t public evidence that any of these companies have deployed their in-house chips at large scales.

Non-chip hardware: As noted above, we subtract Nvidia’s networking revenue to find revenue from AI chips. However, the remainder may still include the value of some non-chip equipment, such as non-chip components in NVIDIA’s DGX servers. On the other hand, it is possible that some of this overhead is already included in estimates of H100’s unit price.

Differences between NVIDIA chips: For simplicity, we assume that all NVIDIA AI chip revenue is from H100s. In practice, some of these sales were other chips such as A100s (the flagship AI chip before the H100), H200s (a newer chip with the same peak compute as the H100 and improved memory and bandwidth), L40s (a lower-tier data center chip), and H800s and H20s (lower-quality chips sold to China).

There isn’t good public information breaking down NVIDIA sales by chip, but the large majority of sales have likely been H100s, consistent with projections from last year that NVIDIA would sell 2-2.5 million H100s in 2023 and 2024. The distinctions between these chips should have a small impact on estimates, if price-per-performance was similar across these chip models (e.g. due to discounts on older chips like the A100). This is notably not true of the H20 chip, but they are not a large share of overall sales.

Sales data vs chip ownership: NVIDIA’s sales data may not accurately reflect which end customer ultimately receives their chips, since some companies such as Supermicro are in the business of reselling NVIDIA chips/servers. This contributes to the upside uncertainty on each hyperscaler’s share of NVIDIA chips, since they may own chips that are puchased directly and indirectly.