Our AI Chip Owners hub estimates how the world’s AI compute is distributed among major companies and customer categories. This is an open research project that relies on company disclosures and estimates from third-party analysts, building on our AI Chip Sales hub, which estimates total AI chip sales. This hub tracks ownership, not usage — most frontier AI developers do not own the chips they use for training and inference. This works is intended to inform the researchers and the public about the strategic landscape of the inputs to AI progress.
The data is available on our website as a visualization or table, and is available for download as a ZIP file.
If you would like to ask any questions about the data, or suggest companies that should be added, feel free to contact us at data@epoch.ai.
Epoch’s data is free to use, distribute, and reproduce provided the source and authors are credited under the Creative Commons Attribution license.
To estimate how many AI chips are owned by major technology companies, we build upon estimates of aggregate AI chip volumes from our AI Chip Sales explorer. These aggregate estimates come primarily from chip vendor revenue reporting, reports from media and analysts and industry research firms on average chip prices, and direct disclosures and estimates of chip volumes from companies and analysts. The full methodology is available on the Chip Sales hub.
Building on these aggregate chip volumes, we estimate how chips are distributed among several major compute owners, with the remainder allocated to an ‘Other’ category. We do not have a comprehensive breakdown of how compute is distributed among “Other”; further discussion can be found in the Nvidia methodology (since most of ‘Other’ is Nvidia compute) and in the FAQ. For visual simplicity, ‘Other’ in the explorer’s visualizations also includes smaller compute owners such as Tesla where we have confident estimates.
To allocate these chips to specific owners, we draw primarily on analyst/industry research estimates of chip ownership, company statements, and financial disclosures. We summarize this methodology below, followed by more detailed methodologies for each chip designer.
Nvidia chips comprise most of the world’s AI compute, so most of our modeling focuses on Nvidia. Nvidia also has a relatively diverse customer base. We estimate Nvidia chip ownership for several hyperscalers and large cloud companies, one frontier AI model developer (xAI, now owned by SpaceX), and Chinese customers. For each of these, our approach differs somewhat:
For AMD, disclosures and industry research coverage of customer shares are more limited than for Nvidia. We rely primarily on 2024 estimates that attribute most AMD Instinct GPU sales to Meta and Microsoft. Our extrapolation of these estimates to 2025 carries added uncertainty.
For custom hyperscaler chips (Google TPU/Amazon Trainium), we attribute ownership to the respective hyperscaler. We will revisit this allocation for future quarters, as Broadcom plans to begin selling TPUs to external customers such as Anthropic.
For Huawei, we attribute all Huawei Ascend chips to Chinese customers collectively, absent evidence of significant sales outside China.
The full code for our analyses is available here.
Note: this methodology was originally written in April 2026. While we plan to maintain and update this hub and methodology over time, descriptions of industry context may not reflect more recent developments.
The largest owners of Nvidia compute are the “hyperscalers”, meaning Microsoft, Alphabet (a.k.a. Google)1, Amazon, Meta, and Oracle.
We rely on a mix of Nvidia disclosures, analyst estimates, and capex figures to estimate their Nvidia purchases. In short:
More details and a full list of our parameter values can be found in this sheet: Nvidia ownership shares
We then use these hyperscaler share allocations to model their overall chip purchases using Monte Carlo sampling. This builds on our revenue-based model of overall Nvidia chip sales, which is described here. Most of the explanation of how we implement our uncertainty modeling can be found in the code itself; the rest of this methodology focuses on our parameter choices (point estimates and uncertainty intervals) and the evidence behind them.
Nvidia frequently discloses in earnings commentary what share of their data center revenue comes from “large cloud providers” or “hyperscalers”.2 We assume that “large cloud service providers” means Microsoft, Amazon, Google/Alphabet, and Oracle, while “hyperscalers” means those four plus Meta. This is consistent with analyst estimates discussed below. This became clear when Nvidia said for the quarter ending January 2026 that “hyperscalers” referred to the “top five cloud providers and hyperscalers”.3 Nvidia generally attributes approximately 50% to large CSPs (implying approximately 60% for the top five including Meta, and somewhere in between 50-60% when including Meta but subtracting Oracle), but for some quarters they state that hyperscalers overall are approximately 50%.
These disclosures are listed in this table.
| Date/time period | Disclosure | Category | Presumed company set |
|---|---|---|---|
| Quarter ending July 2024 | “Cloud service providers represented roughly 45% of our Data Center revenue” | Cloud service providers | Microsoft, Amazon, Alphabet, Oracle, though they didn’t qualify with “large”. |
| Quarter ending January 2025 | “Large cloud service providers … represented approximately 50% of our Data Center revenue.” | Large cloud service providers | Microsoft, Amazon, Alphabet, Oracle |
| Quarter ending April 2025 | [large cloud service providers] “remained our largest at just under 50% of Data Center revenue.” | Large cloud service providers | Microsoft, Amazon, Alphabet, Oracle |
| Quarter ending July 2025 | “large cloud service providers… represented approximately 50% of Data Center revenue.” | Large cloud service providers | Microsoft, Amazon, Alphabet, Oracle |
| Quarter ending January 2026 (no disclosure for previous quarter) | “hyperscaler revenue increased and remained our largest customer category at slightly over 50% of Data Center revenue, while growth was led by the rest of our Data Center customers as revenue diversified.” | Hyperscalers Note the claim that growth in this quarter (22% QoQ) was led by non-hyperscaler customers | Microsoft, Amazon, Alphabet, Oracle, and Meta CFO clarified in call this was their top five customers: “the top five cloud providers and hyperscalers… account for a little over 50% of our Data Center revenue” |
Nvidia often sells hardware to its final customers indirectly through server manufacturers or other distributors. We assume these disclosures all measure total direct and indirect revenue, rather than just direct purchases. These disclosures are meant to characterize the diversity of Nvidia’s customer base, so direct revenue alone would underestimate their dependence on hyperscalers, which would be substantially misleading. And a significant share of Nvidia hardware is routed through intermediaries, which makes it implausible that hyperscalers have a direct share of 50% of data center revenue even before their indirect purchases.4
Analyst estimates and other reports
Several financial analysts have independently estimated what proportion of Nvidia’s revenue is attributable to the top four hyperscalers:
We synthesize these reports below using rounded averages. The second row converts share of total revenue to share of data center revenue: this assumes an 88% data center share of revenue (typical in 2024), and that hyperscalers do not buy non-data center Nvidia products like gaming GPUs. We use hyperscaler share of data center revenue to estimate their proportion of AI chip sales; see “Revenue shares to chip sales” for more details.
| Company | Share of Nvidia total revenue through 2024, per analysts | Share of Nvidia data center revenue through 2024 (using 88% data center share) |
|---|---|---|
| Microsoft | 19% | 21% |
| Meta | 9.5% | 11% |
| Amazon | 7% | 8% |
| 7% | 8% | |
| Total | 42.5% | 48% |
Other reports
The Information reports that Nvidia sources told them that Microsoft was Nvidia’s biggest customer in 2024.
Additionally, media reports corroborate that Google and Amazon are still major Nvidia purchasers as of 2025, despite ramping up their custom AI chips. As of August 2024, Google had placed orders for 400,000 GB200s (worth approximately $15 billion) with Nvidia. In January 2025, the same outlet reported that “Microsoft, Amazon, Google, and Meta had each placed Blackwell rack orders worth $10 billion or more”.
$10 billion for GB200s alone would be about 5% of Nvidia’s eventual 2025 data center revenue, so this suggests all four continued to buy a robust share of Nvidia hardware in 2025. At least some of these orders were canceled or delayed due to the heating issues with GB200 NVL72 racks, but there is also upside uncertainty since they presumably also bought B300s and Hoppers.
It is rare for companies to explicitly disclose their Nvidia purchases, or for exact figures to be reported in the media. One exception is that in January 2024, Meta said they planned to buy 340k Hopper GPUs in 2024, aiming for a total compute stock of 600k Hopper-equivalents by the end of 2024. This contrasts with an estimated approximately 2.7M in Hopper sales in 2024, which would imply Meta bought 12.5% of Hoppers in 2024 if these plans were borne out. But Meta did not confirm whether they met or exceeded this goal, so we don’t have independent confirmation of this. In particular, they may have bought more Hoppers and fewer Blackwells in 2024 than anticipated.
The hyperscalers regularly disclose their capital expenditures (capex), but they do not break down how much is AI-related or how much goes to AI chips. But Microsoft and Meta’s capex figures are still informative (more so than Google’s or Amazon’s), since they both primarily rely on Nvidia for AI compute.6
One approach to estimating Meta and Microsoft’s Nvidia purchases based on their capex figures is to look at the ratio of their estimated Nvidia spending (based on extrapolating the analyst estimates) to their total capex, and see whether the share of their total capex going to Nvidia seems plausible.
We implement this approach in this notebook. Based on analyst estimates from 2024, Microsoft had a roughly 21% share of Nvidia’s data center revenue, while Meta had a ~11% share. The graph below plots the ratio between Meta’s/Microsoft’s implied Nvidia spending given these share estimates, and their reported total capex.

As expected, the implied Nvidia share of each company’s capex grows rapidly through 2023 and reaches 30% in 2024, consistent with AI taking over their overall capex. However, assuming a static 11% revenue share for Meta implies that their Nvidia share of capex went down in 2025, which seems directionally incorrect given continued rapid growth in Meta’s overall capex that is presumably driven by AI.
By contrast, when we assume Meta’s share of Nvidia grows to approximately 13-14% (our projection from the “Bottom Line” section below), the trajectory (the light dotted blue line) looks more plausible, growing slightly in 2025. One could ask whether the Nvidia share of Meta’s capex should be growing even more than this, but this looks like the right ballpark.
Stepping outside the black-box approach, can we directly estimate Nvidia spending from a hyperscaler’s overall capex? We sketch out a model for Meta and Microsoft (more details in the appendix). The short answer is that the capex figures for Meta and Microsoft are both quite plausibly consistent with the analyst estimates (yielding point estimates that Meta was responsible for approximately 11% of Nvidia’s data center revenue in 2025, and Microsoft approximately 18%), but we leave a full model of capex and AI chip purchases for future work.
Google’s and Amazon’s capex is harder to interpret than Meta’s and Microsoft’s due to their large custom AI chip programs. Amazon in particular has the highest capital spending of any hyperscaler, but a large minority goes to retail infrastructure like warehouses, not data centers. Meanwhile, Google is in third place in capex with approximately $90 billion in capex in 2025, behind Microsoft (approximately $110 billion) but ahead of Meta ($72 billion), but also spends a significant amount on TPU-based data centers. Google spent approximately $12-13 billion on TPU chips themselves, based on Broadcom disclosures, with total TPU-related capex in 2025 likely several-fold higher (see Appendix: Direct capex model), i.e., significantly greater than the approximately $18 billion gap between Google and Meta’s total capex. This means Google’s total capex figures are consistent with the analyst reports that Google buys less Nvidia equipment than Meta, though they do not put any tight bounds on the exact share.
For 2024 and earlier, we defer to the analyst estimates as point estimates of the top four hyperscalers’ share of Nvidia data center revenue, though we model some uncertainty in these shares. While most of these estimates are from 2024 or early 2025, we extrapolate these backwards to 2023 with higher uncertainty.
However, for 2025 and beyond, we need to take special care about estimating hyperscaler shares, absent more recent figures from analysts. This is because Nvidia’s AI chip sales are growing very rapidly: their data center revenue increased 75% year-on-year in the quarter ending January 2026, and we estimate they shipped almost 3× more H100e in 2025 than 2024!
Here’s the overall state of the evidence for hyperscaler shares in 2025, summarizing relevant content discussed so far:
Our approach to synthesizing this evidence is an exercise in reconciliation. We propose a set of point estimates that meet the following constraints:
This leads to the following adjustments to the analyst point estimates. You can see more details and specific numbers in this sheet.
Overall, this exercise only leads to minor adjustments to the analyst estimates for 2025, with Microsoft staying the same at 21%, Meta increasing its share from 11% to approximately 13%, and Amazon and Google increasing their shares to approximately 9%. We estimate that just over 50% of Nvidia’s data center revenue in the past year came from the top four hyperscalers, and over 55% from the top four hyperscalers plus Oracle.
We have some remaining uncertainty about Amazon and Google. For one, we treat them symmetrically, but the distribution of Nvidia between Amazon and Google may have changed. Both companies ramped up their custom AI chip programs in 2025, with Google’s TPU achieving very high volumes, raising the possibility that they might have scaled back their Nvidia purchases relative to the other hyperscalers. While this hypothesis is in tension with the fact that hyperscalers overall grew their share of Nvidia between 2024 and 2025, it’s possible that Microsoft/Meta/Oracle grew their shares by even more than we estimate, squeezing out Amazon and Google.
But we judge that we have enough evidence about their Nvidia purchases in 2025 specifically (e.g., The Information on their large Blackwell orders) to mostly rule out a large decline (e.g., <=5% share for either company). In the broader context of the AI cloud industry, it also seems unlikely that either company would let up dramatically on their Nvidia investments, risking a loss of market share in AI cloud since Nvidia chips are still the most popular choice among AI chip customers. One industry analyst recently claimed that supply constraints on TPUs have forced Google to buy Nvidia GPUs in large quantities.7
We also think a large increase (e.g. to 15%) is unlikely for either Amazon or Google in 2025 because of the growth in custom chips and because this would squeeze out the Microsoft/Meta shares too much to be consistent with their capex trends.
Oracle is a major AI hyperscaler, and is one of OpenAI’s major compute providers via its Stargate initiative. We discuss Oracle separately because our evidence about Oracle is sparser due to relatively little analyst coverage, with just an Omdia estimate of Hopper purchases for 2024, so we proceed with caution and model Oracle with more uncertainty.
In short:
Our bottom-line parameter values for Oracle can be found in this sheet.
Capex
For the four quarters through November 2025, Oracle spent approximately $35 billion in capital expenditures, around 50% of Meta’s $72 billion in capex in calendar year 2025. Oracle’s capex has grown very rapidly, with approximately $11 billion in capex in 2024, and approximately $7 billion in each of 2022 and 2023.
There are two relevant notes for interpreting these numbers.
These both suggest that Oracle’s capex relative to other hyperscalers (e.g., 50% of Meta’s capex) understates its relative share of Nvidia spending, and we adjust our point estimates accordingly (to 7-8% in 2025, rather than the 6% implied by the ratio of capex between Oracle, Meta, and Microsoft).
Researcher and media coverage
The industry research firm SemiAnalysis describes the history of Oracle’s cloud and AI efforts in this article. Oracle’s aggressive AI ramp reportedly began around 2024: starting in late 2023, Oracle acquired more datacenter leases (>2 GW) in the US than any other company in the period from November 2023 to January 2025.10
In terms of quantitative evidence, Omdia estimated Oracle acquired 125,000 Hopper GPUs in 2024, compared to 224,000 for Meta, 485,000 for Microsoft, and 169,000 for Google. Normalizing by other estimates of Meta’s and Microsoft’s Nvidia spending, this suggests that Oracle made up roughly 5% of Nvidia’s datacenter revenue in 2024. This is only one estimate from one firm, so we are less confident about Oracle’s share for 2024 than for the other hyperscalers.
Before 2024, our information is sparser. SemiAnalysis describes how Oracle’s capex doubled in 2022 compared to 2021, but as part of a general transition to cloud computing, not necessarily AI/GPU compute specifically. Nvidia did announce in October 2022 that Oracle would add “tens of thousands” of Nvidia A100 and H100 GPUs (compared to over 1 million Nvidia GPU sales in 2023), suggesting that Oracle reached multiple percentage points of Nvidia sales in 2023.
Another sanity check is the Stargate campus in Abilene, Texas, Oracle’s single largest data center project in 2025.11 We estimate that Stargate Abilene had approximately 100,000 Blackwells, or approximately 250,000 H100-equivalents installed by September 2025. By comparison, our point estimate of Oracle buying 7-8% of Nvidia’s AI chips in 2025 implies that it bought just over 500,000 Nvidia H100-equivalents in the first three quarters of 2025, meaning that around half of those chips were sent to Abilene (setting aside deployment lags). This seems plausible since we should expect Stargate Abilene to absorb a large but not overwhelming share of Oracle’s AI compute: OpenAI is Oracle’s largest cloud customer and Abilene was the only active “Stargate” site in 2025, but Oracle also has other AI cloud customers (e.g. ByteDance).
The Financial Times also reported in 2025 that Oracle would spend $40 billion on 400,000 Nvidia GB200 chips for Stargate Abilene. This dollar figure is much higher than 8% of Nvidia’s data center revenue, but it also seems inconsistent with the reported chip count, given that GB200 racks (including networking) typically cost <$50,000 per GPU for a total cost of approximately $20 billion. So it may actually include an expansion of Abilene Stargate beyond 400,000 GPUs, or be a total cost including non-computing equipment. In any case, the 400,000 GPU phase of Abilene Stargate has not been completed as of April 2026, and Oracle’s overall capex is not large enough to support $40 billion in spending on Nvidia equipment for a single project to date.
At a first pass, a hyperscaler’s share of Nvidia’s data center revenue in dollars should be roughly equal to their share of Nvidia AI chip counts. But we need to check whether “datacenter” corresponds to AI chips, and account for China-spec chips and lower prices.
“Datacenter” vs AI chips: We assume hyperscaler share of Nvidia data center revenue is equal to their share of Nvidia’s flagship AI chip revenue.
Nvidia’s “data center” segment is predominantly AI-related hardware, including compute and networking. Since Nvidia networking is predominantly for servers and clusters in AI data centers, we assume hyperscalers purchase networking at a similar proportion to Nvidia compute purchases.12 “Data center” also includes a small amount of software/cloud revenue; while it’s unlikely that hyperscalers buy much cloud compute from Nvidia, we ignore this for simplicity.
Chip composition: US hyperscalers have no need for downgraded China-spec Nvidia chips (A800, H800, and H20).13 Otherwise, we assume US hyperscalers buy Nvidia’s flagship chips in the same mix as Nvidia’s overall non-China customer base (they may get priority access to new chips, but so do at least some neoclouds).14
Cheaper prices: Hyperscalers also presumably pay lower prices than others due to their size and market power, which bumps up chip counts per dollar. How big is this discount?
One handle is that SemiAnalysis reports that hyperscalers pay around 5-10% less for Hopper and Blackwell servers than two categories of neoclouds. Assuming that these neoclouds are reasonably representative of non-hyperscale Nvidia customers, how should we adjust hyperscaler chip purchases?
Let d be the hyperscaler discount and h be the hyperscaler share of the market. A hyperscaler pays full_price × (1 − d) per chip, while everyone else pays full_price.
The observed average price is then a weighted mix of the two:
avg_price = h × full_price × (1 − d) + (1 − h) × full_price
Factoring out full_price:
avg_price = full_price × [h(1 − d) + (1 − h)]
= full_price × [1 − h×d]
We want to know how many more chips a hyperscaler gets per dollar compared to what the average price would imply. That ratio is:
factor = avg_price / hyperscaler_price
= full_price × (1 − h×d) / [full_price × (1 − d)]
= (1 − h×d) / (1 − d)
With hyperscaler share h roughly equal to 50%, and a discount d of 5% to 10%, the factor varies from approximately 1.03 to 1.06. This means that the hyperscaler discount yields approximately 1.03× to 1.06× more chips per dollar compared to if hyperscalers paid the overall market price. Accordingly, in our model we boost hyperscaler chip counts by this factor, relative to the chip counts implied by their share of revenue.
xAI is a frontier AI developer that recently merged with SpaceX. For now, we will continue to describe its compute as owned by “xAI” rather than “SpaceX” because xAI is more publicly associated with AI and because the xAI subsidiary is presumably still relatively operationally independent from the rest of SpaceX (by comparison, DeepMind and Google seem much more intertwined).
xAI owns and operates two large data centers in the Memphis, Tennessee area: Colossus 1 and Colossus 2. Note that Colossus 2 is jointly owned by xAI and outside investors; we attribute it to xAI for simplicity. From what we can tell, Colossus 1 and 2 make up nearly all of the compute capacity owned by xAI, though xAI reportedly also built a smaller data center in Atlanta with approximately 12,000 GPUs. xAI has also used cloud compute, at least historically. For example, Grok 2 was trained with Oracle-owned chips.
We use estimates from our Frontier Data Centers hub of the collective compute capacity of Colossus 1 and Colossus 2 over time, as well as company statements on the chip splits in these data centers. These estimates are tied to specific dates, so we map them to calendar quarters. As of the end of 2025, Colossus 1 and 2 have approximately 550,000 H100-equivalents in compute capacity. This is around 3% of all estimated Nvidia AI chip sales since 2022, and a fraction (around one-seventh to one-quarter) of the Nvidia compute owned by each of the top four hyperscalers.
Because these figures are based primarily on when data center capacity comes online, they are likely lumpy and lagged relative to when xAI actually received these chips. This means that these data center-based figures are usually lagged estimates (that is, underestimates) of the chips xAI actually owns over time. In one case, however, we incorporate B200s delivered to Colossus 2 in fall 2025 shortly before we estimate they became operational.
We attribute Nvidia sales to customers based in mainland China, building on work done in the AI Chip Sales explorer.
Nvidia chips designed for the Chinese market are assumed to have been sold to customers in China; this is a slight simplification (Nvidia sold approximately $500 million in H20s to a customer outside China in the last quarter in which H20 was sold), but one we believe has a negligible impact on the overall estimates.
For the H800s and A800s, which were more competitive with Nvidia’s flagships, it is less clear whether they were exclusively sold in China, but our estimates of H800 and A800 volumes are actually inferred from Chinese Nvidia purchases in the relevant time period, not from any H800/A800 revenue breakout (which Nvidia has not provided to our knowledge). In addition, Chinese customers likely bought between 10% and 20% of Nvidia A100s before export controls were first implemented in 2022.
Note that while Nvidia has now received US government licenses approving H200 exports to China, shipments have not yet begun as of the latest Nvidia fiscal quarter, ending in January 2026.
A longer write-up on our methodology for these estimates can be found here.
Note that our China ownership figures exclude unofficial exports to China in contravention of US export controls on leading US-designed AI chips. Prior work by Grunewald & Fist (2025) estimated that over 100,000 Nvidia A100s and H100s were shipped to China in 2024, though there is significant uncertainty in this figure. Several reports indicate that Chinese imports of export-controlled chips continued in 2025, including:
As with all of our ownership figures in this hub, we do not measure cloud compute sales. Some Chinese companies reportedly rent leading US AI chips via offshore cloud compute, which, as of early 2026, is allowed under US export controls.
CoreWeave is the largest AI “neocloud” by a significant margin. CoreWeave recently cited a third-party estimate that they were larger than the next 15 neoclouds in the US and Europe put together. One industry analyst describes CoreWeave as a “neocloud giant,” alongside Lambda, Crusoe, and Nebius, and a cursory review suggests that none of these other neoclouds is close to CoreWeave in scale.15 As a publicly traded company focused almost entirely on AI compute, CoreWeave discloses a large amount of information about their AI compute capacity.
CoreWeave primarily uses Nvidia compute. We estimate their GPU fleet composition over time: CoreWeave directly discloses its deployed GPU unit counts and data center power capacity in annual S-1 filings, and has given updates on power capacity for every quarter of 2025. We derive or estimate the breakdown into specific chip types (allowing us to find the total compute capacity) using GPU power consumption data and our estimates of Nvidia’s chip sales mix.
Code for analysis can be found here.
2022–2024: S-1 Disclosures
CoreWeave disclosed their total deployed GPU count and data center power capacity at year-end in their S-1 filing with the SEC in March 2025. Throughout, we assume their reported power capacity refers to IT power (electricity consumed by servers and compute equipment) rather than total facility power. This is standard practice for data center compute providers whose disclosures are primarily meant to characterize their compute infrastructure. For example, Core Scientific reports their power capacity in terms of critical IT power.
2022: CoreWeave disclosed that they ended the year with 17,000 GPUs and 10 MW of capacity. The low implied average IT power per GPU (~588 W) indicates a mixed fleet of A40s, RTX cards, L40s, A100s, and other lower-power chips. We exclude 2022 from our estimates for three reasons: the GPU mix is unclear and hard to anchor, many of these cards have likely been retired, and even if still in service they would represent a negligible share of CoreWeave’s current compute capacity.
2023: CoreWeave ended the year with 53,000 GPUs and 70 MW of capacity. Even in the most conservative scenario — assuming all 17,000 chips from 2022 were still running — the implied IT power per GPU is 70 MW / 53,000 = 1,321 W, which points strongly to a high proportion of H100s, since previous Nvidia chips such as the A100 had significantly lower average power draws even accounting for server and other IT overheads. CoreWeave had also placed a $100M order for H100s in summer 2022 and made them available in Q1 2023. We treat all GPUs added in 2023 as H100s.
Rather than back-solving GPU counts from the raw 2023 MW figure, we apply our 2024-derived IT power assumption of 1,472 W per GPU (see below), which we have more confidence in. We also assume roughly half of the 2022 fleet was retired during 2023, freeing up approximately 5 MW of capacity. This implies 65 MW of gross new H100 capacity deployed in 2023, or approximately 44,200 H100s added.
2024: CoreWeave ended 2024 with 250,000 GPUs and 360 MW of capacity. Assuming no 2023 GPUs were retired, this means they added 197,000 GPUs and 290 MW of capacity, implying 1,472 W per GPU. We believe virtually all of these were H100/H200 for two reasons: Blackwell chips were not meaningfully sold by Nvidia until Q4 2024 and were not deployed by CoreWeave until early 2025, and our Nvidia chip sales estimates show 95% of Nvidia GPU units sold in 2024 were Hoppers. The H100 and H200 share the same 700 W TDP so we group them together. We use the implied 1,472 W as our H100/H200 IT power assumption going forward, equivalent to a GPU/IT power ratio of ~48%. As a cross check, we derive the H100/H200 IT power using the methodology from our Frontier Data Centers tracker: taking the H100/H200 server power of 1,275 W and applying a 1.14x factor to account for additional IT power overhead beyond the GPU itself. This gives 1,454 W, which closely aligns to the 1,472 W implied by CoreWeave’s GPU count and power-capacity disclosures.
2025: Quarterly Power Capacity
For 2025, CoreWeave disclosed their total operational data center power capacity in each of their quarterly earnings, but not GPU counts. Their cumulative capacity reached 420 MW, 470 MW, 590 MW, and 850 MW at the end of Q1 through Q4, implying quarterly additions of 60, 50, 120, and 260 MW.
To estimate GPU counts by generation, we use our Nvidia chip sales estimates as a proxy for CoreWeave’s chip mix in a given quarter, under the assumption that CoreWeave’s procurement broadly tracks the overall Nvidia sales mix for non-China chips. We apply a lag between when Nvidia books revenue for a chip and when CoreWeave deploys it, and model this lag as shortening over the course of 2025 for two reasons.
First, the high capital cost of holding GPUs idle makes a long lag increasingly implausible as deployment scale grows. In Q3 2025 CoreWeave added 120 MW of capacity, and in Q4 2025 they added 260 MW — at $30–40k+ per GPU, pre-purchasing a full quarter’s worth of chips before the data center is ready to receive them would mean holding billions of dollars of assets in storage with no return. We therefore model CoreWeave as timing their GPU purchases to land close to when power capacity is available, rather than accumulating large advance inventories.
Second, we assume CoreWeave’s installation process became faster and more streamlined as 2025 progressed. On their Q4 2025 earnings call, management noted that after resolving delays flagged in Q3, they deployed over 50,000 Grace Blackwells “within weeks of receiving access to the requisite data center infrastructure.”
We model a median lag of one quarter for Q1 and Q2 2025, which is consistent with observed deployment timelines: CoreWeave made GB200 NVL72 generally available in February 2025, aligning with Nvidia first booking meaningful B200 revenue in Q4 2024, and deployed GB300 NVL72 in July 2025, aligning with B300 sales beginning in Q2 2025. The one-quarter lag is further supported by CoreWeave’s Q1 and Q2 2025 earnings presentations, in which they indicate a 3 month lag between purchasing GPUs and making revenue from them. We shorten our median lag to 0.8 quarters for Q3 2025 and 0.4 quarters (~5 weeks) for Q4 2025.
To the best of our knowledge, CoreWeave does not purchase or host China-spec chips such as the H20, H800, and A800. Thus, we exclude these chips from the sales mix and renormalize the remaining GPU types — H100/H200, B200, and B300 — to 100% before applying their IT power-weighted shares to CoreWeave’s quarterly MW additions. We weight by IT power rather than unit count because chips with higher IT power per GPU account for a disproportionate share of capacity for a given number of units.
IT power per GPU is estimated as follows:
| GPU | TDP | GPU/IT Power Ratio | Est. IT Power per GPU |
|---|---|---|---|
| H100/H200 | 700 W | ~48% | 1,472 W |
| B200 | 1,200 W | ~58% | 2,080 W |
| B300 | 1,400 W | ~62% | 2,258 W |
The H100/H200 figure is derived directly from CoreWeave’s disclosed GPU count and power capacity. For B200 and B300, we estimate IT power by using the power capacity of each GPU’s most common server spec — GB200 NVL72 and GB300 NVL72 — and applying a 1.14x overhead for other IT power costs. This follows the methodology of our Frontier Data Centers tracker; more details can be found in this sheet.
We model uncertainty in these estimates using Monte Carlo simulation with 10,000 draws, capturing three sources of uncertainty: (1) Nvidia unit counts per chip per quarter from our chip sales hub; (2) the lag between procurement of GPUs and deployment; and (3) the B200 and B300 GPU/IT power ratios. The 90% confidence intervals on total GPUs added in 2025 run from approximately 237,000 to 258,000, with H100e ranging from approximately 508,000 to 556,000.
This brings CoreWeave’s estimated cumulative fleet to approximately 497,000 GPUs at the end of 2025, or roughly 781,000 H100-equivalents [90% CI: 758,000–805,000].
Sanity Checks
Revenue: Our GPU count estimates imply approximately $4.93B in 2025 revenue [90% CI: $4.65B–$5.23B], assuming H100/H200s rent at $2/hr and B200/B300s at $3/hr, 80% utilization, and linear ramp-up of newly added GPUs within each quarter. CoreWeave’s actual 2025 revenue of $5.1B is reasonably consistent with our estimates.
Power to Compute Capacity: In our Frontier Data Centers hub, we estimate that OpenAI Stargate Abilene data center has 590 MW of IT power capacity and ~510,000 H100e as of March 2026, which is 864 H100e per MW. Applying this ratio to CoreWeave’s end-of-2025 deployed capacity of 850 MW implies 734,000 H100e, which aligns with our median estimate of 781,000 H100e.
End-of-2025 GPU Inventory
The figures above reflect CoreWeave’s deployed capacity. However, CoreWeave likely holds additional GPUs in inventory at the end of 2025, which are purchased but awaiting power capacity coming online in early 2026. To estimate these, we model the GPUs CoreWeave would have procured for their Q1 2026 deployments.
CoreWeave has guided to ending 2026 with approximately 1.7 GW of capacity, implying roughly 850 MW of additions over the course of 2026. To estimate Q1 2026’s share, we use their capex guidance: full-year 2026 capex of $30–35B and Q1 2026 capex of $6–7B, implying Q1 accounts for approximately 17–23% of annual capex. We model this share as uniform over that range, giving a Q1 2026 MW addition of approximately 145–200 MW.
Importantly, we model Q1 2026 capacity coming online gradually throughout the quarter. With a procurement lag of 0.4 quarters (~5 weeks) and assuming a linear ramp in capacity additions, CoreWeave would have pre-purchased only the fraction of Q1 2026 capacity coming online within 5 weeks of year-end — approximately 40% of Q1’s total MW addition. We assume no capacity retirements in 2026 given continued strong demand for older GPU generations.
This adds 79,000 GPUs to CoreWeave’s end-of-2025 inventory, predominantly B200s and B300s given the Q4 2025 Nvidia mix. Combined with the 249,000 from 2023–2024 and ~247,000 deployed in 2025, this brings our median estimate for CoreWeave’s total GPU stock at end of 2025 to approximately 520,000 GPUs [90% CI: 507,000–534,000], or roughly 852,000 H100-equivalents [90% CI: 811,000–866,000].
We estimate that the five hyperscalers own around 60% of total Nvidia AI compute (measured in H100 equivalents), with xAI owning at least 3%. CoreWeave, which we track separately above, is likely the sixth-largest owner. The remaining Nvidia compute — roughly a quarter of the total — is distributed across several other owner categories, though we do not know of a comprehensive breakdown.
Neoclouds, of which there are at least dozens, may own a majority of the remaining Nvidia compute.
Another large customer category for Nvidia, likely overlapping heavily with neoclouds, is “Sovereign AI”. This segment made up 15% of the company’s data center revenue in 2025. “Sovereign AI” is a marketing term without a clear definition, but in practice, it refers to compute buildouts to enhance a non-US nation’s AI capabilities or independence. These projects are generally championed by national governments but owned by corporations, meaning that many of these projects are actually owned by non-US neoclouds. Nvidia has said that this category is “driven primarily by customers based in Canada, France, the Netherlands, Singapore, and the UK”. We don’t know for sure that there is zero overlap between “sovereign AI” and hyperscaler compute located outside the US, though Nvidia’s description of customer locations suggests that this overlap is small at best.
Finally, Tesla disclosed in a 2025 investor presentation that it had 120,000 H100-equivalents of AI training compute capacity as of the end of September 2025. This capacity is likely mostly or entirely Nvidia GPUs. While Tesla specified training capacity (we take this to mean all R&D compute) as opposed to inference, Tesla’s main AI product is autonomous driving, and its AI inference is mostly done on board its vehicles using custom hardware, not leading datacenter chips. Tesla also doesn’t specify how it converts compute to H100-equivalents; we take their conversion at face value.
At any given time, Nvidia holds significant volumes of hardware as unsold or undelivered inventory: at the end of 2025, its total inventory was worth $21 billion (approximately 30% of its recent quarterly sales). Our topline ownership figures don’t include this inventory.
Our ownership figures also don’t account for shipment lags at the server/distributor level. This means we may be overestimating hyperscaler purchases for any particular quarter. Under generally accepted accounting principles, Nvidia recognizes revenue when it delivers hardware to customers, not when it receives purchase orders. However, the first layer of customers are often server OEMs, which assemble the chips into servers, rather than the company that will ultimately deploy the chips. Nvidia’s customer attribution to the hyperscalers likely accounts for this intermediate step, but at a lag; i.e., the company reports an estimate of how many of the chips sold in a given quarter to server assemblers will eventually end up at the hyperscalers. The time lag between when the server is assembled and delivered to the final customer is not accounted for in our figures. This is in addition to the time lag between revenue recognition and when chip clusters are actually installed and operational in data centers, which can take weeks or months.
Beyond inventory, we are not aware of Nvidia itself owning and deploying its own chips (which would be excluded from its sales data) on a large scale. Nvidia has a major AI research and model development arm that works on the Nemotron language model family as well as robotics models. We know that the company uses neocloud compute for at least some of these workloads. Nvidia also has an AI cloud division that reportedly runs on rented capacity from Oracle and neoclouds. We also aren’t aware of any Nvidia-owned data centers or clusters approaching the scale of frontier data centers.
For Amazon’s custom Trainium and Inferentia chips, we attribute ownership solely to Amazon itself, absent any known reports of direct sales to external customers.
Historically, TPUs are exclusively owned by Google, which rents them to external customers via the cloud. Accordingly, we attribute all Google TPU ownership to Google itself through 2025. (By “Google” we mean its parent company Alphabet, which also includes Waymo and other subsidiaries).
However, we expect Anthropic to directly purchase a significant share of TPUs in 2026.16 Per Broadcom, which is Google’s design partner and vendor for TPUs, Anthropic placed $21 billion in purchase orders for TPUv7 chips in late 2025 and early 2026.
We assume there were little or no completed deliveries in 2025, as Broadcom executives mentioned Anthropic purchase orders but not deliveries in a December 2025 earnings call, which we interpret as indicating that deliveries had not yet begun.
However, some initial deliveries may have been completed in January 2026: on Broadcom’s March 2026 earnings call for the fiscal quarter through February 2, 2026, CEO Hock Tan said that “For Anthropic, we are off to a very good start in 2026 for 1 gigawatt (GW) of TPU compute”. From this statement, we interpret both a 1 GW delivery target for 2026 and that deliveries may have already begun (“off to a very good start’).17
While Tan’s comment suggests that deliveries to Anthropic have already begun, it does not provide a clear lower bound on volume. Assuming a ramp-up period for deliveries, 1 GW in 2026 suggests a loose upper bound of 250 megawatts through in the first quarter of deliveries.
1 GW of capacity would be a substantial share of overall TPU shipments in 2026, translating to roughly 600,000 TPUv7 chips, assuming approximately 1,700 watts of IT power per chip.18 By comparison, Google took delivery of roughly 2 million TPU chips per year in 2024 and 2025, with some analysts forecasting over 4 million total TPU units in 2026. If these projections hold, Anthropic will own 10% to 20% of 2026 TPU production.
AMD’s flagship AI chip line is the Instinct GPU family. Based on industry research reports as well as announced partnerships, the largest customers of AMD’s Instinct GPUs are Meta, Microsoft, and Oracle. We are not aware of reports of any major Instinct purchases from Amazon and Google.19
We estimate the allocation of Instinct GPUs across Meta, Microsoft, Oracle, and a remaining “Other” category for 2024-2025. We also attribute sales of their Chinese-market GPU, the MI308X, to Chinese customers generically. We obtain total AMD AI GPU shipment volumes and average prices by chip type from our Chip Sales Explorer.
For 2024, we use estimates from the research firm Omdia of MI300X purchases by customer: Omdia estimates that Meta and Microsoft respectively purchased approximately 50% and 30% of all MI300X in 2024, while Oracle bought just over 10%.20 We aren’t confident in these estimated shares, and we treat them as high-end estimates given a prior that 80% market share for two buyers is unusually high. The MI300X chip made up the majority of Instinct sales in 2024; most of the remainder was the MI300A, which was primarily tailored for scientific computing, e.g., at US national laboratories. We allocate MI300X sales to Meta/Microsoft accordingly, while assuming they did not buy any MI300A.
For 2025, we make two carve-outs before allocating units to customers. First, in line with our 2024 estimates, MI300A shipments are assigned entirely to “Other.” Second, AMD disclosed $390 million in MI308X sales to China in 2025. We convert this to units using the MI308X ASP of $12,000 and assign those units to a separate “China” category, removing the associated revenue from the pool available to other customers.
For the remaining data center GPU types sold in 2025 (MI300X, MI325X, MI350X, and MI355X), we extrapolate Omdia’s estimated Meta and Microsoft shares from 2024 to 2025 as high-end estimates, while factoring in significant uncertainty about how the sales split may have changed in 2025. For example, we model Meta’s 50% share with a confidence interval of [30%, 60%] — we don’t have enough information to pin down this range more precisely, but we are fairly confident that Meta and Microsoft remain the largest customers of AMD’s data center GPUs. Meanwhile, in 2025, Oracle announced plans to build two large AMD clusters of 50,000 MI450X GPUs by Q3 2026 and 30,000 MI355X GPUs with an unclear timeframe. This appears to be broadly consistent with a persistent ~10% share, though this also carries significant uncertainty.
A full list of parameters and values can be found in the code notebook for our AMD model and accompanying data sheet.
For Huawei, we use our estimates of total shipments of Huawei’s Ascend AI chips from our AI Chip Sales explorer, which are synthesized from analyst estimates and media reports. More details can be found here. We attribute all Huawei AI chips to Chinese companies collectively, rather than to specific companies. Estimating Huawei volumes for specific customers is left for future work. We also assume Huawei was able to sell its full production — however, this may not be correct.
While there are some reports of Huawei marketing Ascend chips in moderate quantities (“low thousands”) to international customers, two features of US semiconductor export controls make it unlikely that Huawei has sold large volumes of AI chips to non-Chinese customers to date:
We have uncertainty in both overall Nvidia chip sales and each hyperscaler’s share of those chips. In our model these uncertainties are modeled as independent, absent a clear reason to think they are correlated.
The overall share of Nvidia revenue that goes to the top 4-5 hyperscalers is close to ground truth due to the Nvidia disclosures, but their wording admits uncertainty: they usually describe this share as “approximately 50%” or slightly above or below 50% or similar. Nvidia describes these numbers as estimates on their part, but Nvidia presumably has excellent visibility into their largest customers.21 In addition, these share disclosures are not made in every quarter (and sometimes omit Meta), which requires some interpolation.
So assuming approximately 10% uncertainty in each direction (e.g. point estimate of 50% to a credence interval of [approximately 45%, approximately 55%]) seems reasonable.
How would this translate to uncertainty in each company’s individual share of revenue? If these shares are independent, there’s a closed form solution, because the variance of the sum of independent variables is equal to the sum of the variances of each variable.
Let S be the total share of four hyperscalers. If each company’s share Xᵢ is independent with equal variance σ², then:
Var(S) = Var(X₁) + Var(X₂) + Var(X₃) + Var(X₄) = 4σ²
The standard deviation of S, σ_S, is the square root of Var(S) and is proportional to the width of the CI. So σ_S = 2σ, or equivalently σ = σ_S / 2: each company’s absolute uncertainty is half the uncertainty in the sum.
Because each company’s share is also about one-fourth of the total (on average), the relative uncertainty is larger: approximately double the overall relative uncertainty in S. So if the total hyperscaler share has ±10% relative uncertainty, individual company shares have roughly ±20% relative uncertainty on average.
This assumes that each company’s share is independent (uncorrelated). If their shares are positively correlated, then uncertainty in each company’s share is closer to the overall uncertainty in hyperscaler share (i.e. less uncertainty), since their values vary jointly. If they’re anti-correlated, then uncertainty in individual company shares should go up. Arguably, uncertainty in company shares should be anti-correlated, given the Nvidia information about overall hyperscaler share.22 If we learned that (e.g.) Amazon’s share was surprisingly low, we may want to bump up Microsoft’s share. But we don’t model any correlation, partly for simplicity, and partly because the analyst estimates provide independent grounding for individual company shares, versus merely inferring point estimates and uncertainty from the overall hyperscaler share.
Stepping beyond the black-box approach, is it plausible that Nvidia spending is approximately one-third of Meta’s and Microsoft’s capex? Should that be higher, or lower?
Here we attempt to model Nvidia chip purchases based on capex numbers. Let’s discuss Meta, which spent $72 billion on capex in 2025, as an illustrative example. How might this translate to their spending on Nvidia hardware?
This is the product of the following parameters. Each listed parameter multiplies the previous value, so some implicit qualifiers are omitted for brevity:

The diagram below illustrates these using several point estimates. We also build a simple Monte Carlo model based on these parameter values, yielding a point estimate that Meta purchased around 11% of Nvidia’s data center equipment in 2025 [CI: 8% to 14%]. For Microsoft, this share is 18% [CI: 15% to 22%]. This corroborates the information from analysts, but the uncertainty is quite high, and our parameter value choices are not necessarily well-grounded. So, for now this model should be viewed more as rough, suggestive evidence that corroborates evidence from analysts, and provides a directional guide to how Microsoft and Meta’s shares of Nvidia revenue might have changed when analyst estimates become outdated.
More on the timing adjustment
While we won’t model the timing adjustment in full detail, this is largely a function of growth rate in capex and data center construction timelines. For illustration, imagine a hyperscaler with the following properties:
Then for every $100 worth of data centers (measured in terms of total lifecycle cost) the hyperscaler constructs in year 1, it constructs $170 worth of data centers in year 2. In year 2, they spend $80 on IT equipment for the data centers constructed the previous year, and 20% * $170 = $34 on construction for data centers that will be finished in year 3. This means that in year 2, the hyperscaler incurred $80 + $34 = $114 in capex, and the IT share of capital expenses is ($80 / $80 + $34) = 70.1%. The baseline lifecycle IT share was 80%, so the timing adjustment is 70.1/80 = 0.876x.
Some more illustrative values can be found below, courtesy of Claude Opus 4.6. Lower lifecycle IT share, longer construction timelines, and higher capex growth rates demand higher timing adjustments, though most plausible parameter values land in the range of a 70% to 90% timing multiplier.

As with the IT share of capex, the non-IT timing adjustment will also vary with a corporation’s accounting standards and business model. Oracle in particular states that “the vast majority of our CapEx investments are for revenue-generating equipment that is going into our data centers and not for land, buildings or power that collectively are covered via leases. Oracle does not pay for these leases until the completed data centers… are delivered to us.”
Where do these ownership numbers come from?
The AI Chip Owners explorer builds on our AI Chip Sales explorer, which compiles our estimates of the total sales and deliveries of leading AI chips based on company earnings commentary, estimates from analysts and industry research firms, and media reports.
For the Owners hub, we then allocate these AI chips to owners using disclosures and earnings commentaries from chip vendors such as Nvidia, alongside analyst estimates and media reporting of major purchases. We rely significantly on the subset of analyst/industry research results that is shared with the public, so our work should not be viewed as a substitute for proprietary industry research.
More information can be found in the methodology.
How do you define ownership?
The goal of this hub is to describe which companies control the world’s AI compute. We attribute ownership informally without strict legal/financial criteria. We do not recommend using our figures to understand assets owned by these companies from a financial perspective.
Some of the compute we attribute to an AI or technology company may actually be jointly owned by other players like the data center developer or external investors (often via a joint venture or special purpose vehicle). This is true of some of xAI’s data centers, and will be true of Meta’s upcoming Hyperion data center. Beyond joint ownership structures, an AI data center or compute cluster may be wholly owned by a tech company but also serve as collateral for debt financing.
However, when an AI lab or other company uses compute provided by a cloud company, we attribute that compute to the cloud company, not to the lab itself.
What is the purpose of this work?
We want to estimate trends in AI compute, and how this compute is allocated, due to its strategic implications for AI progress and societal impacts. Our intended audience is primarily other researchers, policymakers, and the general public. This hub is not intended to serve as a tool for financial analysis of the companies mentioned.
What does “H100e compute capacity” mean?
H100-equivalent (H100e) compute capacity is total computing power measured in terms of the equivalent number of Nvidia H100 GPUs. We divide the peak number of dense 8-bit operations (FP8 or INT8) each chip can perform by the Nvidia H100’s spec, and then multiply this ratio by the number of chips. For example, since a TPUv7 can perform approximately 2.3× as many operations as an H100, then one million TPUv7s have a compute capacity of 2.3 million H100e. For chips that do not support 8-bit precision, we use their 16-bit (FP16 or BF16) performance instead.
How is chip power measured?
The power figures are based on each chip’s thermal design power (TDP). It refers roughly to the maximum sustained power draw of the chip. Importantly, the total power draw of an AI data center is typically much higher than just the total TDP of the chips inside (roughly twice as high), due to overheads at the server, cluster, and facility level.
Who are the “Other” owners besides the companies listed?
We don’t have a definitive, comprehensive breakdown of AI compute ownership, with a significant minority attributed to “Other” customers that are not among the largest hyperscalers and cloud companies, frontier AI developers, or Chinese customers.
Other categories of important compute owners include:
Do frontier AI companies like OpenAI and Anthropic own their own AI chips?
Before xAI was acquired by SpaceX, xAI was the only major pure-play AI developer that owned most of the compute it uses. By contrast, OpenAI and Anthropic do not own the vast majority of the compute they use for training and inference. For now, their compute is currently provided by cloud companies (companies that rent out compute)—primarily Microsoft, Oracle (including “Stargate”), and CoreWeave for OpenAI, and primarily Amazon and Google for Anthropic. Frontier labs housed inside larger companies, like Google DeepMind and Meta’s AI divisions, presumably mostly use compute owned by their parent companies, though Meta has signed deals to rent cloud compute.
Both OpenAI and Anthropic are reportedly interested in investing in their own compute in the future, with Anthropic planning to buy a large number of TPUs in 2026.
Does “Google” ownership describe the compute available to Google DeepMind?
No. The compute owned by Google (which is shorthand for Alphabet) is split across Google DeepMind (including research and training for Gemini models and other efforts), other internal uses at Google, and cloud compute for external customers. Similarly, the compute owned by Meta is not exclusively used by Meta’s frontier AI efforts.
Why do Google, Microsoft, Amazon, and Meta own so much compute?
Most of the world’s AI compute is owned by the “hyperscalers”—usually describing Google, Microsoft, Amazon, Meta, and sometimes Oracle. Note that the hyperscalers do not own all the chips they use: Meta has signed cloud deals with Google and CoreWeave, and Microsoft acquires some compute from neoclouds that they then rent out to their own customers.
Hyperscaler compute usage for a hyperscaler mostly falls within three broad categories. The relative ranking of these three categories is quite unclear to us.
First, renting out AI compute on the cloud to AI companies and other customers. Hyperscaler cloud compute is the primary source of compute for most pure-play AI labs, e.g. OpenAI and Anthropic. Amazon, Microsoft, and Google are the top three cloud companies overall, and almost certainly the top three AI clouds as well. Providing cloud compute is also the primary use of Oracle’s AI chips. Meanwhile, Meta does not have a cloud division.
Second, powering internal AI labs for training models, research, and inference. The four largest hyperscalers all have internal AI divisions working on large-scale foundation models. The internal labs at Google (Google DeepMind) and Meta (MSL/FAIR) appear to be much more advanced than the labs at either Microsoft or Amazon, and presumably use much more compute.
Third, powering AI products or AI enhancements to products. This is significant for both Meta and Google who use AI compute for content/ad recommender systems as well as AI features like AI overviews in Google Search. Many recommender algorithms are now powered by large-scale transformer models like Meta’s GEM. While these companies don’t disclose hard numbers, these AI systems are likely highly lucrative and compute-intensive for Google and Meta, with both companies attributing recent growth in their traditional advertising businesses to AI advances. This category also includes product features such as AI features in Google search and a wide variety of other internal uses.24
The hyperscaler clouds all host their own model APIs for models from e.g. OpenAI and Anthropic as well as open-weight models.
What does the “China” category mean?
This is a bucketed category for the AI chips that we infer were sold to companies based in the People’s Republic of China. For example, this is true of the vast majority of the Nvidia and AMD chips designed around US export controls (e.g. the H800, H20, and MI308X), and of Huawei Ascend chips. It does not mean Chinese government ownership of compute.
Our estimates of Chinese compute do not include the smuggling of export-controlled chips. See our methodology for more discussion on potential volumes of these imports.
Do you know in which countries these chips are located?
This is outside the scope of the current phase of this project as of early 2026. A company headquartered in one country may build AI data centers elsewhere; this is fairly common for the US hyperscalers. Our Frontier AI Data Centers hub shows the location of many large AI data centers, though the total coverage is currently still a minority of the estimated global AI compute stock. And most chips owned by Chinese companies are likely physically located in China (though they can also procure offshore cloud compute).
Does this hub track deployed compute, delivered chips, or purchase orders?
Following our AI Chip Sales methodology, our headline numbers primarily come from recognized revenue by the relevant chip vendor (e.g. Nvidia, AMD, and Broadcom for Google TPUs), though in some cases (like xAI’s Colossus) we use operational data center capacity as a proxy metric. These vendors recognize revenue when they deliver hardware to their immediate customers, which can be server assemblers rather than the company actually deploying the chips. So there may be a time lag for when the server is assembled and delivered to the final customer that is not accounted for in our figures. This is on top of the time lag between hardware delivery and when chip clusters are installed and operational in data centers, which can take weeks or months.
On the flip side, our headline figures do not count unsold inventory held by Nvidia and others. Further discussion of inventory volumes can be found here.
How do you measure uncertainty and confidence intervals?
We estimate chip ownership using Monte Carlo models incorporating uncertainty in both the total volume and computing power of AI chips, and in the case of Nvidia and AMD, each company’s share of those chips. For some owners and time periods, we have significant uncertainty about how many chips they acquired (for example, most hyperscaler shares of Nvidia pre-2024, and Oracle’s pre-2025); we do our best to widen the uncertainty in share parameters when we are extrapolating or operating from limited evidence. Our headline figures are the median predictions of these models, and we sample from our models and report the 5th and 95th percentiles as confidence intervals. More details can be found in the methodology and in our code notebooks.
When comparing confidence intervals, one should consider whether the uncertainty may be correlated. For example, if our median estimate is that company A has more Nvidia compute than company B, but their Nvidia confidence intervals overlap, that does not necessarily mean it is plausible that company B has more Nvidia compute than company A. This is because the confidence intervals account for uncertainty in both the total number of Nvidia chips and each company’s share of Nvidia, so when holding the total stock of Nvidia chips constant, it is possible that the CIs will no longer overlap.
When CIs are not available for a figure, this means that we do not have a probabilistic model for that figure, not that we are completely confident in our reported estimate.
The AI Chip Owners dataset describes estimated AI chip ownership broken down by time period, chip manufacturer, owner, and chip type. We provide three views of the data: cumulative ownership by chip type, cumulative ownership by chip designer, and quarterly acquisitions by chip type.
We provide a comprehensive guide to the database’s fields below. This includes example field values as reference.
If you would like to ask any questions about the database, or request a field that should be added, feel free to contact us at data@epoch.ai.
Cumulative estimates of AI chip ownership over time, broken down by owner and chip type.
| Column | Type | Definition | Example value | Coverage |
|---|---|---|---|---|
Cumulative estimates of AI chip ownership over time, aggregated by chip designer/manufacturer. Includes power estimates.
| Column | Type | Definition | Example value | Coverage |
|---|---|---|---|---|
Estimates of AI chip acquisitions per quarter, broken down by owner and chip type.
| Column | Type | Definition | Example value | Coverage |
|---|---|---|---|---|
2026-04-15
We corrected some stale and inconsistent data in our Nvidia ownership estimates, where Oracle’s Nvidia compute was not being properly subtracted from our estimate of “Other” compute. This inflated “Other”, and the overall total, by around 1 million H100-equivalents, or around 5% of the estimated cumulative total in compute capacity. This update brings our AI Chip Owners in line with the totals in AI Chip Sales.
Download the AI Chip Sales dataset as individual CSV files for specific data types, or as a complete package containing all datasets.
Henceforth we use “Google” as shorthand for Alphabet, though Google technically does not include some Alphabet divisions like Waymo.
Nvidia also discloses in regulatory filings the revenue share of their largest (anonymized) customers, but these are generally direct customers including server OEMs and other intermediaries, so these disclosures are not very helpful here.
Their Form 10-K from January 2026 contextualizes how they estimate indirect customer revenue, which presumably applies to their hyperscaler-related claims on earnings commentary: “Indirect customer revenue is an estimation based upon multiple factors including customer purchase order information, product specifications, internal sales data, and other sources. Indirect customers primarily purchase our products through system integrators and distributors. We generate a significant amount of our revenue from a limited number of indirect customers, some individually representing 10% or more of our revenue.”
Specifically, the count of five confirms that Oracle is included in hyperscalers.
For example, per the most recent 10-K, around 20% of Nvidia’s total revenue in the past year went to Taiwan-headquartered customers, with presumably a large chunk going to the server ODM Foxconn, and Nvidia says “we estimate 76% of Data Center revenue from Taiwan-headquartered customers was attributed to end customers based in the United States and Europe”. Dell and Supermicro are also major server companies and large direct customers of Nvidia.
Omdia’s estimates of total Nvidia chip sales in 2024 are significantly lower than ours, though this may have been a partial year-to-date figure.
They are also the two largest AMD AI chip customers, but AMD is much smaller than Nvidia, and both have relatively minor custom AI chip programs. Meta has deployed hundreds of thousands of custom MTIA chips
“Even though [Google’s] TPU is just better for them to deploy, they have to deploy a crap load of GPUs because they don’t have enough TPUs to fill up their data centers.”
“You are well aware of the strong demand for AI infrastructure. But multiple segments across [Oracle Cloud Infrastructure] are also contributing to this accelerating growth rate, including cloud natives, dedicated regions, and multi-cloud.”
“The vast majority of our CapEx investments are for revenue-generating equipment that is going into our data centers and not for land, buildings or power that collectively are covered via leases. Oracle does not pay for these leases until the completed data centers and accompanying utilities are delivered to us.”
This refers to new acquisition/commitments to data center sites, not completed data centers. So this doesn’t necessarily mean Oracle built more data centers than anyone else, or even that these new leases will manifest to Oracle becoming the #1 data center company in the US at any point in the future.
Oracle is reportedly buying all of the compute for Stargate Abilene, though the site itself is owned or co-owned by Crusoe and Blue Owl. OpenAI does not have an ownership stake in Stargate Abilene to our knowledge, nor does the “Stargate” joint venture announced in January 2025, which as of early 2026 appears to be largely defunct.
Hyperscalers may buy slightly more networking than this share since smaller-scale compute deployments need less networking per chip (think individual GPUs or 8-GPU servers vs 72-GPU racks, and standalone servers vs larger clusters), but we set this aside for simplicity. Even non-hyperscaler purchases probably mostly go to commercial-scale data centers.
They would need them for data centers located physically in the PRC, but we aren’t aware of this happening. Oracle does serve AI compute to Chinese customers, but these data centers are reportedly located outside China. Nvidia has disclosed a small number of sales of H20s to non-Chinese customers.
A relevant dynamic here is that Nvidia actively supports and cultivates neoclouds, including making direct investments, in order to prevent a few powerful buyers (hyperscalers) from dominating the AI chip market. This may mitigate the pricing advantage that hyperscalers enjoy as well.
Our impression is that it is the largest neocloud by a large margin: one industry analyst describes CoreWeave as a “neocloud giant” along with Lambda, Crusoe, and Nebius. Crusoe is a collaborating partner with large data center projects like Stargate Abilene and its wholly-owned compute capacity is not easily available, while Lambda’s and Nebius’s revenue run rates are both relatively small fractions of CoreWeave’s.
Anthropic has been renting TPUs on Google Cloud for a while.
The Broadcom/Anthropic TPU deal appears to be tied to a data center we are tracking that is being developed by Fluidstack in partnership with Anthropic, though the ownership structure of this project appears complicated and a full breakdown is out of scope of this writeup.
Calculated as 960 watts of thermal design power (TDP) per chip multiplied by a typical IT power overhead factor of 1.7×, assuming that the “1 GW” measures data center IT power rather than chip power or facility power.
One sign is that Google Cloud’s and AWS’s partnership pages with AMD omit any mention of AMD Instinct GPUs, though this doesn’t preclude internal use of AMD Instinct. Google and Amazon were considering offering Instinct GPUs on their cloud platforms but decided against it after weak customer demand.
Omdia technically only reports the breakdown among four large customers; their total MI300X count closely matches our estimate of MI300X sales in 2024, but it’s not clear whether they estimate more MI300X sales to customers other than the listed four. We are not sure how reliable this single source is; we discuss Omdia further in the Nvidia methodology. But we think this estimate is a real signal pointing to large Meta and Microsoft shares of AMD in 2024.
As opposed to some smaller clouds that may or may not be participating in chip smuggling.
Note that our uncertainty here may be evidential, not causal, so one shouldn’t think too hard about the causal mechanics of Amazon buying fewer chips freeing up other hyperscalers to buy more, etc.
The estimates in the linked sheet are used in our Frontier Data Centers capital cost model.
This article gives some flavor on how Amazon’s retail division allocates its AI compute. Another article qualitatively describes Google’s struggle to allocate compute between Cloud, Google DeepMind, and other products and internal uses.