AI Supercomputers

Our database of over 500 supercomputers (also known as computing clusters) tracks large hardware facilities for AI training and inference and maps them across the globe.

By Konstantin Pilz, Robi Rahman, James Sanders and Lennart Heim

Last updated July 15, 2025

Disclaimer: Our dataset covers an estimated 10–20% of existing global aggregate AI supercomputer performance as of March 2025. Planned systems are subject to changes and inherently lower confidence. While coverage varies across companies, sectors, and hardware types due to uneven reporting, we believe the overall distribution remains broadly representative. Future country shares may change dramatically as exponential growth continues. Chinese systems are anonymized and their specifications are rounded.

Data insights

Selected insights from this dataset.

See all our insights

The computational performance of leading AI supercomputers has doubled every nine months

The computational performance of the leading AI supercomputers has grown by 2.5x annually since 2019. This has enabled vastly more powerful training runs: if 2020’s GPT-3 were trained on xAI’s Colossus, the original two week training run could be completed in under 2 hours.

This growth was enabled by two factors: the number of chips deployed per cluster has increased by 1.6x per year, and performance per chip has also improved by 1.6x annually.

Learn more

Acquisition costs of leading AI supercomputers have doubled every 13 months

AI supercomputers have become increasingly expensive. Since 2019, the cost of the computing hardware for leading supercomputers has increased at a rate of 1.9x per year. In June 2022, the most expensive cluster was Oak Ridge National Laboratory Frontier, with a reported cost of $200M. Three years later, as of June 2025, the most expensive supercomputer is xAI’s Colossus, estimated to use over $7B of hardware.

Learn more

Power requirements of leading AI supercomputers have doubled every 13 months

Leading AI supercomputers are becoming ever more energy-intensive, using more power-hungry chips in greater numbers. In January 2019, Summit at Oak Ridge National Lab had the highest power capacity of any AI supercomputer at 13 MW. Today, xAI’s Colossus supercomputer uses 280 MW, over 20x as much.

Colossus relies on mobile generators because the local grid has insufficient power capacity for so much hardware. In the future, we may see frontier models trained across geographically distributed supercomputers, to mitigate the difficulty of delivering enormous amounts of power to a single location, similar to the training setup for Gemini 1.0.

Learn more

Private-sector companies own a dominant share of AI supercomputers

The private sector’s share of global AI computing capacity has grown from 40% in 2019 to 80% in 2025. Though many leading early supercomputers such as Summit were run by government and academic labs, the total installed computing power of public-sector supercomputers has only increased at 1.8x per year, rapidly outpaced by private-sector supercomputers, whose total computing power has grown at 2.7x per year. The rising economic importance of AI has spurred the private sector to build more and faster supercomputers for training and inference.

As of May 2025, the largest known public AI supercomputer, Lawrence Livermore’s El Capitan, achieves less than a quarter of the computational performance of the largest known industry AI supercomputer, xAI’s Colossus.

Learn more

The US hosts the majority of AI supercomputers, followed by China

As of May 2025, the United States contains about three-quarters of global AI supercomputer performance, with China in second place with 15%. Meanwhile, traditional high-performance computing leaders like Germany, Japan, and France now play marginal roles in the AI supercomputing landscape. This shift largely reflects the increased dominance of major technology companies, which are predominantly based in the United States.

Learn more

FAQ

What is an AI supercomputer?

A supercomputer is a system of many processors (such as CPUs and GPUs) that can efficiently work together and achieve a high level of performance. When supercomputers use specialized AI chips and support large-scale AI training and deployment workloads, we refer to them as AI supercomputers. They are sometimes also referred to as “GPU clusters” or “AI datacenters”.

AI supercomputers are used for training or serving neural network models. Therefore, they typically support number formats favorable for AI training and inference, such as FP16 or INT8, contain compute units optimized for matrix multiplication, have high-bandwidth memory, and rely on AI accelerators rather than CPUs for most of their calculations. A more detailed definition can be found in our documentation and in section 2 of our paper.

How do you measure performance for AI supercomputers?

We provide the total computational rate of the hardware in the computing cluster, which is the performance of each ML hardware chip times the number of chips.

Non-AI supercomputers are often evaluated using the LINPACK benchmark, which measures the speed of serial computations in FP64 arithmetic format. However, AI workloads use parallelized computations and smaller number representations, so benchmarks like LINPACK are not an ideal measure of their performance. Furthermore, most organizations do not measure their performance on LINPACK or other benchmarks such as MLPerf, so we do not record these.

Which types of organizations own AI supercomputers?

Historically, government and academic research organizations such as Oak Ridge National Laboratory and Sunway have owned many of the top supercomputers. In recent years, most AI supercomputers are owned by cloud computing providers such as AWS, Google, and Microsoft Azure, or AI laboratories such as Meta and xAI.

How was the AI supercomputers dataset created?

The data was primarily collected from machine learning papers, publicly available news articles, press releases, and existing lists of supercomputers.

We created a list of potential supercomputers by using the Google Search API to search key terms like “AI supercomputer” and “GPU cluster” from 2019 to 2025, then used GPT-4o to extract any supercomputers mentioned in the resulting articles. We also added supercomputers from publicly available lists such as Top500 and MLPerf, and GPU rental marketplaces. For each potential supercomputer, we manually searched for public information such as number and type of chips used, when it was first operational, reported performance, owner, and location. A detailed description of our methods can be found in the documentation and Appendix A of our paper.

How do you estimate details like performance, cost, or power usage?

Performance is sometimes reported by the owner of the cluster, or in news reports. Otherwise, it is calculated based on the performance per chip of the hardware used in the cluster, times the number of chips.

Costs are sometimes reported by the owner or sponsor of the cluster. Otherwise, costs are estimated from the cost per chip of the hardware, times the number of chips, multiplied by adjustment factors for intra- and inter-server network hardware.

Power draw is sometimes reported by the owner or operator of the cluster. Otherwise, it is estimated from the power draw per chip, times the number of chips, multiplied by adjustment factors for other hardware and the power usage efficiency of the datacenter.

Detailed methodology definitions can be found in the paper and documentation.

How accurate is the information about each supercomputer?

We strive to accurately convey the reported specifications of each supercomputer. The Status field indicates our assessment of whether the supercomputer is currently operational, not yet operational, or decommissioned. The Certainty field indicates our assessment of the likelihood that the cluster exists in roughly the form specified in the dataset and the linked sources. If you find mistakes or additional information regarding any supercomputers in the dataset, please email data@epoch.ai.

How is the dataset licensed?

We have released a public dataset with a CC-BY license. This public dataset includes all of our data on supercomputers outside of China and Hong Kong, along with anonymized data on supercomputers within China, with values rounded to one significant figure and names and links removed. This dataset is free to use, distribute, and reproduce provided the source and authors are credited under the Creative Commons Attribution license.

Data on Chinese supercomputers is stored privately to protect the data sources. For inquiries about this data, please contact Robi Rahman at robi@epoch.ai.

How up-to-date is the data?

Although we strive to maintain an up-to-date database, new supercomputers are constantly under construction, so there will inevitably be some that have not yet been added. Generally, major supercomputers should be added within one month of their announcement, and others are added periodically during reviews. If you notice a missing supercomputer, you can notify us at data@epoch.ai.

How can I access this data?

Download the data in CSV format.
Explore the data using our interactive tools.
View the data directly in a table format.

Who can I contact with questions or comments about the dataset?

For general inquiries about the project and the paper, please reach out to Konstantin at kfp15@georgetown.edu. For inquiries about the dataset, please contact Robi at robi@epoch.ai.

Documentation

This dataset tracks AI supercomputers, identified from sources including model training reports, news articles, press releases, and web search results. Additional information about our approach to identifying supercomputers and collecting data about them can be found in the accompanying documentation.

Read the complete documentation

Use this work

Licensing

Epoch AI's data is free to use, distribute, and reproduce provided the source and authors are credited under the Creative Commons Attribution license.

Citation

Konstantin Pilz, Robi Rahman, James Sanders, Lennart Heim, ‘Trends in AI Supercomputers’. Published online at epoch.ai. Retrieved from ‘https://epoch.ai/data/ai-supercomputers’ [online resource]. Accessed .”

BibTeX Citation

@misc{EpochAISupercomputers2025,
  title = {Trends in AI Supercomputers},
  author = {Konstantin Pilz, Robi Rahman, James Sanders, Lennart Heim},
  year = {2025},
  month = {04},
  url = {https://epoch.ai/data/ai-supercomputers},
  note = {Accessed: }
}

Download this data

AI Supercomputers

CSV, Updated July 15, 2025