AI chips

AI chips are the specialized hardware behind modern AI, designed to handle the massive computational demands of training and running advanced models. They are at the center of a global competition for compute, with performance improving rapidly and demand surging. Epoch tracks trends in AI chip performance, energy efficiency, and price-performance over time, as well as the supply chain dynamics and geopolitical factors shaping who has access to the most advanced hardware.

Filter

Type
0 results
What you need to know about AI chips
Topic Overview
May 1, 2026
What you need to know about AI chips

A look at the specialized hardware driving modern AI — why chips cost tens of thousands of dollars each, and why demand continues to outstrip supply.

Diversion and resale: estimating compute smuggling to China
Report
Apr. 29, 2026
Diversion and resale: estimating compute smuggling to China

We estimate that between 290,000 and 1.6 million H100-equivalents (H100e) were smuggled to China through 2025. Our median estimate of 660,000 H100e would be roughly a third of China's total compute.

By Isabel Juniewicz

Five hyperscalers now own over two-thirds of global AI compute
Data Insight
Apr. 14, 2026
Five hyperscalers now own over two-thirds of global AI compute

By Luke Emberson, Josh You, and Venkat Somala

What does the war in Iran mean for AI?
Newsletter
Apr. 10, 2026
What does the war in Iran mean for AI?

A prolonged Hormuz crisis probably won't derail the compute buildout, but it could slow data center expansion and disrupt Gulf investment flows into AI.

By Josh You

Google controls the most AI computing power, driven by its custom TPUs
Data Insight
Apr. 7, 2026
Google controls the most AI computing power, driven by its custom TPUs

By Luke Emberson, Josh You, and Venkat Somala

Introducing the AI Chip Owners Explorer
Update
Apr. 6, 2026
Introducing the AI Chip Owners Explorer

We announce our new AI Chip Owners explorer, showing which companies own the world’s leading AI chips.

By Josh You and Venkat Somala

Total AI chip memory bandwidth has grown 4.1x per year, now reaching 70 million terabytes per second
Data Insight
Mar. 24, 2026
Total AI chip memory bandwidth has grown 4.1x per year, now reaching 70 million terabytes per second

By Luke Emberson

Advanced packaging and HBM, not logic dies, were the bottlenecks on AI chip production in 2025
Data Insight
Mar. 12, 2026
Advanced packaging and HBM, not logic dies, were the bottlenecks on AI chip production in 2025

By Venkat Somala

Introducing the AI Chip Sales Data Explorer
Update
Jan. 13, 2026
Introducing the AI Chip Sales Data Explorer

We announce our new AI Chip Sales data explorer, which uses financial reports, company disclosures, and more to estimate compute, power usage, and spending over time for a wide variety of AI chips.

By The Epoch AI Team

Global AI computing capacity is doubling every 7 months
Data Insight
Jan. 9, 2026
Global AI computing capacity is doubling every 7 months

By Josh You, Venkat Somala, Yafah Edelman, and Luke Emberson

GPUs account for about 40% of power usage in AI data centers
Data Insight
Dec. 18, 2025
GPUs account for about 40% of power usage in AI data centers

By Luke Emberson and Ben Cottier

NVIDIA’s B200 costs around $6,400 to produce, with memory accounting for half
Data Insight
Dec. 10, 2025
NVIDIA’s B200 costs around $6,400 to produce, with memory accounting for half

By Venkat Somala

Compute scaling will slow down due to increasing lead times
Newsletter
Sep. 5, 2025
Compute scaling will slow down due to increasing lead times

A heavily underappreciated dynamic when thinking about AI timelines.

By Yafah Edelman and Anson Ho

Frontier AI performance becomes accessible on consumer hardware within a year
Data Insight
Aug. 15, 2025
Frontier AI performance becomes accessible on consumer hardware within a year

By Venkat Somala and Luke Emberson

Newsletter
Jul. 26, 2025
Why China isn’t about to leap ahead of the West on compute

Chinese hardware is closing the gap, but major bottlenecks remain

By Veronika Blablová and Robi Rahman

Widespread adoption of new numeric formats took 3-4 years in past cycles
Data Insight
May 28, 2025
Widespread adoption of new numeric formats took 3-4 years in past cycles

By Venkat Somala and Luke Emberson

Trends in AI supercomputers
Paper
Apr. 23, 2025
Trends in AI supercomputers

AI supercomputers double in performance every 9 months, cost billions of dollars, and require as much power as mid-sized cities. Companies now own 80% of all AI supercomputers, while governments’ share has declined.

By Konstantin F. Pilz, Robi Rahman, James Sanders, and Lennart Heim

Leading AI chip designs are used for around four years in frontier training
Data Insight
Mar. 5, 2025
Leading AI chip designs are used for around four years in frontier training

By Luke Emberson, Ben Snodin, and David Owen

The stock of computing power from NVIDIA chips is doubling every 10 months
Data Insight
Feb. 13, 2025
The stock of computing power from NVIDIA chips is doubling every 10 months

By Luke Emberson and David Owen

Training compute growth is driven by larger clusters, longer training, and better hardware
Data Insight
Jan. 8, 2025
Training compute growth is driven by larger clusters, longer training, and better hardware

By Luke Emberson and David Owen

Newsletter
Dec. 6, 2024
What did US export controls mean for China’s AI capabilities?

Export controls on China give the US a hardware lead of around 4 years in training frontier models, but essentially no lead in serving those models to users.

By Ege Erdil

Hardware failures won’t limit AI scaling
Report
Nov. 22, 2024
Hardware failures won’t limit AI scaling

Our analysis shows hardware failures won't limit AI training scale. GPU memory-based checkpointing enables training beyond millions of GPUs.

By Alexander Erben and Ege Erdil

Data movement bottlenecks to large-scale model training: Scaling past 1e28 FLOP
Paper
Nov. 2, 2024
Data movement bottlenecks to large-scale model training: Scaling past 1e28 FLOP

Data movement bottlenecks limit LLM scaling beyond 2e28 FLOP, with a "latency wall" at 2e31 FLOP. We may hit these in ~3 years. Aggressive batch size scaling could potentially overcome these limits.

By Ege Erdil

Introducing Epoch AI's machine learning hardware database
Update
Oct. 23, 2024
Introducing Epoch AI's machine learning hardware database

Our new database covers hardware used to train AI models, featuring over 100 accelerators (GPUs and TPUs) across the deep learning era.

By The Epoch AI Team

Performance improves 13x when switching from FP32 to tensor-INT8
Data Insight
Oct. 23, 2024
Performance improves 13x when switching from FP32 to tensor-INT8

By Robi Rahman and David Owen

Leading ML hardware becomes 40% more energy-efficient each year
Data Insight
Oct. 23, 2024
Leading ML hardware becomes 40% more energy-efficient each year

By Robi Rahman

The NVIDIA A100 has been the most popular hardware for training notable machine learning models
Data Insight
Oct. 23, 2024
The NVIDIA A100 has been the most popular hardware for training notable machine learning models

By Robi Rahman

The computational performance of machine learning hardware has doubled every 2.3 years
Data Insight
Oct. 23, 2024
The computational performance of machine learning hardware has doubled every 2.3 years

By Robi Rahman

Performance per dollar improves around 30% each year
Data Insight
Oct. 23, 2024
Performance per dollar improves around 30% each year

By Robi Rahman

Leading AI companies have hundreds of thousands of cutting-edge AI chips
Data Insight
Oct. 9, 2024
Leading AI companies have hundreds of thousands of cutting-edge AI chips

By Josh You and David Owen

Can AI scaling continue through 2030?
Report
Aug. 20, 2024
Can AI scaling continue through 2030?

We investigate the scalability of AI training runs. We identify electric power, chip manufacturing, data and latency as constraints. We conclude that 2e29 FLOP training runs will likely be feasible by 2030.

By Jaime Sevilla, Tamay Besiroglu, Ben Cottier, Josh You, Edu Roldán, Pablo Villalobos, and Ege Erdil

Limits to the energy efficiency of CMOS microprocessors
Paper
Dec. 15, 2023
Limits to the energy efficiency of CMOS microprocessors

How far can the energy efficiency of CMOS microprocessors be pushed before we hit physical limits? Using a simple model, we find that there is room for a further 50 to 1000x improvement in energy efficiency.

By Anson Ho, Ege Erdil, and Tamay Besiroglu

Trends in machine learning hardware
Report
Nov. 9, 2023
Trends in machine learning hardware

FLOP/s performance in 47 ML hardware accelerators doubled every 2.3 years. Switching from FP32 to tensor-FP16 led to a further 10x performance increase. Memory capacity and bandwidth doubled every 4 years.

By Marius Hobbhahn, Lennart Heim, and Gökçe Aydos

Predicting GPU performance
Report
Dec. 1, 2022
Predicting GPU performance

We develop a simple model that predicts progress in the performance of field-effect transistor-based GPUs under the assumption that transistors can no longer miniaturize after scaling down to roughly the size of a single silicon atom. Our model forecasts that the current paradigm of field-effect transistor-based GPUs will plateau sometime between 2027 and 2035, offering a performance of between 1e14 and 1e15 FLOP/s in FP32.

By Marius Hobbhahn and Tamay Besiroglu

Trends in GPU price-performance
Report
Jun. 27, 2022
Trends in GPU price-performance

Using a dataset of 470 models of graphics processing units released between 2006 and 2021, we find that the amount of floating-point operations/second per $ doubles every ~2.5 years.

By Marius Hobbhahn and Tamay Besiroglu