2026-03-23
Trainium model updates
Updated Trainium2 model to incorporate volume disclosures from Amazon, rather than using inferences from Project Rainier deployments. This reduced our estimate of Trainium2 volumes substantially, from ~2.5M to ~1.4M units. See methodology for more information.
2026-01-29
TPU model updates
Updated TPU volume and compute estimates with revised assumptions on production mix and margins based on expert feedback.
In summary, we now model more rapid transitions to new TPU generations. Specific changes to our production mix estimates include:
Revisions to Broadcom’s gross profit margins:
Implications:
Code refactoring
Ported our TPU and AMD models to Python Jupyter notebooks, which we previously used for Nvidia, and refactored all three models to use similar formats and helper functions. These can be found in Epoch AI’s ai-chip-counts repository.
The new models are functionally the same as before, with some specific modeling updates:
Data refactoring
Reorganized our “Timelines by chip” tables into a single table, with substantially similar columns, instead of multiple tables broken down by designer. In addition, the rows in the table are now all broken down by calendar quarter, where previously they were often broken down by fiscal quarters that do not line up with calendar quarters (e.g. for Google TPU and Nvidia) or other non-quarter periods such as years or half-years. This does not affect our graph views, which previously already interpolated the table results into calendar quarters.
Our models still output intermediate results broken down by fiscal quarter when they are distinct from calendar quarters, which you can view in the respective code notebooks (see full methodology).