The computational performance of machine learning hardware has doubled every 1.9 years

Measured in 16-bit floating point operations, ML hardware performance has increased at a rate of 43% per year, doubling every 1.9 years. A similar trend exists in 32-bit performance. Optimized ML number formats and tensor cores provided additional improvements.

The improvement was driven by increasing transistor counts and other semiconductor manufacturing improvements, as well as specialized design for AI workloads. This improvement lowered cost per FLOP, increased energy efficiency, and enabled large-scale AI training.

Published

October 23, 2024

Last updated

February 21, 2025

Epoch’s work is free to use, distribute, and reproduce provided the source and authors are credited under the Creative Commons BY license.