All publications

Frontier AI performance becomes accessible on consumer hardware within a year
How Much Power Will Frontier AI Training Demand in 2030?
The power required to train the largest frontier models is growing by more than 2x per year, and is on trend to reaching multiple gigawatts by 2030.
Compute is not a bottleneck for robotic manipulation
We didn’t learn much from the IMO
Quantifying the algorithmic improvement from reasoning models
Training open-weight models is becoming more data intensive
Why China isn’t about to leap ahead of the West on compute
Frontier training runs will likely stop getting longer by around 2027
Evaluating Grok 4’s Math Capabilities
It's good at involved computations, improving at proofs, and useful for literature search. It still favors low-level grinds and leans on background knowledge.
After the ChatGPT Moment: Measuring AI’s Adoption
How to run SWE-bench Verified in one hour on one machine
We are releasing a public registry of optimized Docker images for SWE-bench. This allows us to run SWE-bench Verified in 62 minutes on a single GitHub actions VM.
What will the IMO tell us about AI math capabilities?
How big could an “AI Manhattan Project” get?
LLMs now accept longer inputs, and the best models can use them more effectively
AI and explosive growth redux
Inference Economics of Language Models
We investigate how speed trades off against cost in language model inference. We find that inference latency scales with the square root of model size and the cube root of memory bandwidth, and other results.
Do the biorisk evaluations of AI labs actually measure the risk of developing bioweapons?
What skills does SWE-bench Verified evaluate?
We take a deep dive into SWE-bench Verified, a prominent agentic coding benchmark. While one of the best public tests of AI coding agents, it is limited by its focus on simple bug fixes in familiar open-source repositories.
LLM providers offer a trade-off between accuracy and speed
Over 30 AI models have been trained at the scale of GPT-4
Beyond benchmark scores: Analyzing o3-mini’s mathematical reasoning
Power requirements of leading AI supercomputers have doubled every 13 months
Private-sector companies own a dominant share of GPU clusters
The US hosts the majority of GPU cluster performance, followed by China
Acquisition costs of leading AI supercomputers have doubled every 13 months
The computational performance of leading AI supercomputers has doubled every nine months
What is Epoch?
Our director explains Epoch AI’s mission and how we decide our priorities. In short, we work on projects to understand the trajectory of AI, share this knowledge publicly, and inform important decisions about AI.
GPQA Diamond: What’s left?
How Many AI Models Will Exceed Compute Thresholds?
We project how many notable AI models will exceed training compute thresholds. Model counts rapidly grow from 10 above 1e26 FLOP by 2026, to over 200 by 2030.
Widespread adoption of new numeric formats took 3-4 years in past cycles
Is AI already superhuman on FrontierMath?
How fast can algorithms advance capabilities?
How far can reasoning models scale?
Where’s my ten minute AGI?
The case for multi-decade AI timelines
Trends in AI Supercomputers
AI supercomputers double in performance every 9 months, cost billions of dollars, and require as much power as mid-sized cities. Companies now own 80% of all AI supercomputers, while governments’ share has declined.
LLM responses to benchmark questions are getting longer over time
The combined revenues of leading AI companies grew by over 9x in 2023-2024
Is it 3 Years, or 3 Decades Away? Disagreements on AGI Timelines
In this podcast episode, two Epoch AI researchers with relatively long and short AGI timelines candidly examine the roots of their disagreements.
The real reason AI benchmarks haven’t reflected economic impacts
Most AI value will come from broad automation, not from R&D
GATE: Modeling the Trajectory of AI and Automation
We introduce a compute-centric model of AI automation and its economic effects, illustrating key dynamics of AI development. The model suggests large AI investments and subsequent economic growth.
FrontierMath Competition: Setting Benchmarks for AI Evaluation
We are hosting a competition to establish rigorous human performance baselines for FrontierMath. With a prize pool of $10,000, your participation will contribute directly to measuring AI progress in solving challenging mathematical problems.
LLM inference prices have fallen rapidly but unequally across tasks
What AI can currently do is not the story
Train Once, Deploy Many: AI and Increasing Returns
AI's “train-once-deploy-many” advantage yields increasing returns: doubling compute more than doubles output by increasing models' inference efficiency and enabling more deployed inference instances.
Leading AI chip designs are used for around four years in frontier training
The promise of reasoning models
Biology AI models are scaling 2-4x per year after rapid growth from 2019-2021
AI progress is about to speed up
Algorithmic progress likely spurs more spending on compute, not less
The stock of computing power from NVIDIA chips is doubling every 10 months
US models currently outperform non-US models
Models with downloadable weights currently lag behind the top-performing models
Accuracy increases with estimated training compute
How much energy does ChatGPT use?
A more systematic and transparent AI Benchmarking Hub
We've overhauled our AI benchmarking infrastructure to provide more transparent, systematic, and up-to-date evaluations of AI model capabilities.
What went into training DeepSeek-R1?
Announcing our Expanded Biology AI Coverage
We've expanded our Biology AI Dataset, now covering 360+ models. Our analysis reveals rapid scaling from 2017-2021, followed by a notable slowdown in biological model development.
AGI could drive wages below subsistence level
Clarifying the Creation and Use of the FrontierMath Benchmark
We clarify that OpenAI commissioned Epoch AI to produce 300 math questions for the FrontierMath benchmark. They own these and have access to the statements and solutions, except for a 50-question holdout set.
Chinese language models have scaled up more slowly than their global counterparts
AI in 2030, Scaling Bottlenecks, and Explosive Growth
Epoch AI presents their first podcast, exploring AI scaling trends, discussing power demands, chip production, data needs, and how continued progress could transform labor markets and potentially accelerate global economic growth to unprecedented levels.
How has DeepSeek improved the Transformer architecture?
2024 Impact Report
Epoch's Impact Report for 2024 highlights influential research on AI's trajectory, the launch of FrontierMath, an expanded AI data hub, engagement with leaders, $7M raised, and more.
Frontier open models may surpass 10²⁶ FLOP of training compute before 2026
The economic consequences of automating remote work
Training compute growth is driven by larger clusters, longer training, and better hardware
Moravec’s paradox and its implications
How do mixture-of-experts models compare to dense models in inference?
Frontier language models have become much smaller
Announcing Gradient Updates: Our New Weekly Newsletter
We are announcing Gradient Updates, Epoch AI’s new weekly newsletter focused on timely and important questions in AI.
What did US export controls mean for China’s AI capabilities?
What is the Future of AI in Mathematics? Interviews with Leading Mathematicians
How will AI transform mathematics? Fields Medalists and other leading mathematicians discuss whether they expect AI to automate advanced math research.
Introducing the Distributed Training Interactive Simulator
We introduce and walk you through an interactive tool that simulates distributed training runs of large language models under ideal conditions.
Introducing Epoch AI's AI Benchmarking Hub
We are launching the AI Benchmarking Hub: a platform presenting our evaluations of leading models on challenging benchmarks, with analysis of trends in AI capabilities.
Hardware Failures Won’t Limit AI Scaling
Hardware failures won't limit AI training scale - GPU memory checkpointing enables training with millions of GPUs despite failures.
FrontierMath: A Benchmark for Evaluating Advanced Mathematical Reasoning in AI
FrontierMath: a new benchmark of expert-level math problems designed to measure AI's mathematical abilities. See how leading AI models perform against the collective mathematics community.
How Far Behind Are Open Models?
Analysis of open vs. closed AI models reveals the best open model today matches closed models in performance and training compute, but with a one-year lag.
Data Movement Bottlenecks to Large-Scale Model Training: Scaling Past 1e28 FLOP
Data movement bottlenecks limit LLM scaling beyond 2e28 FLOP, with a "latency wall" at 2e31 FLOP. We may hit these in ~3 years. Aggressive batch size scaling could potentially overcome these limits.
AI training cluster sizes increased by more than 20x since 2016
Performance per dollar improves around 30% each year
The computational performance of machine learning hardware has doubled every 2.4 years
The NVIDIA A100 has been the most popular hardware for training notable machine learning models
Leading ML hardware becomes 40% more energy-efficient each year
Performance improves 12x when switching from FP32 to tensor-INT8
Introducing Epoch AI’s Machine Learning Hardware Database
Our new database covers hardware used to train AI models, featuring over 100 accelerators (GPUs and TPUs) across the deep learning era.
Leading AI companies have hundreds of thousands of cutting-edge AI chips
The power required to train frontier AI models is doubling annually
Interviewing AI researchers on automation of AI R&D
AI could speed up AI R&D, especially in coding and debugging. We explore predictions on automation and researchers' suggestions for AI R&D evaluations.
Can AI Scaling Continue Through 2030?
We investigate four constraints to scaling AI training: power, chip manufacturing, data, and latency. We predict 2e29 FLOP runs will be feasible by 2030.
The length of time spent training notable models is growing
Language models compose the large majority of large-scale AI models
Most large-scale models are developed by US companies
The pace of large-scale model releases is accelerating
Almost half of large-scale models have published, downloadable weights
The size of datasets used to train language models doubles approximately every six months
Training compute costs are doubling every eight months for the largest AI models
The training compute of notable AI models has been doubling roughly every six months
Training compute has scaled up faster for language than vision
Announcing Epoch AI’s Data Hub
We're launching a hub for data and visualizations, featuring our databases on notable and large-scale AI models for users and researchers.
Will We Run Out of Data? Limits of LLM Scaling Based on Human-Generated Data
If trends continue, language models will fully utilize the stock of human-generated public text between 2026 and 2032.
How Much Does It Cost to Train Frontier AI Models?
The cost of training top AI models has grown 2-3x annually for the past eight years. By 2027, the largest models could cost over a billion dollars.
Training Compute of Frontier AI Models Grows by 4-5x per Year
Our expanded AI model database shows that training compute grew 4-5x/year from 2010 to 2024, with similar trends in frontier and large language models.
Do the Returns to Software R&D Point Towards a Singularity?
Returns to R&D are key in growth dynamics and AI development. Our paper introduces new empirical techniques to estimate this vital parameter.
Chinchilla Scaling: A Replication Attempt
We replicate Hoffmann et al.’s parametric scaling law estimates, finding issues and providing better-fitting estimates that align with their other methods.
Tracking Large-Scale AI Models
We present a dataset of 81 large-scale models, from AlphaGo to Gemini, developed across 18 countries, at the leading edge of scale and capabilities.
Optimally Allocating Compute Between Inference and Training
AI labs should spend comparable resources on training and inference, assuming they can flexibly balance compute between the two to maintain performance.
Algorithmic Progress in Language Models
Progress in pretrained language model performance outpaces expectations, occurring at a pace equivalent to doubling computational power every 5 to 14 months.
2023 Impact Report
In 2023, Epoch published nearly 20 reports on AI, added hundreds of models to our database, helped with government policies, and raised over $7 million.
Biological Sequence Models in the Context of the AI Directives
Our expanded database now includes biological sequence models, highlighting potential regulatory gaps and the growth of training compute in these models.
How Predictable Is Language Model Benchmark Performance?
We investigate large language model performance, finding that compute-focused extrapolations are a promising way to forecast AI capabilities.
Limits to the Energy Efficiency of CMOS Microprocessors
How far can the energy efficiency of CMOS microprocessors be pushed before hitting physical limits? We find room for a further 50 to 1000x improvement.
AI Capabilities Can Be Significantly Improved Without Expensive Retraining
While scaling compute is key to improving LLMs, post-training enhancements can offer gains equivalent to 5-20x more compute at less than 1% of the cost.
Who Is Leading in AI? An Analysis of Industry AI Research
Industry has emerged as a driving force in AI. We compare top companies on research impact, training runs, and contributions to algorithmic innovations.
Challenges in Predicting AI Automation
Economists propose various approaches to predicting AI's automation of valuable tasks, but disagreements persist, with no consensus on the best method.
Trends in Machine Learning Hardware
FLOP/s performance in 47 ML hardware accelerators doubled every 2.3 years. Switching from FP32 to tensor-FP16 led to a further 10x performance increase.
Announcing Epoch AI’s Updated Parameter, Compute and Data Trends Database
Our database now spans over 700 ML systems, tracking parameters, datasets, and training compute details for notable machine learning models.
Explosive Growth from AI: A Review of the Arguments
Our new article explores whether deployment of advanced AI systems could lead to growth rates ten times higher than those of today’s frontier economies.
Trading Off Compute in Training and Inference
We characterize techniques that induce a tradeoff between spending resources on training and inference, outlining their implications for AI governance.
The Limited Benefit of Recycling Foundation Models
Reusing pretrained models can save on training costs, but it's unlikely to significantly boost AI capabilities beyond modest improvements.
Epoch AI and FRI Mentorship Program Summer 2023
We’re launching the Epoch and FRI mentorship program for women, non-binary, and transgender people interested in AI forecasting.
Direct Approach Interactive Model
When could transformative AI be achieved? We present a simple, user-adjustable model of key inputs that forecasts the date TAI could be deployed.
A Compute-Based Framework for Thinking About the Future of AI
AI’s potential to automate labor could alter the course of human history. The availability of compute is the most important factor driving progress in AI.
Please Report Your Compute
Compute is essential for AI performance, yet often underreported. Adopting reporting norms would improve research, forecasts, and policy decisions.
The Direct Approach
We propose a method using neural scaling laws to estimate the compute needed to train AI models to reach human-level performance on various tasks.
Power Laws in Speedrunning and Machine Learning
Our model suggests ML benchmarks aren’t near saturation. While large improvements are rare, we find 1OOM gains happen roughly once in every 50 instances.
Announcing Epoch AI’s Dashboard of Key Trends and Figures in Machine Learning
Our dashboard provides key data from our research on machine learning and is a valuable resource for understanding the present and future of the field.
2022 Impact Report
Our impact report for 2022.
Trends in the Dollar Training Cost of Machine Learning Systems
How much does it cost to train AI models? Looking at 124 ML systems from between 2009 and 2022, we find the cost has grown by approximately 0.5OOM/year.
Scaling Laws Literature Review
I have collected a database of scaling laws for different tasks and architectures, and reviewed dozens of papers in the scaling law literature.
An Interactive Model of AI Takeoff Speeds
We have developed an interactive website showcasing a new model of AI takeoff speeds.
Literature Review of Transformative Artificial Intelligence Timelines
We summarize and compare several models and forecasts predicting when transformative AI will be developed.
Revisiting Algorithmic Progress
Examining over 100 computer vision models, we find that every 9 months, better algorithms contribute the equivalent of a doubling of compute budgets.
Will We Run Out of ML Data? Evidence From Projecting Dataset Size Trends
We project dataset growth in language and vision domains, estimating future limits to training by evaluating the availability of unlabeled data over time.
The Longest Training Run
Training runs of large ML systems will likely last less than 14-15 months, as shorter runs starting later use better hardware and algorithms.
A Time-Invariant Version of Laplace’s Rule
We discuss estimating event probabilities with past data, addressing issues with Laplace’s rule and proposing a modification to improve accuracy.
Machine Learning Model Sizes and the Parameter Gap
Since 2018, the size of ML models has been growing 10 times faster than before. Around 2020, model sizes saw a significant jump, increasing by 1OOM.
Trends in GPU Price-Performance
Improvements in hardware are central to AI progress. Using data on 470 GPUs from 2006 to 2021, we find that FLOP/s per dollar doubles every ~2.5 years.
Announcing Epoch AI: A Research Initiative Investigating the Road to Transformative AI
We are a new research initiative forecasting developments in AI. Come join us!
Compute Trends Across Three Eras of Machine Learning
We’ve compiled a comprehensive dataset of the training compute of AI models, providing key insights into AI development.
Estimating Training Compute of Deep Learning Models
We describe two approaches for estimating the training compute of Deep Learning systems, by counting operations and looking at GPU time.
What’s the Backward-Forward FLOP Ratio for Neural Networks?
Determining the backward-forward FLOP ratio for neural networks, to help calculate their total training compute.
How to Measure FLOP for Neural Networks Empirically?
Computing the utilization rate for multiple Neural Network architectures.