Compute Demand Driving Growth of AI Datacenters
AI’s appetite for compute — from training giant models to powering daily inference — is exploding, and that growth is expected to accelerate all the way to 2030. This surge is triggering a global race to build massive new data centers packed with GPUs, power infrastructure, and capital. The big question now is whether supply can keep up. Optimists believe aggressive investment and rapid innovation will stay ahead of demand. But if the buildout slows or bottlenecks hit, we could be staring at a serious compute crunch by the second half of the decade.
In this blog, we’ll explore some intriguing taxonomy, facts, and stats around this emerging landscape.
Training and Inference Compute
Training Compute
AI Training Compute is the heavy, one-off computational effort to "teach" a model using large datasets. It's highly intensive, often requiring thousands of accelerators running for weeks.
- Training: GPT-4 (2023) used ~70x more compute than GPT-3 (2020), costing an estimated $80-100 million for that single training run.
 
Training is the heavy one-off investment to build the model; Inference is the continuous cost to run it in the real world.
AI Inference Compute is the continuous, ongoing cost of deploying that model for users. It happens *every time* you make a query or the model generates an output.
- Inference: By 2025, ChatGPT handles 2.5 billion queries per day. This consumes ~850 MWh daily, equivalent to the daily charging of 14,000 electric vehicles.
 
Inference has become "the center of gravity" for AI infrastructure growth—deploying models for real-time use is the dominant factor.
Growing Demand Projections (2024-2030)
The demand for AI compute is projected to grow explosively, driving massive increases in energy consumption and market spending.
Computational Facts
How to estimate the compute needed for training an LLM?
A common rule of thumb is:
Training FLOPs ≈ 6 × (Params) × (Tokens in Training Data)
How may GPUs did GPT-4's training take?
An estimated 2.1 × 10^25 FLOPs (70× more than GPT-3!). To train this in one week (at 30% utilization) would require roughly:
- 29,000 NVIDIA H200 or 250,000 Google TPU v5p GPUs
 
How much power is consumed per ChatGPT query?
An rough estimate of 0.34 Wh.
An equivalent of 12 seconds of running a ceiling fan (120 W) or ~2 minutes of streaming YouTube on a smartphone.
What's the future power demand for AI look like in a real-world equivalent?
By 2030, the projected growth in AI inference alone could require building ~38 new 1-GW class data centers,
an equivalent of 40 new nuclear reactors.
Disclaimer: This blog and dashboard are intended solely for educational purposes. While every effort is made to source data from reliable and publicly available platforms, certain datasets, charts, or insights may contain inaccuracies, or discrepancies. If you find any issues or need any clarification, please write to the Author