TL;DR: AI data centers consumed roughly 500 TWh of electricity in 2025 — equivalent to France's entire annual power usage. The IEA projects that number doubles by 2030. Microsoft pledged $80B+ in data center construction for 2025 alone. OpenAI signed a 20-year nuclear deal for a restarted Three Mile Island reactor. Grid operators in Virginia, Ireland, and the Netherlands are already signaling capacity constraints through 2030. Models are getting more efficient — but history says efficiency gains get absorbed by expanded deployment. The grid is not a software problem.
The AI boom has a power problem nobody wants to put at the front of a press release. The language of AI progress — faster models, lower latency, expanded context windows — is the language of computation. And computation, at the scale now being deployed, is not free. It consumes electricity. A lot of it. More than most people in the industry have been willing to say plainly.
This article is the plain version.
The Numbers: Where We Are in 2026
The International Energy Agency estimates that global data center electricity consumption reached approximately 500 TWh in 2025. To put that in context: France's entire national power consumption — homes, industry, transportation — sits around 450 TWh per year. Data centers now consume more. And the IEA's baseline projection for 2030 is roughly double that figure, driven primarily by the expansion of AI infrastructure.
These numbers do not represent a steady state. They represent an acceleration. The AI industry's compute requirements are not growing linearly — they are growing with each new model generation, each new use case, each new enterprise deployment. The power infrastructure being built or planned today will shape the grid's reality through the 2030s.
What Actually Drives the Consumption
There are two distinct demand drivers, and conflating them produces a distorted picture of the problem.
Model Training
Training a single frontier model — GPT-5, Claude Opus 4.7, Gemini Ultra 2, or their equivalents — can consume tens of millions of kWh. One training run for a large model is roughly enough to power a mid-sized European city for several weeks. These runs happen repeatedly during development: initial training, fine-tuning runs, safety evaluations, capability assessments. For the largest labs, training is a continuous and significant power draw.
Inference at Scale
Inference — running the model to answer a user's query — is individually cheaper than training. But it is continuous, and scale is the multiplier that makes it the dominant long-term concern. When ChatGPT crossed 500 million weekly users, OpenAI's inference infrastructure was reportedly drawing the equivalent of a dedicated power plant's full output, around the clock. Every additional 100 million users adds another layer of demand that doesn't turn off at night. As AI becomes embedded in enterprise workflows, inference load compounds.
The Hyperscaler Construction Race
The three largest cloud providers are not waiting for the grid to catch up. They are building ahead of it and betting that power supply will follow.
Microsoft announced over $80 billion in data center investment for 2025 alone — the largest single-year capital commitment in the company's history. The facilities are spread across North America, Europe, and Asia-Pacific, with dedicated AI campuses in Virginia, Wisconsin, Arizona, and the United Kingdom. Each campus requires a dedicated substation. Several require entirely new power transmission corridors that must be negotiated with local utilities and regulators.
Google accelerated its own infrastructure build-out following strong AI revenue growth, announcing new facilities across three continents. Its campuses in Belgium, Denmark, and Singapore are among the largest power consumers in their respective regions. Amazon Web Services is building across Ohio, Oregon, Singapore, and Frankfurt, with a particular focus on capacity reserved for its Bedrock AI platform.
The capital being committed through 2025–2026 will lock in grid demand profiles through the late 2030s. These are not temporary facilities. They are permanent infrastructure, and they will draw power for as long as the AI services they support remain in production.
Sam Altman's Nuclear Bet
OpenAI's CEO Sam Altman has been the most public voice in the AI industry on the energy question — and the most direct about the scale of what he believes is coming.
His position, stated repeatedly in interviews and public forums: AI will require nuclear power. Not as one option among many, but as a structural necessity. Renewables cannot provide the firm, reliable baseload power that large-scale AI inference requires. Solar and wind are variable by nature. Battery storage at the scale needed to smooth that variability is not yet economically viable at data center scale. Nuclear delivers what AI needs: continuous, carbon-free, predictable power output.
Altman's investment in Oklo — a next-generation nuclear startup building small modular reactors — predates his tenure at OpenAI and signals a long-held conviction rather than a reactive position. But the most significant move was operational: in 2024, OpenAI signed a 20-year power purchase agreement with Constellation Energy for electricity generated by a restarted reactor at Three Mile Island in Pennsylvania.
The symbolism: Three Mile Island was the site of the most serious nuclear accident in US history, in 1979. The reactor in question was the undamaged unit that continued operating after the accident and was shut down for economic reasons in 2019 — not safety reasons. Its restart carries real symbolic weight: the AI industry is not just endorsing nuclear power in the abstract, it is funding the revival of America's most fraught nuclear site to meet its own energy demands.
Altman's logic is technically defensible. The counter-argument is not whether nuclear power works — it does — but whether it can be built fast enough to matter. New nuclear plants in the US and Europe have historically taken 10 to 15 years from groundbreaking to first power. The grid constraints being flagged by utilities right now are materializing in 3 to 5 years. Small modular reactors, which Oklo is developing, promise shorter build times — but they remain largely unproven at commercial scale. The timing mismatch is real, and no one in the industry has a clean answer to it.
The Grid Bottleneck Is Already Operational
The energy demand projections are not a future risk. In several of the world's most significant data center markets, the constraint is already affecting build plans and approval timelines.
Virginia is the most consequential example. Loudoun County, in Northern Virginia, is home to the largest concentration of data centers anywhere in the world — approximately 30% of global internet traffic passes through it. Dominion Energy, the primary utility serving the region, has issued formal warnings about grid strain through 2030. Interconnection queues — the waiting list for projects seeking to connect new generation capacity to the grid — are creating 2-to-3-year delays for new facilities. Some projects are being turned away entirely.
The constraint is not unique to Virginia. Ireland's national grid operator, EirGrid, has imposed restrictions on new data center connections in Dublin — already home to a large share of European cloud infrastructure — citing projected power shortages in the 2030s. The Netherlands placed a moratorium on new data centers in the Amsterdam region in 2022, citing grid congestion, and has maintained restrictions since. Texas, despite its abundant wind and solar generation, faces reliability concerns as summer AI inference loads coincide with air conditioning demand peaks.
| Region | Status | Main Constraint | Timeline |
|---|---|---|---|
| Virginia (USA) | Grid strain | Interconnection queue delays, substation capacity | Active through 2030 |
| Ireland | Restricted | National grid capacity — Dublin moratorium | Restrictions ongoing |
| Netherlands | Restricted | Amsterdam data center moratorium since 2022 | Restrictions ongoing |
| Texas (USA) | Monitoring | Summer peak demand overlap with AI inference load | Seasonal risk |
| UK | Managed growth | Grid connection timelines — 5+ year waits in some areas | Active |
| Singapore | Reopened | Prior moratorium lifted 2022 — managed under new standards | Monitored |
Efficiency vs. Jevons' Paradox
The most common counterargument to AI's energy problem is efficiency. Models are getting more capable per unit of compute. DeepSeek's sparse attention architecture, adopted by several 2026 frontier models including GLM-5, dramatically reduces inference costs relative to earlier dense transformer designs. Techniques like model distillation, quantization, and speculative decoding compress large models into smaller computational footprints without proportional quality degradation. The hardware side is improving too: NVIDIA's Blackwell architecture and custom ASICs from Google (TPUs), Amazon (Trainium), and Microsoft (Maia) deliver substantially better performance per watt than the previous generation.
This is all true. And it is insufficient as a solution, for a reason economists identified in 1865.
William Stanley Jevons observed that when the efficiency of coal steam engines improved, total coal consumption increased rather than decreased — because lower operating costs expanded the range of economically viable applications. The efficiency gains were real. The conservation gains were not. This dynamic, now called Jevons' paradox, has repeated across the history of energy technology: more efficient cars lead to more driving, more efficient appliances lead to more appliances, more efficient lighting leads to more lights left on.
AI follows the same pattern. More efficient inference lowers the cost per query. Lower cost per query makes AI commercially viable for more use cases. More use cases mean more queries. The total energy footprint grows. The trajectory of AI deployment strongly suggests that efficiency improvements will be absorbed by expansion, not conservation. OpenAI's own inference volumes have grown by orders of magnitude over the past three years. Each hardware generation made inference cheaper. None of them bent the total consumption curve downward.
What Green AI Actually Looks Like
Several major AI companies have made carbon-neutral or net-zero commitments, and the substance behind those commitments varies significantly. Understanding the difference matters for anyone evaluating these claims.
The most meaningful approach is power purchase agreements for additional renewable capacity — signing long-term contracts that fund the construction of new wind or solar farms, rather than purchasing credits from existing projects. Microsoft, Google, and Amazon have all made commitments of this type, though the timelines and scale of additionality vary.
The less meaningful approach is carbon credit purchasing — buying offsets that claim to represent carbon sequestration or avoided emissions elsewhere. This approach does not reduce the data center's grid demand; it purchases an accounting offset. For operational emissions at the scale of a hyperscaler, this is closer to reputation management than environmental management.
The honest assessment is that current renewable build-out is not keeping pace with AI-driven demand growth. The IEA's own analysis notes that data center demand is growing faster than new clean energy capacity in most markets. Nuclear — Altman's preferred answer — is the only currently available technology that can deliver firm, zero-carbon baseload power at data center scale. Its deployment timeline is the core problem.
What This Means If You're Building on AI
For developers, product teams, and organizations integrating AI into their infrastructure, the energy question has practical implications that are often overlooked in technical planning.
First, compute costs will not fall as fast as they have historically. The efficiency improvements are real, but they are partially offset by the rising cost of power infrastructure and the grid constraints that are increasing the capital cost of new data center construction. API pricing may not follow the steep downward curve that developers experienced between 2022 and 2025.
Second, latency and availability guarantees are increasingly tied to regional grid conditions. Data centers in constrained markets face real-time grid management requirements that can affect performance during demand peaks. This is not a theoretical risk for enterprise AI deployments — it is an emerging operational consideration.
Third, efficiency choices at the model level have real downstream impact. Choosing a smaller, well-quantized model for tasks that don't require frontier capability is not just a cost decision — it is a genuine contribution to reduced infrastructure demand. At aggregate scale, these choices matter.
TechVernia Verdict
The AI energy crisis is not a headline risk. It is an operational reality. Grid constraints are already affecting data center build timelines in the world's most important markets. The IEA's doubling projection by 2030 assumes no major policy intervention — and no major policy intervention is currently in place.
The efficiency counterargument is real but historically insufficient. Jevons' paradox has defeated it in every previous technology cycle. Nuclear power is the most credible long-term solution — and the one with the longest lead time. The gap between what the grid can deliver over the next five years and what the industry's deployment plans demand is the central constraint on how fast the current AI cycle can sustain itself.
The industry knows this. The honest ones are saying it. The question now is whether infrastructure investment — in generation, in transmission, in grid modernization — can be mobilized at a pace that matches the ambition of what's being built on top of it.
Frequently Asked Questions
A single ChatGPT query consumes approximately 0.001 to 0.01 kWh depending on query length and model version — roughly 10 times more than a Google search. At 100 million daily users sending multiple queries each, the cumulative inference load becomes substantial. The individual query seems trivial; the aggregate is a power plant running continuously.
Unit 1 at Three Mile Island — the undamaged reactor that continued operating after the 1979 accident — was shut down in 2019 for economic reasons, not safety reasons. Constellation Energy owns the facility and reached a 20-year power purchase agreement with OpenAI that makes the restart economically viable. The reactor is expected to deliver approximately 835 megawatts of carbon-free baseload power. The deal is significant both operationally and symbolically: it demonstrates that the AI industry is willing to fund nuclear infrastructure at a scale and timeline that governments and utilities have been unable to achieve on their own.
They are real in varying degrees. The most credible commitments involve power purchase agreements that fund the construction of new renewable capacity — "additionality" in carbon accounting terms. Google and Microsoft have signed agreements of this type. The less credible commitments involve purchasing existing renewable energy certificates or carbon offsets, which shift accounting without changing grid dynamics. The honest benchmark is whether a company's renewable procurement is growing at or above the pace of its data center demand growth. For most hyperscalers, the answer currently is no.
Efficiency improvements reduce the energy required per unit of AI output, which is genuinely valuable. But the historical pattern — documented across steam engines, automobiles, lighting, and semiconductor chips — is that efficiency gains expand the economically viable application space, leading to increased total consumption. AI is following this pattern: GPT-4 was dramatically more efficient than GPT-3 per capability unit, but total AI inference energy consumption grew sharply in the years following GPT-4's release because the lower cost enabled vastly broader deployment. Efficiency is necessary but not sufficient as a solution to aggregate demand growth.
Northern Virginia (USA), Dublin (Ireland), and the Amsterdam metropolitan area (Netherlands) face the most active constraints. Virginia has the highest density of data center infrastructure globally and formal grid strain warnings through 2030. Ireland imposed connection restrictions in Dublin following projections of data center loads exceeding 30% of national electricity demand by 2030. The Netherlands has maintained a de facto moratorium on new hyperscale data centers in the Amsterdam region since 2022. Singapore, which also imposed a moratorium, has reopened under stricter efficiency standards.
Jevons' paradox is the economic observation that efficiency improvements in resource consumption tend to increase total resource use, not decrease it, because lower cost per unit enables broader adoption. Named after 19th-century economist William Stanley Jevons, who observed it in coal consumption. For AI, the implication is that making inference cheaper per query — through better hardware, more efficient architectures, or model compression — expands the range of commercially viable AI applications, leading to more total queries and more total energy consumption, even as energy per query falls. It is the central reason why efficiency alone cannot be expected to bend the AI energy curve downward.
Conclusion
The grid is not a software problem. It does not respond to clever architecture choices or optimization passes. It is physical infrastructure — transmission lines, substations, generation capacity — and it requires capital investment, regulatory approval, and construction timelines measured in years, not sprint cycles.
The AI industry is building at a pace that assumes the energy supply will follow. In some markets, that assumption is already proving incorrect. The bottleneck is not theoretical or distant. It is showing up in interconnection queues, in utility capacity warnings, in regulatory moratoria on new data center connections in some of the world's most important digital infrastructure markets.
The companies that take this constraint seriously — that plan infrastructure around realistic power availability, that invest in efficiency not just for cost but for sustainability, that engage with nuclear and grid modernization as genuine strategic priorities — will be better positioned in a decade than those treating energy as a footnote to their compute roadmaps.
The AI energy question is not a footnote to the AI story. It may be its defining constraint.
Related Articles: