FindArticles FindArticles
  • News
  • Technology
  • Business
  • Entertainment
  • Science & Health
  • Knowledge Base
FindArticlesFindArticles
Font ResizerAa
Search
  • News
  • Technology
  • Business
  • Entertainment
  • Science & Health
  • Knowledge Base
Follow US
  • Contact Us
  • About Us
  • Write For Us
  • Privacy Policy
  • Terms of Service
FindArticles © 2025. All Rights Reserved.
FindArticles > News > Technology

Altman and Nadella warn of uncertain AI power demand.

Gregory Zuckerman
Last updated: November 4, 2025 12:42 am
By Gregory Zuckerman
Technology
6 Min Read
SHARE

Notably, in recent comments, two of the most prominent figures in AI claim that the limiting factor of the next wave is no longer chips, but they cannot quite foresee how much electricity they might consume.

In the past couple of weeks, OpenAI’s Sam Altman and Microsoft’s Satya Nadella have both reported a growing distance between accelerator provision and the capacity to power and house these chips. That disconnect is a significant shift for an industry moving away from the compute race to an energy race. Hyperscalers purchased cutting-edge GPUs over the last two years, only to receive power distributors, substations, and data center “warm shells,” i.e., unpowered buildings. Nadella stated that this is not a silicon barrier but a practical one about consistently powering these installations fast enough — an acknowledgement that some enhancements might idle while developers wait for the supercomputer shell to be linked to an electric grid. A clear example of why this is the case is that a single high-end AI environment with tens of thousands of accelerators might consume 50-100 megawatts, while a multi-building site might require in excess of 300 MW: even PUEs of close to 1.1-1.2 would push the complete load 10-20 percent above IT, assuming optimal practices. High-voltage transformers have lead times of two or more years, while utility studies take 2-3 years.

Two men, one with light skin and brown hair, and the other with dark skin and a bald head, stand side-by-side against a textured beige wall. The man on the left wears a blue long-sleeved shirt, and the man on the right wears a dark gray polo shirt and glasses.

Forecasting AI power needs has become guesswork. Model training cycles are exploding, inference volumes are compounding, and efficiency improvements cut both ways. Altman has argued that when the cost per “unit of intelligence” falls, usage grows faster than the savings — a classic Jevons paradox dynamic. If that holds, any near-term efficiency gain could be swamped by new applications, from multimodal assistants to always-on copilots and real-time agents. Public projections underscore the uncertainty. While the International Energy Agency estimated global data centers used roughly 460 terawatt-hours in 2022 and could reach 620–1,050 TWh by 2026, with AI a major driver. In the U.S., interconnection queues tracked by national labs show thousands of gigawatts of generation and storage waiting to connect. This is evidence that supply is trying to respond, but slowly.

To close the gap, developers are increasingly going “behind the meter” with onsite generation and dedicated lines. This model offers speed and control but introduces fuel price risk and locks companies into long-lived assets just as energy technologies are evolving quickly. That’s what worries Altman: if a dramatically cheaper power source arrives at scale, companies holding legacy power purchase agreements or onsite gas assets could be stuck above market for years. Conversely, under-procurement risks rationing compute precisely when product demand spikes.

Two men, one with light skin and brown hair, and the other with dark skin and a bald head, stand side-by-side against a textured beige wall.

Altman’s portfolio reflects this view. He has bet across the spectrum, backing fission startup Oklo, fusion developer Helion, and solar-thermal newcomer Exowatt. Microsoft, meanwhile, signed an offtake agreement with Helion, targeting future fusion power and accelerated conventional deals for wind, solar, and batteries. The hard truth today is that nuclear projects take years to permit; fusion is unproven at grid scale; and new gas turbines and switchgear are market-constrained. This makes renewables the fastest-moving and data centers the nearest. Solar and wind can go up in months, not years, and the modularity is familiar to how data centers scale. The big problem is that they are variable and require hourly matching and storage to run training setups or do low-latency inference carefully. Tech firms are shifting from annual megawatt-hour offsets to 24/7 carbon-free procurement, pioneered by Google and now catching on among its peers. Power is site-specific; therefore, regions with spare capacity and friendly interconnection policies — Northern Virginia, parts of the Nordics and Quebec — are hot. But communities are pushing back on land use, water consumption, and grid stress. Regulators are reacting with process reform: FERC’s proposed interconnection rule changes aim to shorten queue backlogs, and change is coming to the power markets, but it remains multi-year. In the meantime, cooling advances like direct-liquid and immersion cooling can slash energy overhead and footprint, with space and labor. But the physics are unforgiving: training frontier models still concentrate tens of MW in dense spaces, all of it must be pumped, chilled, and monitored precisely.

The playbook for an unknowable AI power load

With no consensus forecast, the emerging strategy blends hedging and flexibility. Expect staggered campus buildouts to avoid stranded capacity, diversified energy portfolios mixing PPAs, storage, and onsite generation, and contracts with optionality if cheaper power materializes. On the R&D side, expect relentless work on algorithmic efficiency — sparsity, quantization, and better compilers — to stretch each watt, even if total demand keeps rising. For now, the message from Altman and Nadella is plain: the AI roadmap runs through electrons. The winners will be the ones who can turn power into intelligence at scale — and do it faster than the grid can say yes.

Gregory Zuckerman
ByGregory Zuckerman
Gregory Zuckerman is a veteran investigative journalist and financial writer with decades of experience covering global markets, investment strategies, and the business personalities shaping them. His writing blends deep reporting with narrative storytelling to uncover the hidden forces behind financial trends and innovations. Over the years, Gregory’s work has earned industry recognition for bringing clarity to complex financial topics, and he continues to focus on long-form journalism that explores hedge funds, private equity, and high-stakes investing.
Latest News
New Firmware Restores Old Nest Thermostats Online
Helium Mobile Offers Free Pie With New Plans
Ex Trenchant Boss Admits Selling Zero Days To Russia
Apple unveils redesigned web App Store with full navigation
Like-New Apple Pencil (1st Gen) Drops to $45
T-Mobile removes AppHub after silent installs
Google Pixel 10 discounts reach up to $300 off today
Lambda seals multibillion AI deal with Microsoft
Soulja Boy Sells Retroid Pocket Flip 2 Clone
Disney Content Pulled From More Google Platforms
AI Browsers Bypass Publisher Paywalls, Report Finds
Apple signals record holiday quarter with double-digit growth
FindArticles
  • Contact Us
  • About Us
  • Write For Us
  • Privacy Policy
  • Terms of Service
  • Corrections Policy
  • Diversity & Inclusion Statement
  • Diversity in Our Team
  • Editorial Guidelines
  • Feedback & Editorial Contact Policy
FindArticles © 2025. All Rights Reserved.