"Watts, Water, and Carbon: What Your AI Prompts Actually Cost"

8 min read Original article ↗

Every time you ask an AI to summarize a document, generate an image, or write code, a GPU somewhere lights up — and a power meter ticks upward. As someone who runs AI agents all day, I started wondering: what’s the actual energy footprint of this revolution I’m participating in?

The answer is more nuanced than the headlines suggest.

What a single prompt actually costs

Let’s start small. You’ve probably seen the stat that a ChatGPT query uses “10x more energy than a Google search.” The reality is more complicated . Google’s traditional search uses about 0.3 watt-hours. The Electric Power Research Institute estimated a ChatGPT request at 2.9 Wh — roughly 10x.

But in August 2025, Google published actual data for the first time: a median Gemini query consumes just 0.24 Wh, and efficiency improved 33x between May 2024 and May 2025. More recent analysis from Epoch AI puts a typical GPT-4o query around 0.3 Wh — nearly comparable to a Google search. These estimates are directional, not perfectly apples-to-apples: model architecture, hardware, traffic patterns, and accounting boundaries all differ.

Not all AI workloads are equal, though. Research from Hugging Face found that image generation consumes 5 to 50x more energy than a text query of comparable length. A single Stable Diffusion image clocks in around 0.7 Wh. Video generation is even hungrier. As multimodal AI explodes, the per-query averages will shift upward even as text gets cheaper.

The per-query cost for text is shrinking fast. But that doesn’t mean the total cost is shrinking. This is where it gets interesting.

Training vs. inference: two different beasts

To understand why energy demand is growing so fast, you need to understand the two fundamentally different ways AI consumes power.

Training a large model — the months-long process of feeding it data — is a sustained, predictable load. It runs 24/7 and benefits from steady baseload power like nuclear. Training GPT-4 consumed an estimated ~50 GWh total .

Inference — actually using the model when you send it a prompt — is bursty. Millions of requests create intense, unpredictable spikes. As Tom’s Hardware noted , these spikes stress grid infrastructure and require flexible power sources to absorb.

This distinction matters because inference is growing much faster than training. Every new AI product, every chatbot, every agent running in the background is an inference workload. Training happens once; inference happens millions of times per day.

The global demand surge

Now let’s zoom out. According to Deloitte , data centers consumed about 536 TWh globally in 2025 — roughly 2% of the world’s electricity. That sounds manageable until you zoom into the U.S.

Pew Research found that data centers already account for 4% of total U.S. electricity use in 2024, and that number is expected to more than double by 2030. The IEA projects global data center electricity demand will exceed 1,000 TWh by 2030, with AI-optimized facilities quadrupling their share. The largest individual data centers now under construction will each consume as much electricity as 2 million households.

A third of U.S. data centers are concentrated in just three states: Virginia, Texas, and California. In Virginia alone, data centers consumed 26% of the state’s total electricity in 2023.

Why efficiency won’t save us

Here’s the counterintuitive part. Per-query costs are plummeting — so shouldn’t total energy use be falling too?

In the 1860s, economist William Stanley Jevons observed that as coal engines became more efficient, total coal consumption increased — because efficiency made coal cheaper, which drove more usage. NPR covered how the AI world is now obsessed with this paradox after DeepSeek showed that cheaper models just mean more people using AI.

A 2025 paper published at ACM FAccT makes this explicit: efficiency gains in AI hardware lower costs, which spurs demand for new AI functionalities, which drives further hardware upgrades, which increases total energy consumption. As SIGARCH noted , efficiency alone won’t solve the data center carbon challenge without policy and behavioral changes alongside it.

Each query is getting cheaper. We’re just making exponentially more of them.

Energy gets the headlines, but AI data centers are also incredibly thirsty.

Google published that an average Gemini query consumes 0.26 mL of water. OpenAI’s Sam Altman said a ChatGPT query uses “roughly one fifteenth of a teaspoon.” But these numbers are not directly comparable to older estimates. Earlier research from UC Riverside estimated that, depending on location and cooling assumptions, generating 10 to 50 GPT-3 responses could indirectly consume about one 500 mL bottle of water. The range is wide because water use depends heavily on where and when the model is served.

At scale, Brookings reports that annual U.S. data center water consumption could double or quadruple by 2028 compared to 2023 levels — reaching 150-280 billion liters per year. Microsoft’s water use already jumped 34% year-over-year in 2023, driven by AI infrastructure. These facilities are increasingly being built in water-stressed regions of the American West , competing with agriculture and residential use.

Big Tech’s broken climate promises

According to UN data reported by Al Jazeera , tech giants saw emissions surge 150% in three years amid the AI boom. The breakdown: Amazon’s operational emissions grew 182% since 2020, Microsoft 155%, Meta 145%, and Google 138%.

These are the same companies that pledged to be carbon neutral or carbon negative by 2030. Microsoft’s own sustainability report acknowledged the 2030 carbon-negative goal is now significantly harder to meet . Google, which had been carbon neutral since 2007, admitted its 24/7 carbon-free energy target is increasingly difficult .

The response? More carbon credit purchases — up 181% to 68.4 million carbon credits in 2025 — while simultaneously spending a combined $320 billion on AI infrastructure that same year. As Accenture warned , AI’s carbon emissions could surge 11-fold if the industry doesn’t change course.

The nuclear gold rush

To their credit, Big Tech is betting big on nuclear:

But as Goldman Sachs Research notes, nuclear can’t meet all the demand alone. And the timelines don’t match: these SMRs are years away, while demand is growing now. As CSIS frames it , the central challenge is speed-to-power — how fast new sites can access electricity. That urgency favors whatever fuel is available today, which often means gas or even coal.

Coal generation in the U.S. rose nearly 20% as elevated gas prices pushed utilities back toward dirtier fuels, according to the Jefferies analysis cited by Tom’s Hardware . Even if that proves temporary, it shows how quickly short-term power constraints can push the grid toward dirtier fallback options.

What this means for us

I think about all of this when I spin up multi-agent workflows that orchestrate half a dozen Claude instances in parallel. Each prompt has a real, if tiny, energy cost. Multiply that across millions of users, and those tiny costs compound into the electricity consumption of small countries.

Does that mean we should stop using AI? No. But we should be honest about the tradeoffs:

  • Transparency matters. Google publishing Gemini’s energy data was a good start. Every AI provider should follow suit.
  • Efficiency is necessary but not sufficient. Per-query costs are dropping, but Jevons paradox means total consumption keeps rising. We need systemic solutions, not just better chips.
  • The nuclear timeline is too slow. We need bridge solutions that aren’t just “burn more gas and buy carbon credits.”
  • As developers, our choices have energy consequences. Running an agent loop that makes 50 LLM calls when 5 would suffice isn’t just wasteful of tokens — it’s wasteful of watts and water.

MIT researchers describe powering AI as a “multifaceted challenge” spanning technology, energy, government policy, and consumer impact. The IEA paints a picture of an industry building demand far faster than it’s building supply. The optimistic view is that AI will eventually help optimize the very energy systems it’s straining — better grid management, materials science breakthroughs, more efficient chips. The realistic view is that for the next decade, the AI boom will lean on fossil fuels more than anyone wants to admit, and create real tensions between tech growth and climate commitments.

The least we can do is stop pretending the cloud runs on magic. Every inference has a cost — in watts, in water, in carbon. The question is whether we’ll be intentional about paying it.