WaPo is very wrong on ChatGPT energy use

January 6, 2025

(disclaimer: I've worked at OpenAI since 2020)

The Washington Post claims generating a 100 word email with ChatGPT takes 0.14 kWh of electricity, resulting in 519 mL of water use (article link, archive link).

This number is off by many orders of magnitude, maybe around 1000x.1

I'm writing this as a thing I can link to family and friends who ask me about this.

Cost of electricity

One easy way the Washington Post could have known how mistaken they were: simply compare the cost OpenAI charges for generating text to the cost of electricity. You can use OpenAI's GPT-4o model to generate 1 million output tokens at a cost of $5 or $10 per OpenAI's pricing page.

Generating a 100 word email (or about 120 tokens) therefore costs you between 0.06 and 0.12 cents.

If the Washington Post's 0.14kWh number were true, and assuming a data centre electricity cost of $0.06 per kWh, it would take 0.84 cents of electricity to generate a 100 word email.

In other words, for every dollar OpenAI charges you to generate text, OpenAI would be losing $7 to $14 just on electricity, before even taking into account the cost of computer chips, personnel, operations, partnerships, etc. This is obviously not the case.

How did they get it so wrong?

The article cites Making AI Less “Thirsty”: Uncovering and Addressing the Secret Water Footprint of AI Models as the source of its methodology.

That paper quotes two sources for energy cost of inference (aka using a model):

A page of text is conventionally 500 words (600 tokens), so plugging in the numberssee 3, Washington Post's number is already off by a factor of 175x from the paper it cites.

There is no explanation I could find for the discrepancy between the Washington Post and the paper the Washington Post cites — the quoted number is explicitly about inference, not training. My guess is that they're attempting to account for the difference between GPT-3 and GPT-4o, so they pull a multiple from a hat, with no discernible connection with any rumoured or real numbers.

We'll spend the rest of this post talking about why it's actually much cheaper to use GPT-4o in 2024 than it was to use GPT-3 in 2020, and so even the GPT-3 paper's estimate of 0.0008 kWh per 100 word email is much too high — let alone the number fabricated by the Washington Post that's 175x higher than that.

While I think the WaPo number is so far off as to be inexcusable, it obviously doesn't help that OpenAI hasn't published official energy use numbers (in part, because energy use can reveal quite a lot of proprietary information and because the numbers are changing constantly).

Edit June 2025: Sam Altman included an estimate of ChatGPT energy use per query as an aside in one of his blogposts. That number also lines up with a 1000x mistake from WaPo, once you correct for the average ChatGPT query being more than 100 words.

One more quick methodology note: all sources involved are primarily looking at energy usage of just the GPUs. Somewhat surprisingly, in practice at inference time, the energy draw from other parts of the system can actually add up meaningfully. But I'll stick apples to apples with WaPo and its sources and its sources' sources.

Using models is much cheaper than it was in 2020

When the GPT-3 paper came out, approximately no one cared about generative language models. Now that people do, there is a lot more incentive to make them efficient. There have been a number of algorithmic, hardware and engineering improvements in the last five years that make inference and training much cheaper. I'll give a few random fully publicly known examples, but feel free to skip past this if it's too much detail.

...or just look at what OpenAI charges

GPT-4 was introduced in the OpenAI API in 2023 at the same cost as GPT-3 was originally. Since then, the cost OpenAI charges to use GPT-4 quality models in the API has fallen by a factor of 6x, which suggests a lower bound on efficiency improvements.

From the Washington Post's number, you go down 175x to the GPT-3 paper's number. From there, you go down 6x based on OpenAI's pricing changes. This 6 x 175 alone gets you to the 1000x I claim in the intro.

Inference energy use contextualised

I'd estimate that using ChatGPT to generate a 100 word email uses energy roughly similar to:

One thing to take home: for a given level of capability, today is the most expensive AI will ever be. Writing your emails will continue to get cheaper, likely rapidly.

What about water usage?

"Making AI less Thirsty"'s water usage numbers are derived from the electricity use numbers. They essentially multiply the electricity usage by a water usage constant. 5 This means that the water usage estimates are off by the same factor as the electricity usage estimates.

Indeed, the paper's main contribution appears to be estimating those water usage constants. I have not independently verified those numbers, but assuming everything else is correct, the water usage per 100 word email is literally just a few drops.

One other thing I noticed: "Making AI less Thirsty"'s training numbers for GPT-3 come from Carbon Emissions and Large Neural Network Training. That paper already accounts for data centre PUE — the double counting means water usage numbers for training are inflated by 17% (but this is relatively a small potatoes error when compared to errors of 100000%).

What about training?

So far we've focussed just on model usage, because the headline number in the Washington Post article is explicitly just about the ongoing costs of using a model.

However, the one-time training process for language models is energy intensive. Whether training a model is worth the energy to do so depends on how much the model gets used.

The 300 million weekly ChatGPT users send over a billion messages a day — in addition to all usage of a model via the OpenAI API — so even 10 GWh scale energy use in training would amortise reasonably.

Good public numbers here are hard to come by, so you'll need to plug in your best guesses, but you'll likely find the amortised energy cost to be a low multiple of marginal inference cost.

What about the future?

While for a given level of capability, AI is currently the most expensive it will ever be, what remains to be seen is a) how much capability will increase, b) what demand will be for these higher levels of capabilities.

For instance, OpenAI's o-series models require more energy at inference time compared to the traditional GPT models. Terence Tao reckons o1 is "on par with a mediocre but not completely incompetent graduate student". The recently announced o3 is in the 99.8th percentile on Codeforces, a competitive programming website. What uses — beyond email writing — will these models and future even more powerful models unlock?

One MRI scan takes about 20 kWh, or five orders of magnitude more than a ChatGPT email. How much energy is it worth spending to broaden access to better medical care?

CERN currently uses about 1.3 TWh of electricity annually to power particle accelerators and detectors, much more than generative AI usage today. How much energy is it worth spending to push the frontiers of science?


1. I'll make the case in this post for 1000x using fully public information. The exact power draw and token usage numbers I see internally are proprietary information and can be used to infer proprietary details of OpenAI models.
2. The BLOOM paper measures inference cost in a low usage deployment. The BLOOM paper is really good about contextualising this, talking about how the energy use of their deployment is just a little bit more than the baseline idle energy use (80% of the energy use in their 18 day period can be attributed to baseline idle use). The BLOOM paper actually doesn't even provide the energy per request number since it would be misleading.
3. The Washington Post claims 0.14 kWh per 100 words (120 tokens). The GPT-3 paper cited claims 0.004 kWh per page (600 tokens). `(0.14 / 120) / (0.004 / 600) = 175`
4. What is definitely an over-claim is this Nvidia 2024 presentation which claims 120x joules per token improvement for GPT models just from using Hopper chips instead of Volta — not sure where they got that number.
5. In the paper, this is Electricity Water Intensity x Power Usage Effectiveness + Water Usage Effectiveness = 3.142 * 1.17 + 0.55 on average in the U.S.