AI Is Thirsty. Each chat with a large language-model… | by Clive … – Medium
Member-only story
Clive Thompson
Follow
—
40
Share
Today I used ChatGPT to get some help making a browser plugin. I posted my queries, then watched as the code and text spilled down the screen. This is the part of large language-models that I dig! As a hobbyist developer, getting suggestions of customized lines of software can be a powerful way to learn.
But as it turns out, using ChatGPT consumes a lot of an unexpected resource:
Water.
The code wasn’t quite what I was looking for, so I chatted with ChatGPT for 15 minutes or so, slowly coaxing it to revise. By the time I was done, we’d gone back and forth about 20 times.
And during that exchange? Microsoft’s servers probably used about as much water as if I’d just bought a half-liter bottle … and spilled it on the ground.
AI, it turns out, is incredibly thirsty tech — ploughing through torrents of fresh water every day. Given that we’re likely to see large-language-model AI woven into ever more apps and appliances these days, it’s worth pondering just how much water our booming use of AI will consume.
Why precisely does large-language-model AI require water? Back in April, a group of researchers pondered this question as they created an estimate of AI’s water consumption. As they note in their paper (which is here free in full), the main use of water is when tech firms train their AI, and when the firms are running inferences (i.e. when you, I or anyone else interacts with the model).
Tech firms like Microsoft and Google and Meta do all that training (and inferring) on their huge computational farms. That computation requires a ton of energy, which generates heat. To remove that heat from server farms, the tech firms generally use cooling towers, where water is evaporated to send the heat out into the outside world. That evaporation? That’s how AI consumes water. It is, it’s worth noting, mostly all freshwater.
Tech firms do not publish specific stats on how much freshwater they use for different forms of computation. So the academics did some estimates. They calculated how much energy it would take to train one of the well-known large language-models (and…
—
—
40
I write 2X a week on tech, science, culture — and how those collide. Writer at NYT mag/Wired; author, “Coders”. @clive@saturation.social clive@clivethompson.net
Clive Thompson
—
53
Clive Thompson
—
36
Clive Thompson
—
13
Clive Thompson
—
21
Pau Blasco i Roca
in
Towards Data Science
—
54
Thomas Smith
in
The Generator
—
58
David Goudet
—
49
Jeremy Arancio
in
Towards AI
—
34
Tomas Pueyo
—
49
Devansh
in
DataDrivenInvestor
—
17
Help
Status
About
Careers
Blog
Privacy
Terms
Text to speech
Teams