AI Is Thirsty. Each chat with a large language-model… | by Clive … – Medium

Member-only story
Clive Thompson
Follow

40
Share
Today I used ChatGPT to get some help making a browser plugin. I posted my queries, then watched as the code and text spilled down the screen. This is the part of large language-models that I dig! As a hobbyist developer, getting suggestions of customized lines of software can be a powerful way to learn.
But as it turns out, using ChatGPT consumes a lot of an unexpected resource:
Water.
The code wasn’t quite what I was looking for, so I chatted with ChatGPT for 15 minutes or so, slowly coaxing it to revise. By the time I was done, we’d gone back and forth about 20 times.
And during that exchange? Microsoft’s servers probably used about as much water as if I’d just bought a half-liter bottle … and spilled it on the ground.
AI, it turns out, is incredibly thirsty tech — ploughing through torrents of fresh water every day. Given that we’re likely to see large-language-model AI woven into ever more apps and appliances these days, it’s worth pondering just how much water our booming use of AI will consume.
Why precisely does large-language-model AI require water? Back in April, a group of researchers pondered this question as they created an estimate of AI’s water consumption. As they note in their paper (which is here free in full), the main use of water is when tech firms train their AI, and when the firms are running inferences (i.e. when you, I or anyone else interacts with the model).
Tech firms like Microsoft and Google and Meta do all that training (and inferring) on their huge computational farms. That computation requires a ton of energy, which generates heat. To remove that heat from server farms, the tech firms generally use cooling towers, where water is evaporated to send the heat out into the outside world. That evaporation? That’s how AI consumes water. It is, it’s worth noting, mostly all freshwater.
Tech firms do not publish specific stats on how much freshwater they use for different forms of computation. So the academics did some estimates. They calculated how much energy it would take to train one of the well-known large language-models (and…


40
I write 2X a week on tech, science, culture — and how those collide. Writer at NYT mag/Wired; author, “Coders”. @clive@saturation.social clive@clivethompson.net
Clive Thompson

53
Clive Thompson

36
Clive Thompson

13
Clive Thompson

21
Pau Blasco i Roca
in
Towards Data Science

54
Thomas Smith
in
The Generator

58
David Goudet

49
Jeremy Arancio
in
Towards AI

34
Tomas Pueyo

49
Devansh
in
DataDrivenInvestor

17
Help
Status
About
Careers
Blog
Privacy
Terms
Text to speech
Teams

source

Jesse
https://playwithchatgtp.com