this post was submitted on 30 Jul 2024
172 points (94.3% liked)

Fuck AI

1148 readers
504 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

founded 6 months ago
MODERATORS
 

One million Blackwell GPUs would suck down an astonishing 1.875 gigawatts of power. For context, a typical nuclear power plant only produces 1 gigawatt of power.

Fossil fuel-burning plants, whether that's natural gas, coal, or oil, produce even less. There's no way to ramp up nuclear capacity in the time it will take to supply these millions of chips, so much, if not all, of that extra power demand is going to come from carbon-emitting sources.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 39 points 1 month ago (3 children)

I've had multiple people on Lemmy tell me that the amount of energy LLMs use will be trivial. They always base it on the amount of energy used to train the LLMs, not the millions (billions? trillions?) of calculations those LLMs have to do every second they're used by who knows how many people 24 hours a day.

Then you bring up the water wasting and the best they can do is say something like, "okay, that's a problem... but only in some places!"

(Some places including much of the United States. Guess where lots of the data centers are?)

[–] [email protected] 0 points 1 month ago* (last edited 1 month ago) (1 children)

I don't disagree, but it is useful to point out there are two truths in what you wrote.

The energy use of one person running an already trained model on their own hardware is trivial.

Even the energy use of many many people using already trained models (ChatGPT, etc) is still not the problem at hand (probably on the order of the energy usage from a typical search engine).

The energy use in training these models (the appendage measuring contest between tech giants pretending they're on the cusp of AGI) is where the cost really ramps up.

[–] [email protected] 1 points 1 month ago (1 children)

(probably on the order of the energy usage from a typical search engine).

I find that hard to believe. Search engines just regurgitate what is in a database. LLMs have to do calculations to create the sentences they produce. That takes more energy.

[–] [email protected] 2 points 1 month ago* (last edited 1 month ago)

Believe what you will. I'm not an authority on the topic, but as a researcher in an adjacent field I have a pretty good idea. I also self host Ollama and SearXNG (a metasearch engine, to be clear, not a first party search engine) so I have some anecdotal inclinations.

Training even a teeny tiny LLM or ML model can run a typical gaming desktop at 100% for days. Sending a query to a pretrained model hardly even shows up on HTop unless it's gigantic. Even the gigantic models only spike the CPU for a few seconds (until the query is complete). SearXNG, again anecdotally, spikes my PC about the same as Mistral in Ollama.

I would encourage you to look at more explanations like the one below. I'm not just blowing smoke, and I'm not dismissing the very real problem of massive training costs (in money, energy, and water) that you're pointing out.

https://www.baeldung.com/cs/chatgpt-large-language-models-power-consumption

load more comments (1 replies)