←back to thread

56 points trott | 2 comments | | HN request time: 0s | source
Show context
ofrzeta ◴[] No.40714106[source]
"If trends continue, language models will fully utilize this stock between 2026 and 2032" - that will require data centers with their own nuclear reactors (or other power plants) as hinted at by Marc Zuckerberg?
replies(2): >>40714248 #>>40714295 #
trott ◴[] No.40714248[source]
If you take Llama-3-400B, and 30x its data (hitting the data ceiling, AFAICT), 30x its size to match, and the hardware improves by, say, 3x, then you'll use up about a year's worth of energy from a typical nuclear power plant.
replies(3): >>40714269 #>>40714294 #>>40714809 #
1. mathsmath ◴[] No.40714269[source]
I don’t know much about LLMs, but is it possible to throttle their training?

Solar has gotten pretty cheap, and I’m just wondering if you can throttle up and down based on how much output the panels are producing.

replies(1): >>40714416 #
2. moi2388 ◴[] No.40714416[source]
Of course it is, but the trade-off is time.