←back to thread

346 points swatson741 | 1 comments | | HN request time: 0.207s | source
Show context
brcmthrowaway ◴[] No.45788805[source]
Do LLMs still use backprop?
replies(3): >>45788887 #>>45790627 #>>45791775 #
1. samsartor ◴[] No.45791775[source]
Yes. Pretraining and fine-tuning use standard Adam optimizers (usually with weight-decay). Reinforcement learning has been the odd-man out historically, but these days almost all RL algorithms also use backprop and gradient descent.