Another question is - will it ever become less costly to train?
Let to see opinions from someone in the know
Another question is - will it ever become less costly to train?
Let to see opinions from someone in the know
So to keep up with times the models have to be constantly trained.
One thing though is that right now it's not just incremental training, the whole thing gets updated - multiple parameters and how the model is trained is different.
This might not be the case in the future where the training could become more efficient and switch to incremental updates where you don't have to re-feed all the training data but only the new things.
I am simplifying here for brevity, but I think the gist is still there.
They're training new models because the (software) technology keeps improving, (proprietary) data sets keep improving (through a lot of manual labelling but also synthetic data generation), and in general researchers have better understanding of what's important when it comes to LLMs.