←back to thread

3 points yalogin | 1 comments | | HN request time: 0.315s | source

OpenAI, Anthropomorphic and every company is putting a lot into training. How is the functionality pipeline filled out for LLMs? What is missing in today’s LLMs that they need to plan for? Just trying to get some insights into the planning process and also what the community sees as the northstar for LLMs without saying AGI
Show context
jqpabc123 ◴[] No.45059878[source]
For the most part, LLMs just remember.

They don't think or learn or create on their own --- at least not anywhere close to a human level. Otherwise, they wouldn't require so much "training"

Essentially, they are best characterized as a huge database with a natural language interface.

Once the internet had been consumed and indexed, this sort of approach starts to hit a wall. There is no more data readily available for "training".

I don't know what the next breakthrough will be but I firmly believe one will be required to push performance to any significantly higher level.

replies(1): >>45061396 #
1. pillefitz ◴[] No.45061396[source]
In terms of bits seen during training, LLM are more akin to a 3 year old. Robots roaming around and learning to interact with the in environment and sharing knowledge might be a game changer, assuming that the current methodologies are sufficient (LLM + RL)