I've seen a deer on a road maybe once. I've seen a rabbit on a road zero times. But I know what to do if I see one.
Is that because the "video" of my perception has many "frames"? Even if that's true at some level, I think it's massively missing the point. Yeah, so I saw that one deer from a lot of angles. But current AI training is like the equivalent of taking every deer that has ever been on camera in the history of the human species.
Somehow I'm still dramatically better at generalization than the AI. Surely that's an algorithm difference.
But we have seen from AlphaGo that when training data is extensive, it can rediscover strategy on its own and even surpass us. It's not inherently worse than human learning.
Which pre-human animals evolved instincts for swerving a car to avoid a deer?
There are definitely teams working on applying reinforcement learning to LLMs. Maybe that will unlock new potential from finite training data.
"""
Intuitively, an overparameterized model will generalize well if the model’s representations capture the essential information necessary for the best model in the model class to perform well
"""
https://iclr-blogposts.github.io/2024/blog/double-descent-de...