Humans receive an enormous amount of training data in forms not currently available to LLMs.
If you locked baby Einstein in a room with the collected works of humanity and left him there for a lifetime, I doubt he'd have even learnt to read on his own.
Secondarily, I would expect this effect to be swamped by other factors (e.g. conquest).
All of the information accumulated by evolution gets passed through DNA. For humans, that's well under 1GB. Probably a very tiny fraction of that determines how the brain works at the algorithmic level. You should think of this information as the "software" of the brain, not pretrained LLM weights (350GB for GPT-3).
How much of that is cognitively useful for learning English? On top of the textual content, audio gives you emphasis and mood. Not a lot of information in that -- a few bits per sentence.
Better even measure in bytes. And remember that kids look at video, not at individual pictures (even if these are videos of pictures).