←back to thread

An LLM is a lossy encyclopedia

(simonwillison.net)
509 points tosh | 2 comments | | HN request time: 0.006s | source

(the referenced HN thread starts at https://news.ycombinator.com/item?id=45060519)
Show context
kgeist ◴[] No.45101806[source]
I think an LLM can be used as a kind of lossy encyclopedia, but equating it directly to one isn't entirely accurate. The human mind is also, in a sense, a lossy encyclopedia.

I prefer to think of LLMs as lossy predictors. If you think about it, natural "intelligence" itself can be understood as another type of predictor: you build a world model to anticipate what will happen next so you can plan your actions accordingly and survive.

In the real world, with countless fuzzy factors, no predictor can ever be perfectly lossless. The only real difference, for me, is that LLMs are lossier predictors than human minds (for now). That's all there is to it.

Whatever analogy you use, it comes down to the realization that there's always some lossiness involved, whether you frame it as an encyclopedia or not.

replies(6): >>45102030 #>>45102068 #>>45102070 #>>45102175 #>>45102917 #>>45103645 #
A_D_E_P_T ◴[] No.45102068[source]
> If you think about it, natural "intelligence" itself can be understood as another type of predictor: you build a world model to anticipate what will happen next so you can plan your actions accordingly and survive.

Yes.

Human intelligence consists of three things.

First, groundedness: The ability to form a representation of the world and one’s place in it.

Second, a temporal-spatial sense: A subjective and bounded idea of self in objective space and time.

Third: A general predictive function which is capable of broad abstraction.

At its most basic level, this third element enables man to acquire, process, store, represent, and continually re-acquire knowledge which is external to that man's subjective existence. This is calculation in the strictest sense.

And it is the third element -- the strength, speed, and breadth of the predictive function -- which is synonymous with the word "intelligence." Higher animals have all three elements, but they're pretty hazy -- especially the third. And, in humans, short time horizons are synonymous with intellectual dullness.

All of this is to say that if you have a "prediction machine" you're 90% of the way to a true "intelligence machine." It also, I think, suggests routes that might lead to more robust AI in the future. (Ground the AI, give it a limited physical presence in time and space, match its clocks to the outside world.)

replies(1): >>45102212 #
quonn ◴[] No.45102212[source]
"Prediction" is hardly more than another term for inference. It's the very essence of machine learning. There is nothing new or useful in this concept.
replies(1): >>45102421 #
A_D_E_P_T ◴[] No.45102421[source]
Point is that it's also exactly analogous to human intelligence. There's almost nothing else to it.
replies(1): >>45102802 #
devmor ◴[] No.45102802[source]
This is how you spot hype nonsense - claims that anything is analogous to human intelligence. Even absent all other objections, we don't understand the human mind well enough to make a claim like that.
replies(1): >>45103278 #
A_D_E_P_T ◴[] No.45103278[source]
You don't need to understand the human mind on a mechanistic level. You only need to examine how the whole organism learns, acts, and reacts to stimulus and situation.

Even something as simple as catching a ball is basically predictive. You predict where the ball will be along its arc when it reaches a point in space where you can catch it. Then, strictly informed by that prediction, you solve a problem of motion through space -- and some very simple-seeming problems of motion through space can't be cracked in a general case without a very powerful supercomputer -- to physically catch the ball.

That's a very simple example. The major component of what we call intelligence is purely predictive. Of course Bayesian inference also works the same way, etc.

replies(2): >>45103983 #>>45104033 #
devmor ◴[] No.45104033[source]
> The major component of what we call intelligence is purely predictive.

Making more unfounded, nonsensical claims does not reinforce your first unfounded, nonsensical claim.

I'm sure statisticians would love it if the human mind were an inference machine, but that doesn't make it one. Your point of view on this is faith-based.

replies(1): >>45106834 #
1. cubefox ◴[] No.45106834{3}[source]
His view aligns both with a leading neuroscience explanation of the brain (predictive coding [1]) and with Active Inference / the Free Energy principle [2] from optimal control theory. A similar theory of intelligence, called H-JEPA (hierarchical joint embedding predictive architecture) [3] is also put forward by Yann LeCun, a major AI pioneer. Another AI pioneer, Jürgen Schmidhuber, subsequently criticized LeCun's theory, but not for being faulty, but for rehashing ideas published in several papers from the 1990s onward. [4]

1: https://en.wikipedia.org/wiki/Predictive_coding

2: https://en.wikipedia.org/wiki/Free_energy_principle#Active_i...

3: https://openreview.net/forum?id=BZ5a1r-kVsf

4: https://people.idsia.ch/~juergen/lecun-rehash-1990-2022.html

replies(1): >>45119711 #
2. devmor ◴[] No.45119711[source]
The Bayesian brain model is an unfalsifiable, faith-based mechanism - as I alluded to in my previous comment.

Real science is done with it as a starting point, but it is not real science and claiming that it is an accurate representation of the human mind carries as much merit as claiming that "the soul" is what powers human intellect.