←back to thread

385 points vessenes | 10 comments | | HN request time: 0.971s | source | bottom

So, Lecun has been quite public saying that he believes LLMs will never fix hallucinations because, essentially, the token choice method at each step leads to runaway errors -- these can't be damped mathematically.

In exchange, he offers the idea that we should have something that is an 'energy minimization' architecture; as I understand it, this would have a concept of the 'energy' of an entire response, and training would try and minimize that.

Which is to say, I don't fully understand this. That said, I'm curious to hear what ML researchers think about Lecun's take, and if there's any engineering done around it. I can't find much after the release of ijepa from his group.

1. janalsncm ◴[] No.43366161[source]
I am an MLE not an expert. However, it is a fundamental problem that our current paradigm of training larger and larger LLMs cannot ever scale to the precision people require for many tasks. Even in the highly constrained realm of chess, an enormous neural net will be outclassed by a small program that can run on your phone.

https://arxiv.org/pdf/2402.04494

replies(3): >>43366173 #>>43366198 #>>43368618 #
2. throw310822 ◴[] No.43366173[source]
> Even in the highly constrained realm of chess, an enormous neural net will be outclassed by a small program that can run on your phone.

This is true also for the much bigger neural net that works in your brain, and even if you're the world champion of chess. Clearly your argument doesn't hold water.

replies(1): >>43366869 #
3. thewarrior ◴[] No.43366198[source]
Any chance that “reasoning” can fix this
replies(1): >>43366940 #
4. janalsncm ◴[] No.43366869[source]
For the sake of argument let’s say an artificial neural net is approximately the same as the brain. It sounds like you agree with me that smaller programs are both more efficient and more effective than a larger neural net. So you should also agree with me that those who say the only path to AGI is LLM maximalism are misguided.
replies(2): >>43366985 #>>43367106 #
5. janalsncm ◴[] No.43366940[source]
It kind of depends. You can broadly call any kind of search “reasoning”. But search requires 1) enumerating your possible options and 2) assigning some value to those options. Real world problem solving makes both of those extremely difficult.

Unlike in chess, there’s a functionally infinite number of actions you can take in real life. So just argmax over possible actions is going to be hard.

Two, you have to have some value function of how good an action is in order to argmax. But many actions are impossible to know the value of in practice because of hidden information and the chaotic nature of the world (butterfly effect).

replies(1): >>43367516 #
6. jpadkins ◴[] No.43366985{3}[source]
smaller programs are better than artificial or organic neural net for constrained problems like chess. But chess programs don't generalize to any other intelligence applications, like how organic neural nets do today.
7. throw310822 ◴[] No.43367106{3}[source]
> It sounds like you agree with me that smaller programs are both more efficient and more effective than a larger neural net.

At playing chess. (But also at doing sums and multiplications, yay!)

> So you should also agree with me that those who say the only path to AGI is LLM maximalism are misguided.

No. First of all, it's a claim you just made up. What we're talking about is people saying that LLMs are not the path to AGI- an entirely different claim.

Second, assuming there's any coherence to your argument, the fact that a small program can outclass an enormous NN is irrelevant to the question of whether the enormous NN is the right way to achieve AGI: we are "general intelligences" and we are defeated by the same chess program. Unless you mean that achieving the intelligence of the greatest geniuses that ever lived is still not enough.

8. artificialprint ◴[] No.43367516{3}[source]
Isn't something about alphago also involves "infinitely" many possible outcomes? Yet they cracked it, right?
replies(1): >>43367625 #
9. janalsncm ◴[] No.43367625{4}[source]
Go is played on a 19x19 board. At the beginning of the game the first player has 361 possible moves. The second player then has 360 possible moves. There is always a finite and relatively “small” number of options.

I think you are thinking of the fact that it had to be approached in a different way than Minimax in chess because a brute force decision tree grows way too fast to perform well. So they had to learn models for actions and values.

In any case, Go is a perfect information game, which as I mentioned before, is not the same as problems in the real world.

10. ifdefdebug ◴[] No.43368618[source]
Best in class chess program actually is a NN, just not a LLM.