←back to thread

385 points vessenes | 1 comments | | HN request time: 0.245s | source

So, Lecun has been quite public saying that he believes LLMs will never fix hallucinations because, essentially, the token choice method at each step leads to runaway errors -- these can't be damped mathematically.

In exchange, he offers the idea that we should have something that is an 'energy minimization' architecture; as I understand it, this would have a concept of the 'energy' of an entire response, and training would try and minimize that.

Which is to say, I don't fully understand this. That said, I'm curious to hear what ML researchers think about Lecun's take, and if there's any engineering done around it. I can't find much after the release of ijepa from his group.

1. alok-g ◴[] No.43371664[source]
>> believes LLMs will never fix hallucinations because, essentially, the token choice method at each step leads to runaway errors

I may have an actual opinion on his viewpoint, however, I have a nitpick even before that.

How exactly is 'LLM' defined here? Even if some energy-based thing is done, would some not call even that an LLM? If/when someone finds a way to fix it within the 'token choice' method, could some people not just start calling it something differently from 'LLM'.

I think Yann needs to rephrase what exactly he wants to say.