←back to thread

214 points optimalsolver | 7 comments | | HN request time: 0.769s | source | bottom
1. hirako2000 ◴[] No.45770845[source]
Has any one ever found an ML/AI paper that make claims that RLMs can reason?

When I prompt an RLM, I can see it spits out reasoning steps. But I don't find that evidence RLMs are capable of reasoning.

replies(3): >>45770918 #>>45770977 #>>45771339 #
2. Sharlin ◴[] No.45770918[source]
Semantics schemantics.
replies(1): >>45771659 #
3. _heimdall ◴[] No.45770977[source]
That would require the ability to understand what happens inside the system during inference when the output is created and they can't do that today.

There's no evidence to be had when we only know the inputs and outputs of a black box.

4. tempfile ◴[] No.45771339[source]
I don't understand what point you are making. Doesn't the name "Reasoning language models" claim that they can reason? Why do you want to see it explicitly written down in a paper?
replies(2): >>45771590 #>>45774621 #
5. hirako2000 ◴[] No.45771590[source]
This very paper sits on the assumption reasoning (to solve puzzles) is at play. It calls those LLMs RLMs.

Imo the paper itself should have touched on the lack of paper discussing what's in the blackbox that makes them Reasoning LMs. It does mention some tree algorithm supposedly key to reasoning capabilities.

By no means attacking the paper as its intent is to demonstrate the lack of success to even solve simple to formulate, complex puzzles.

I was not making a point, I was genuinely asking in case someone knows of papers I could read on that make claims with evidence that's those RLM actually reason, and how.

6. hirako2000 ◴[] No.45771659[source]
It's a statistical imitation of a reasoning pattern, underlying mechanism is pattern matching. The ability to create a model that can determine two radically different words have strong similarity in meaning doesn't imply emergence of some generalizable, logical model that suddenly can Reason to solve novel problems.

Pattern matching is a component of reason. Not === reason.

7. tekno45 ◴[] No.45774621[source]
By renaming this binary to a "Mind reading language model" We now can read your mind and predict your choices just by chatting.

Don't ask how it works cuz its called a "Mind reading language model" duh.