Most active commenters
  • al_borland(3)

←back to thread

I'm absolutely right

(absolutelyright.lol)
648 points yoavfr | 13 comments | | HN request time: 0.433s | source | bottom
Show context
trjordan ◴[] No.45138620[source]
OK, so I love this, because we all recognize it.

It's not fully just a tic of language, though. Responses that start off with "You're right!" are alignment mechanisms. The LLM, with its single-token prediction approach, follows up with a suggestion that much more closely follows the user's desires, instead of latching onto it's own previous approach.

The other tic I love is "Actually, that's not right." That happens because once agents finish their tool-calling, they'll do a self-reflection step. That generates the "here's what I did response" or, if it sees an error, the "Actually, ..." change in approach. And again, that message contains a stub of how the approach should change, which allows the subsequent tool calls to actually pull that thread instead of stubbornly sticking to its guns.

The people behind the agents are fighting with the LLM just as much as we are, I'm pretty sure!

replies(11): >>45138772 #>>45138812 #>>45139686 #>>45139852 #>>45140141 #>>45140233 #>>45140703 #>>45140713 #>>45140722 #>>45140723 #>>45141393 #
1. al_borland ◴[] No.45140722[source]
In my experience, once it starts telling me I’m right, we’re already going downhill and it rarely gets better from there.
replies(4): >>45141151 #>>45143167 #>>45145334 #>>45146082 #
2. flkiwi ◴[] No.45141151[source]
Sometimes I just ride the lightning to see how off course it is willing to go. This is not a productive use of my time but it sure is amusing.

In fairness, I’ve done the same thing to overconfident junior colleagues.

replies(3): >>45141631 #>>45142374 #>>45165030 #
3. al_borland ◴[] No.45141631[source]
I spent yesterday afternoon doing this. It go to the point where it would acknowledge it was wrong, but would keep giving me the same answer.

It also said it would only try one more time before giving up, but then kept going.

replies(2): >>45142992 #>>45145937 #
4. neilv ◴[] No.45142374[source]
They should've called it "riding the lightning":

https://en.wikipedia.org/wiki/Socratic_dialogue

5. dingnuts ◴[] No.45142992{3}[source]
this happens to me constantly, it's such a huge waste of time. I'm not convinced any of these tools actually save time. It's all a fucking slot machine and Gell-Mann Amnesia and at the end, you often have nothing that works.

I spent like two hours yesterday dicking with aider to make a one line change and it hallucinated an invalid input for the only possible parameter and I wound up using the docs the old fashioned way and doing the task in about two minutes

replies(1): >>45143240 #
6. anthem2025 ◴[] No.45143167[source]
Usually it’s a response to my profanity laden “what are you doing? Why? Don’t do that! Stop! Do this instead”
7. brianwawok ◴[] No.45143240{4}[source]
The mistake was using AI for a two minute fix. It totally helps at some tasks. Takes some failures to realize that it does indeed have flaws.
replies(1): >>45143957 #
8. ◴[] No.45143957{5}[source]
9. lemming ◴[] No.45145334[source]
Yeah, I want a feature which stops my agent as soon as it says anything even vaguely like: "let me try another approach". Right after that is when the wheels start falling off, tests get deleted, etc. That phrase is a sure sign the agent should (but never does) ask me for guidance.
replies(1): >>45147506 #
10. CuriouslyC ◴[] No.45145937{3}[source]
The worst is when it can't do something right, and it does a horrible mock/hack to get it "working." I had a claude fake benchmark data, that pissed me off a bit, though I did make a major architectural improvement to a tool as result (though the real benchmark would have probably made me do it anyhow) so it wasn't all horrible.
11. ◴[] No.45146082[source]
12. al_borland ◴[] No.45147506[source]
I’ve found even giving guidance at this point doesn’t help, as it fundamentally doesn’t get it.

I was down one of these rabbit holes with it once while having it write a relatively simple bash script. Something I had written by hand previously in Python, but wanted a bash version and also wanted to see what AI could do.

It was 98% there, but couldn’t get that last 2% to save its life. Eventually I went through the code myself, found the bug, and I told it exactly what the bug was and where it was at; it was an off-by-one error. Even when spoon feeding it, it couldn’t fix it and I ended up doing it myself just to get it over with.

13. theshrike79 ◴[] No.45165030[source]
It's also educational to play stupid with LLMs

Even if you know exactly where the issue is an it would be a 30 second job to do it manually, you can get a better feel on how to direct it to see what you need to tell it to have it find out the issue and fixing it.