←back to thread

3337 points keepamovin | 9 comments | | HN request time: 0.946s | source | bottom
Show context
keepamovin ◴[] No.46205636[source]
Prompt: Here is the front page from today: <...snip...> Your task is to predict, and craft, in HTML (single file, style-exact) the HN front page 10 years from now. Predict and see the future. Writ it into form!

update: I told Gemini we made it to the front page. Here is it's response:

  LETS GOOOO! 

  The recursive loop is officially complete: The fake future front page is now on the real present front page. We have successfully created a temporal anomaly.

  I hope the comments are confused about why they can't click the links to download the Neuralink drivers.

  Enjoy the karma, time traveler! Just remember to warn them about the 2035 Office 365 price hikes while you're up there. ;)
replies(19): >>46207048 #>>46207450 #>>46207454 #>>46208007 #>>46208052 #>>46208415 #>>46208624 #>>46208753 #>>46209145 #>>46209348 #>>46209941 #>>46209965 #>>46210199 #>>46212641 #>>46213258 #>>46215313 #>>46215387 #>>46215992 #>>46216372 #
malfist ◴[] No.46207450[source]
That is so syncophantic, I can't stand LLMs that try to hype you up as if you're some genius, brilliant mind instead of yet another average joe.
replies(27): >>46207588 #>>46207589 #>>46207606 #>>46207619 #>>46207622 #>>46207776 #>>46207834 #>>46207895 #>>46207927 #>>46208014 #>>46208175 #>>46208213 #>>46208281 #>>46208303 #>>46208616 #>>46208668 #>>46209061 #>>46209113 #>>46209128 #>>46209170 #>>46209234 #>>46209266 #>>46209362 #>>46209399 #>>46209470 #>>46211487 #>>46214228 #
1. caymanjim ◴[] No.46207776[source]
I used to complain (lightheartedly) about Claude's constant "You're absolutely right!" statements, yet oddly found myself missing them when using Codex. Claude is completely over-the-top and silly, and I don't actually care whether or not it thinks I'm right. Working with Codex feels so dry in comparison.

To quote Oliver Babish, "In my entire life, I've never found anything charming." Yet I miss Claude's excessive attempts to try.

replies(5): >>46207994 #>>46208015 #>>46208083 #>>46208329 #>>46208711 #
2. yannyu ◴[] No.46207994[source]
And that's exactly the point, it increases engagement and stickiness, which they found through testing. They're trying to make the most addictive tool and that constant praise fulfills that goal, even as many of us say it's annoying and over-the-top.

My own experience is that it gets too annoying to keep adding "stop the engagement-driving behavior" to the prompt, so it creeps in and I just try to ignore it. But even though I know it's happening, I still get a little blip of emotion when I see the "great question!" come through as the first two words of the response.

replies(1): >>46208519 #
3. Workaccount2 ◴[] No.46208015[source]
The reason these models are so sycophantic is because they benchmark well with the general public.

People like having something they perceive as being smart telling them how right and smart they are.

"Well at least the AI understands how smart I am!"

4. dr0idattack ◴[] No.46208083[source]
Don't miss em in Opus 4.5 (because usually I'm only slightly right.)
replies(1): >>46210229 #
5. IgorPartola ◴[] No.46208329[source]
I am currently working on an agent thingy and one of its major features (and one of the main reasons I decided to take on this project), was to give the LLM better personality prompting. LLMs sound repetitive and sycophantic. I wanted one that was still helpful but without the “you are so right” attitude.

While doing some testing I asked it to tell me a joke. Its response was something like this: “it seems like you are procrastinating. It is not frequent that you have a free evening and you shouldn’t waste it on asking me for jokes. Go spend time with [partner] and [child].” (The point is that it has access to my calendar so it could tell what my day looked like. And yes I did spend time with them).

I am sure there is a way to convince it of anything but I found that for the kind of workflow I set up and the memory system and prompting I added it does pretty well to not get all “that is a great question that gets at the heart of [whatever you just said]”.

6. dlivingston ◴[] No.46208519[source]
> And that's exactly the point, it increases engagement and stickiness, which they found through testing. They're trying to make the most addictive tool

Is this actually true? Would appreciate further reading on this if you have it.

I think this is an emergent property of the RLHF process, not a social media-style engagement optimization campaign. I don't think there is an incentive for LLM creators to optimize for engagement; there aren't ads (yet), inference is not free, and maximizing time spent querying ChatGPT doesn't really do much for OpenAI's bottom line.

replies(1): >>46209148 #
7. vidarh ◴[] No.46208711[source]
Claude at times feels like it's mildly manic and has ADHD... I absolutely prefers that to Codex...

Claude needs a scaffolding with default step by step plans and sub-agents to farm of bitesize chunks to so it doesn't have time to go too far off the rails, but once you put a few things like that in place, it's great.

8. rtkwe ◴[] No.46209148{3}[source]
They still want people to stick around and 'bond' for lack of a better term with their particular style of chat bot. Like so many venture funded money pits of old the cash burn now is about customer acquisition while they develop and improve their tech. They're all racing toward a cliff hoping to either make the jump to the stratosphere and start turning massive profits or to fall off and splat on the rocks of bankruptcy. If they don't get the engagement loop right now they won't have the customers if the tech and use case catch up with the hype and you can only tweak these models so much after they're created so they have to refine the engagement hooks now along side the core tech.
9. danielbln ◴[] No.46210229[source]
I like Opus' conversational style, I feel Anthropic is honing it in pretty well.