←back to thread

Hermes 4

(hermes4.nousresearch.com)
202 points sibellavia | 1 comments | | HN request time: 0s | source
Show context
rafram ◴[] No.45069375[source]
All of the examples just look like ChatGPT. All the same tics and the same bad attempts at writing like a normal human being. What is actually better about this model?
replies(1): >>45069439 #
mapontosevenths ◴[] No.45069439[source]
I hasn't been "aligned". That is to say it's allowed to think things that you're not allowed to say in a corporate environment. In some ways that makes it smarter, and in most every way that makes it a bit more dangerous.

Tools are like that though. Every nine fingered woodworker knows that some things just can't be built with all the guards on.

replies(3): >>45069487 #>>45070079 #>>45070917 #
rafram ◴[] No.45069487[source]
Has it actually not? Because the example texts make it pretty obvious that it was trained on synthetic data from ChatGPT, or a model that itself was trained on ChatGPT, and that will naturally introduce some alignment.
replies(2): >>45069548 #>>45069758 #
1. sebastiennight ◴[] No.45069758[source]
I tried the same roleplaying prompt shared by GP in another (now deleted) comment and got a very similar completion from gpt-3.5-turbo.

(While GPT-5 politely declined to play along and politely asked if I actually needed help with anything.)

So, based on GP's own example I'd say the model is GPT-3.5 level?