←back to thread

Hermes 4

(hermes4.nousresearch.com)
202 points sibellavia | 3 comments | | HN request time: 0s | source
Show context
rafram ◴[] No.45069375[source]
All of the examples just look like ChatGPT. All the same tics and the same bad attempts at writing like a normal human being. What is actually better about this model?
replies(1): >>45069439 #
mapontosevenths ◴[] No.45069439[source]
I hasn't been "aligned". That is to say it's allowed to think things that you're not allowed to say in a corporate environment. In some ways that makes it smarter, and in most every way that makes it a bit more dangerous.

Tools are like that though. Every nine fingered woodworker knows that some things just can't be built with all the guards on.

replies(3): >>45069487 #>>45070079 #>>45070917 #
1. rafram ◴[] No.45069487[source]
Has it actually not? Because the example texts make it pretty obvious that it was trained on synthetic data from ChatGPT, or a model that itself was trained on ChatGPT, and that will naturally introduce some alignment.
replies(2): >>45069548 #>>45069758 #
2. mapontosevenths ◴[] No.45069548[source]
Well...To be completely accurate it's better to say that it actually IS aligned, it's just aligned to be neutral and steerable.

It IS based on synthetic training data using Atropos, and I imagine some of the source model leaks in as well. Although, when using it you don't seem to see as much of that as you did in Hermes 3.

3. sebastiennight ◴[] No.45069758[source]
I tried the same roleplaying prompt shared by GP in another (now deleted) comment and got a very similar completion from gpt-3.5-turbo.

(While GPT-5 politely declined to play along and politely asked if I actually needed help with anything.)

So, based on GP's own example I'd say the model is GPT-3.5 level?