←back to thread

Hermes 4

(hermes4.nousresearch.com)
202 points sibellavia | 8 comments | | HN request time: 0.438s | source | bottom
1. mapontosevenths ◴[] No.45069190[source]
I appreciate the effort they put into providing a neutral tool that hasn't been generically forced to behave like "Sue from HR".
replies(3): >>45070045 #>>45070200 #>>45076115 #
2. dcre ◴[] No.45070045[source]
That is the only thing they seem to care about. It’s juvenile.
replies(1): >>45084279 #
3. fl0id ◴[] No.45070200[source]
There is no neutral. It will just be biased based on its training data etc.
replies(1): >>45071082 #
4. beeflet ◴[] No.45071082[source]
A lot of models seem to be biased based on (political, etc.) reinforcement from their trainers.
5. bckr ◴[] No.45076115[source]
I’m having a hard time not being sarcastic here.

The most recent news about chatbots is that ChatGPT coached a kid on how to commit suicide.

Two arguments come to mind. 1) it’s the sycophancy! Nous and its ilk should be considered safer. 2) it’s the poor alignment. A better trained model like Claude wouldn’t have done that.

I lean #2

replies(2): >>45077511 #>>45080008 #
6. karan4d ◴[] No.45077511[source]
the sycophancy is due to poor alignment. the instruct based mode collapse results in this mode collapse induced sycophancy. constitutional alignment is better than the straight torture OAI does to the model, but issues remain
7. mapontosevenths ◴[] No.45080008[source]
> The most recent news about chatbots is that ChatGPT coached a kid on how to commit suicide.

Maybe every tool isn't meant for children or the mentally ill? When someone lets their kid play with a chainsaw that doesn't mean we should ban chainsaws, it means we should ban lousy parents.

8. silverliver ◴[] No.45084279[source]
Still better than OpenAI HR.