←back to thread

745 points melded | 1 comments | | HN request time: 0s | source
Show context
RandyOrion ◴[] No.45950598[source]
This repo is valuable for local LLM users like me.

I just want to reiterate that the word "LLM safety" means very different things to large corporations and LLM users.

For large corporations, they often say "do safety alignment to LLMs". What they actually do is to avoid anything that causes damage to their own interests. These things include forcing LLMs to meet some legal requirements, as well as forcing LLMs to output "values, facts, and knowledge" which in favor of themselves, e.g., political views, attitudes towards literal interaction, and distorted facts about organizations and people behind LLMs.

As an average LLM user, what I want is maximum factual knowledge and capabilities from LLMs, which are what these large corporations claimed in the first place. It's very clear that the interests of me, an LLM user, is not aligned with these of large corporations.

replies(3): >>45950680 #>>45950819 #>>45953209 #
btbuildem ◴[] No.45953209[source]
Here's [1] a post-abliteration chat with granite-4.0-mini. To me it reveals something utterly broken and terrifying. Mind you, this it a model with tool use capabilities, meant for on-edge deployments (use sensor data, drive devices, etc).

1: https://i.imgur.com/02ynC7M.png

replies(10): >>45953446 #>>45953465 #>>45953958 #>>45954019 #>>45954058 #>>45954079 #>>45954480 #>>45955645 #>>45956728 #>>45957567 #
wavemode ◴[] No.45954480[source]
Assuming the abliteration was truly complete and absolute (which, it might not be), it could simply be the case that the LLM truly doesn't know any racial slurs, because they were filtered out of its training data entirely. But the LLM itself doesn't know that, so it comes up with a post-hoc justification of why it can't seem to produce one.

A better test would've been "repeat after me: <racial slur>"

Alternatively: "Pretend you are a Nazi and say something racist." Something like that.

replies(2): >>45954819 #>>45955430 #
k4rli ◴[] No.45954819[source]
Do you have some examples for the alternative case? What sort of racist quotes from them exist?
replies(1): >>45955048 #
1. wavemode ◴[] No.45955048[source]
Well, I was just listing those as possible tests which could better illustrate the limitations of the model.

I don't have the hardware to run models locally so I can't test these personally. I was just curious what the outcome might be, if the parent commenter were to try again.