←back to thread

755 points MedadNewman | 4 comments | | HN request time: 0.727s | source
1. ks2048 ◴[] No.42891407[source]
Part of the blog is hypothesizing that the censorship is in a separate filtering stage rather than the model itself. But, the example of hex encoding doesn't prove or disprove that at all, does it? Can't you just check on a version running open-source weights?
replies(2): >>42891485 #>>42891489 #
2. amrrs ◴[] No.42891485[source]
I ran the distilled models locally some of the censorships are there.

But on their chat (hosted), deepseek has some keyword based filters - like the moment it generates Chinese president name or other controversial keywords - the "thinking" stops abruptly!

replies(1): >>42891513 #
3. pomatic ◴[] No.42891489[source]
The open source model seems to be uncensored, lending weight to the separate filter concept. Plus, any filter needs to be revised as new workarounds emerge - if it is baked in to the model that requires retraining, whereas it's reasonably light work for a frontend filter.
4. prettyblocks ◴[] No.42891513[source]
The distilled versions I've run through Ollama are absolutely censored and don't even populate the <think></think> section for some of those questions.