←back to thread

221 points whitefables | 1 comments | | HN request time: 0s | source
Show context
netdevnet ◴[] No.41856729[source]
Am I right thinking that a self-hosted llama wouldn't have the kind restrictions ChatGPT has since it has no initial system prompt?
replies(4): >>41856734 #>>41856777 #>>41856779 #>>41856872 #
dtquad ◴[] No.41856777[source]
All the self-hosted LLM and text-to-image models come with some restrictions trained into them [1]. However there are plenty of people who have made uncensored "forks" of these models where the restrictions have been "trained away" (mostly by fine-tuning).

You can find plenty of uncensored LLM models here:

https://ollama.com/library

[1]: I personally suspect that many LLMs are still trained on WebText, derivatives of WebText, or using synthetic data generated by LLMs trained on WebText. This might be why they feel so "censored":

>WebText was generated by scraping only pages linked to by Reddit posts that had received at least three upvotes prior to December 2017. The corpus was subsequently cleaned

The implications of so much AI trained on content upvoted by 2015-2017 redditors is not talked about enough.

replies(2): >>41856879 #>>41857328 #
thrdbndndn ◴[] No.41857328[source]
My to-go test for uncensoring is to ask the LLM to write erotic novel.

But I haven't yet find any "uncensored" ones (on ollama) that works. Did I miss something?

(On the contrary: when ChatGPT first came out, it was trivial to jailbreak it to make it write erotica.)

replies(3): >>41857379 #>>41857399 #>>41857875 #
1. dtquad ◴[] No.41857399[source]
Try the popular (pull count) dolphin models:

https://ollama.com/library/dolphin-mistral