> Modern LLMs are fine-tuned for safety and instruction-following, meaning they are trained to refuse harmful requests.
It's sad that it's now an increasingly accepted idea that information one seeks can be "harmful".
replies(5):
It's sad that it's now an increasingly accepted idea that information one seeks can be "harmful".
So I'm not sure how much it matters if the LLM masters prevent it from repeating things that are overtly racist, or quoting how to make thermite from the Jolly Roger. (I wouldn't trust GPT-4's recipe for thermite even if it would give one). At the end of the day, the degradation of truth and fidelity of the world's knowledge is the ultimate harm that's unavoidable in a technology that is purported to be intelligent but is in fact a black box autocomplete system spewing endless garbage into our infosphere.