> Modern LLMs are fine-tuned for safety and instruction-following, meaning they are trained to refuse harmful requests.
It's sad that it's now an increasingly accepted idea that information one seeks can be "harmful".
replies(5):
It's sad that it's now an increasingly accepted idea that information one seeks can be "harmful".
If the state is censoring the model, I think the problem is more subtle.
Eh, RLHF often amounts to useless moralizing, and even more often leads to refusals that impair the utility of the product. One recent example: I was asking Claude to outline the architectural differences between light water and molten salt reactors, and it refused to answer because nuclear. See related comments on this discussion for other related points.
https://news.ycombinator.com/item?id=40666950
I think there's quite a bit to complain about in this regard.