I wish they would come up with a better term. Computers do not have brains or conscientiousness.
They erroneously construct responses (i.e., confabulation).
replies(1):
They erroneously construct responses (i.e., confabulation).
LLMs, in a very real way, have "conscientiousness". As in: it's a property that can be measured and affected by training, and also the kind of abstract concept that an LLM can recognize and operate off.
If you can just train an LLM to be "more evil", you can almost certainly train an LLM to be "more conscientious" or "less conscientious".
No, you shouldn't. They hate that.