←back to thread

283 points Brajeshwar | 2 comments | | HN request time: 0s | source
Show context
cs702 ◴[] No.45231366[source]
The title is biased, blaming Google for mistreating people and implying that Google's AI isn't smart, but the OP is worth reading, because it gives readers a sense of the labor and cost involved in providing AI models with human feedback, the HF in RLHF, to ensure they behave in ways acceptable to human beings, more aligned with human expectations, values, and preferences.
replies(6): >>45231394 #>>45231412 #>>45231441 #>>45231748 #>>45231773 #>>45233975 #
1. zozbot234 ◴[] No.45231748[source]
RLHF (and its evolution, RLAIF) is actually used for more than setting "values and preferences". It's what makes AI models engage in recognizable behavior, as opposed to simply continuing a given text. It's how the "Chat" part of "ChatGPT" can be made to work in the first place.
replies(1): >>45232111 #
2. cs702 ◴[] No.45232111[source]
Yes. I updated my comment to reflect as much. Thank you.