←back to thread

747 points porridgeraisin | 1 comments | | HN request time: 0s | source
Show context
I_am_tiberius ◴[] No.45062905[source]
In my opinion, training models on user data without their real consent (real consent = e.g. the user must sign a contract or so, so he's definitely aware), should be considered a serious criminal offense.
replies(5): >>45062989 #>>45063008 #>>45063221 #>>45063771 #>>45064402 #
jsheard ◴[] No.45062989[source]
Why single out user data specifically? Most of the data Anthropic and co train on was just scooped up from wherever with zero consent, not even the courtesy of a buried TOS clause, and their users were always implicitly fine with that. Forgive me for not having much sympathy when the users end up reaping what they've sown.
replies(3): >>45063012 #>>45063051 #>>45063335 #
perihelions ◴[] No.45063051[source]
Training on private user interactions is a privacy violation; training on public, published texts is (some argue) an intellectual property violation. They're very different kinds of moral rights.
replies(2): >>45063481 #>>45064161 #
1. jsheard ◴[] No.45064161[source]
I wish I could be so optimistic that there is no private information published unintentionally or maliciously on the open web where crawlers can find it.

(and as diggan said, the web isn't the only source they use anyway. who knows what they're buying from data brokers.)