←back to thread

393 points pyman | 1 comments | | HN request time: 0.213s | source
Show context
codedokode ◴[] No.44492195[source]
If AI companies are allowed to use pirated material to create their products, does it mean that everyone can use pirated software to create products? Where is the line?

Also please don't use word "learning", use "creating software using copyrighted materials".

Also let's think together how can we prevent AI companies from using our work using technical measures if the law doesn't work?

replies(5): >>44492257 #>>44492400 #>>44492975 #>>44493804 #>>44493829 #
redcobra762 ◴[] No.44492400[source]
It's abusive and wrong to try and prevent AI companies from using your works at all.

The whole point of copyright is to ensure you're paid for your work. AI companies shouldn't pirate, but if they pay for your work, they should be able to use it however they please, including training an LLM on it.

If that LLM reproduces your work, then the AI company is violating copyright, but if the LLM doesn't reproduce your work, then you have not been harmed. Trying to claim harm when you haven't been due to some philosophical difference in opinion with the AI company is an abuse of the courts.

replies(4): >>44492530 #>>44492615 #>>44492908 #>>44492935 #
codedokode ◴[] No.44492530[source]
It is not wrong at all. The author decides what to do with their work. AI companies are rich and can simply buy the rights or hire people to create works.

I could agree with exceptions for non-commercial activity like scientific research, but AI companies are made for extracting profits and not for doing research.

> AI companies shouldn't pirate, but if they pay for your work, they should be able to use it however they please, including training an LLM on it.

It doesn't work this way. If you buy a movie it doesn't mean you can sell goods with movie characters.

> then you have not been harmed.

I am harmed because less people will buy the book if they can simply get an answer from LLM. Less people will hire me to write code if an LLM trained on my code can do it. Maybe instead of books we should start making applications that protect the content and do not allow copying text or making screenshots. ANd instead of open-source code we should provide binary WASM modules.

replies(2): >>44492572 #>>44493443 #
1. CaptainFever ◴[] No.44493443[source]
> Maybe instead of books we should start making applications that protect the content and do not allow copying text or making screenshots.

https://en.wikipedia.org/wiki/Analog_hole