This feels like an unwarranted anthropomorphization of what LLMs are doing.
This feels like an unwarranted anthropomorphization of what LLMs are doing.
I don't see why it would be different for LLMs.
The issue is the recall LLMs have over copyrighted contents.
Personally, my read is that the issue with most of these cases is that we are treating and talking about LLMs as if they do things that humans do. They don't. They don't reason. They don't think. They don't know. They just map input to probabilistic output. LLMs are a tool like any other for more easily achieving some outcome.
It's precisely because we insist on treating LLMs as if they are more than an inefficient storage device (with a neat/useful trick) that we run into questions like this. I personally think the illegal status of current models should be pretty clear simply based of the pirated nature of their input material. To my understanding, fair use has never before applied to works that were obtained illegally.