←back to thread

1503 points participant3 | 1 comments | | HN request time: 0s | source
Show context
mlsu ◴[] No.43575950[source]
I was really hoping that the conversation around AI art would at least be partially centered on the perhaps now dated "2008 pirate party" idea that intellectual property, the royalty system, the draconian copyright laws that we have today are deeply silly, rooted in a fiction, and used over and over again, primarily by the rich and powerful, to stifle original ideas and hold back cultural innovation.

Unfortunately, it's just the opposite. It seems most people have fully assimilated the idea that information itself must be entirely subsumed into an oppressive, proprietary, commercial apparatus. That Disney Corp can prevent you from viewing some collection of pixels, because THEY own it, and they know better than you do about the culture and communication that you are and are not allowed to experience.

It's just baffling. If they could, Disney would scan your brain to charge you a nickel every time you thought of Mickey Mouse.

replies(31): >>43576033 #>>43576035 #>>43576039 #>>43576072 #>>43576095 #>>43576129 #>>43576200 #>>43576201 #>>43576223 #>>43576381 #>>43576435 #>>43576475 #>>43576488 #>>43576594 #>>43576625 #>>43576663 #>>43576709 #>>43576768 #>>43576774 #>>43576782 #>>43576815 #>>43576826 #>>43576933 #>>43577120 #>>43577458 #>>43577553 #>>43577827 #>>43577984 #>>43578013 #>>43578038 #>>43581949 #
eaglelamp ◴[] No.43576663[source]
If we are going to have a general discussion about copyright reform at a national level, I'm all for it. If we are going to let billion dollar corporations break the law to make even more money and invent legal fictions after the fact to protect them, I'm completely against it.

Training a model is not equivalent to training a human. Freedom of information for a mountain of graphics cards in a privately owned data center is not the same as freedom of information for flesh and blood human beings.

replies(2): >>43576850 #>>43577289 #
r3trohack3r ◴[] No.43576850[source]
You’re setting court precedent that will apply equally to OpenAI as it does to the llama.cpp and stable diffusion models running on your own graphics card.
replies(3): >>43576934 #>>43576962 #>>43577064 #
munificent ◴[] No.43577064[source]
SGTM.

Honestly, seriously. Imagine some weird Thanos showed up, snapped his fingers and every single bit of generative AI software/models/papers/etc. were wiped from the Earth forever.

Would that world be measurably worse in any way in terms of meaningful satisfying lives for people? Yes, you might have to hand draw (poorly) your D&D character.

But if you wanted to read a story, or look at an image, you'd have to actually connect with a human who made that thing. That human would in turn have an audience for people to experience the thing they made.

Was that world so bad?

replies(3): >>43577131 #>>43578277 #>>43579164 #
1. FeepingCreature ◴[] No.43579164[source]
That world was worse. It wasn't much worse, because we haven't seen most of the benefit of GenAI yet, but yes I would say that it was worse.

It wasn't "so bad", but any history of improvement can be cut into slices that aren't "so bad" to reverse.