It seems odd to complain that computers are using human's artwork to inspire their own creations. Every human artist has done the exact same thing in their lifetime; it's unavoidable.
It seems odd to complain that computers are using human's artwork to inspire their own creations. Every human artist has done the exact same thing in their lifetime; it's unavoidable.
Agree.
But you also have to treat code the same way. We shouldn't be suing Open AI and Microsoft over copilot being trained on open source code. It's no different than models trained on art.
Besides, if Microsoft loses, they actually win. I expect they're one of the few companies with enough code to train the model on completely proprietary data. If they lose the case, they'll still be able to build the tool. The rest of us will be locked out of easy training data and won't be able to compete.
It's like a very complicated form of linear interpolation:
a*x + (1-a)*y
These systems do not "think". Today I spent all day mulling an idea, experimenting with variations, feeling frustrated or excited, imagining it, simulating it, making mistakes, following paths of reasoning, deducing facts, revisiting dead-ends with new insight, daydreaming, talking to my wife about it, etc. That's human thought.These models do not "think" like a human, they do not dream or imagine or feel. They run a feed-forward system of linear equations (matrix multiplications).
They INTERPOLATE HUMAN WORK.
They don't exist without training data (huge amounts of intellectual property) aggregated and interpolated in a monstrous perversion of "fair use":
https://bugfix-66.com/7a82559a13b39c7fa404320c14f47ce0c304fa...
Starve the machine. Without your work, it's got nothing.
I don't find it odd to complain that publishing an artwork on DeviantArt has gone from "I intend humans to look at this" to "I (opt-out!) agree that a corporation may use this to generate new artwork for profit."
I would not complain if a painting of mine were exhibited in a museum and someone came in to look at it and draw something inspired by it.
I would complain if I handed over a painting of mine to that same museum to be exhibited, they scanned it in at high resolution, handed it over to a class of copy artists, who then produced artwork in order to compete with mine, before finally putting it up in a gallery.
Does that still seem odd?
There's loads of BSD code, and to share, all that is required is a link to attributation.
It seems to me, there is money to be made, in getting model data colinked with "saw first" references.
Then, for example, after a github style codebot writes the code for you, it can show a link to "where it learned to help you today!".
There is no technical reason the can't be done, only a business model reason.
That said, I find the comments in this thread strange. Discussion about how tech moves on, and looms and such.
That arg was lost, when people could cut up songs, and slap them together, or cut up 100 textbooks into one. This is settled by endless laws, and caselaw. It isn't a new argument. Microsoft will lose.
We are talking about creative works being shuffled together and remixed as a legal protection for theft. That's all it is. There is ample evidence that these algorithms merely regurgitate what goes in and cannot create something entirely new. Which, is of coursw what you'd expect if you understand what is going on under the hood. But it is not what is being sold.
But again… aren’t people the same way? Noone exists in isolation. The Sir Isaac Newton quote comes to mine:
“If I have seen further, it is by standing on the shoulders of giants”
Edit: to be clear - these algorithms are specifically non-linear and are a far cry from ‘linear interpolation’. Yes they do involve matrix multiplication that does not make them interpolaters unless you want to water down the meaning of interpolation to be so generic it loses its meaning. Having said all that - the sophistication of the algorithm is beyond the point here as long as what they are generating is substantially transformative (which >99% of the possible outputs are legally speaking).
I feel like this post by an HN user is pertinent[1].
> Have you ever done any reading on the Luddites? They weren't the anti technology, anti progress social force people think they were.
> They were highly skilled laborers who knew how to operate complex looms. When auto looms came along, factory owners decided they didn't want highly trained, knowledgeable workers they wanted highly disposable workers. The Luddites were happy to operate the new looms, they just wanted to realize some of the profit from the savings in labor along with the factory owners. When the factory owners said no, the Luddites smashed the new looms.
The Luddites went from middle class business owners and craftsmen to utter destitution. Many of the Luddites were tried for machine breaking and were either executed by the state, or exiled to penal colonies. They risked literally everything, because everything was at stake.
I bring this up because people like to pretend the Luddites were some cult of ignorant technophobes, but the reality is that many of us are in the same situation the Luddites were in, as highly skilled workers that operate complex machinery with comfortable middle class lives, before owners cut them out and their families starved in the streets.
Like a feed-forward chain of matrix multiplications, trained to predict its training data?
No, of course you weren't. That would be FUCKING RIDICULOUS.
Yes. They are angry that their labor was used to create something new and arguably more efficient, but they don't get a appropriate compensation for it.
On the plus side. I can imagine this tech empowering artists to create more stuff they previously couldn’t. I’m imagining a single person producing a whole animation which previously was only accessible to companies and teams.
This is not in good faith, please read HN rules.
Rather than attack me (calling me foolish, swearing at me) why don’t you rebut my ideas and have a conversation if you actually have something to contribute.
I’ve read the papers, I’ve worked personally with these systems. I understand them just fine. Notice that I said earlier: “regardless of how simple they are”. I understand you are trying to water them down to be simple interpolation which they definitely are not but even if they were that simple it wouldn’t change the legal calculus here one bit. New art is being generated (far beyond any ‘transformative’ legal test precedent) and any new art that is substantively different from its inputs is legally protectable.
The "product" that chess players produce is not replaceable by ML systems. The game itself, the "fight" of two minds (or one mind against the machine, in the past) is the "product". Watching two chess AIs play against each other can't replace that.
For artists, the product is their output, the art itself. An approximation of that art can also be produced by a ML system now, making artists an unnecessary cost factor[1] for e.g. simple illustrations.
They are not comparable, IMO. Chess players are not replaced by ML systems, artists will be.
> it's unavoidable.
It really isn't. Of course it would be possible to just outlaw the use of things like "the pile", which includes gigabytes of random texts with unknown copyright status. The same goes for any training set that uses images scraped of the web, ignoring any copyright.
Yes, people would still do it, but it would have the same status that piracy has. You can't build a US multi-billion dollar company on piracy (for long), and you wouldn't be able to do so with ML systems that were trained on random stuff from the internet.
I don't think this, in such broad strokes, would be a good thing, to be clear. Such datasets are great for research! But I have a really hard time understanding this defeatism that there is "nothing we can do".
[1] from the perspective of some customers e.g. magazines or ad companies - I don't agree with this
Personally, I don’t think it is likely that copyright laws will change to protect against algorithmic usage (too much precedent in more general reuse cases and for what is considered transformative). Having said that I also don’t think this will be the death of artists by any stretch, some industries will need to change or evolve but it will be just another tool in an artists belt IMO.
Notably in your example both are certainly legal just varied by the level of controversy around them and on that level I would agree. Scaled up processes do tend to attract more controversy.
For professional artists who do it for the money, yes, that's true.
For amateur artists, the product can be the process, the flow of creating art. Futhermore, I'd say a lot of art isn't about conveying an idea or whatever. You see something, you paint it, because you like it, give it your own spin. Maybe the end result is good, maybe not. Often enough, the art becomes "valuable", because others give it some new context.
> Chess players are not replaced by ML systems, artists will be.
Traditional artists working with real materials won't be. They might even get new interest, because digital art will be flodded with spam.
Or a magazine does, what the art scene has been about forever: Hire artists because of their name or their background.
The job "digital artist" was created roughly 20-30 years ago and now is transformed to something else or might become obsolete. Bummer for digital artists, but not sure if this will destroy "artists" in general.
It's a problem of scale.
> Because if anything, I think AI-generated art is in the process of disproving this exact hypothesis
But it's not creating anything, it's regurgitating it's training material (through a suitably fine blender) in the way that scores best. These models are nothing without the actual art they've appropriated.
It isn't "inspiration". These machine models aren't actually intelligent. There's no expressive element here, with regard to the machine producing the art.
What it really is is just a new tool for producing art. The sculptor had his chisel, the painter had a paintbrush, the photographer had a camera, the graphic designer had Photoshop or whatever, and now you can make art by being skillful in coming up with a prompt. It still requires skill, skill with the tool, just like anything else.
The difference is that this new tool (probably) doesn't enable the creation of anything truly novel.
It would chill the whole ML space significantly for decades, IMO, as the only truly safe data would be synthetic or licensed. This can work for some applications (e.g. Microsoft used synthetic data for facial landmark recognition[1]), but it would kill DALL-E 2 et al.
Humans, artists aren't the machines other human created, they interpret or copy, not interpolate.