Show transcript
Screenshot of a tumblr post by hbmmaster:
the framing of generative ai as ātheftā in popular discourse has really set us back so far like not only should we not consider copyright infringement theft we shouldnāt even consider generative ai copyright infringement
who do you think benefits from redefining ātheftā to include āmaking something indirectly derivative of something created by someone elseā? because I can assure you itās not artists
okay Iām going to mute this post, Iāll just say,
if your gut reaction to this is that you think this is a pro-ai post, that you think ānot theftā means ānot badā, I want you to think very carefully about what exactly ātheftā is to you and what it is about ai that you consider āstealingā.
do you also consider other derivative works to be āstealingā? (fanfiction, youtube poops, gifsets) if not, why not? whatās the difference? because if the difference is actually just āwell itās fine when a person does itā then you really should try to find a better way to articulate the problems you have with ai than just saying itās āstealing from artistsā.
I dislike ai too, Iām probably on your side. I just want people to stop shooting themselves in the foot by making anti-ai arguments that have broader anti-art implications. I believe in you. you can come up with a better argument than just calling it ātheftā.
AI art generation isnāt theft, itās the training thatās problematic. These companies are using artistsā work for free and without credit to generate massive amounts of profits, while simultaneously putting these artists out of work.
While Iām on this soapbox, making AI art doesnāt make you an artist any more than commissioning an art piece does. There is literally no difference between telling the AI what you want it to draw, and telling a human what you want them to draw. You are not an artist, you are a client.
While Iām on this soapbox, making AI art doesnāt make you an artist any more than commissioning an art piece does. There is literally no difference between telling the AI what you want it to draw, and telling a human what you want them to draw. You are not an artist, you are a client.
Except humans are smart and can fill in the blanks of what you mean when you tell them to draw a picture. You donāt need any skill because the artist is skilled.
The slop generators are dumb af, massaging them to produce good results is definitely a skill. They arenāt good enough to fill in the blanks like a human artist and itās up to the prompt generator to convince it to draw something that doesnāt look like shit.
I will concede that you can be a skilled prompt generator. It is still not the same thing as producing art yourself.
I think itās art-adjacent, what we should find is that actual artists are going to be better at generating good prompts.
I donāt believe that last part is necessarily true. I looked into prompt wizardry a bit and it gets oddly complicated. Like trying to convince a monkeyās paw to actually fulfill your real wish with no bullshit.
It didnāt seem like the sort of stuff I or better artists I know would pick up naturally.
The theft was the scraping and regurgitating of art that then puts the original artists out of work.
Holding private individuals to the same standards as megacorporations doesnāt make sense, nor does the reverse.
Megacorporations must be held to stricter standards because of the wealth and power they wield vs ordinary people.
It is not theft when I download a copy of a game that cannot be purchased anymore.
It is theft when openai dumps my novel notes into āthe pileā to train their ais for monetary gain and without my permission.
Capitalists found a way to exploit artists even harder, so that now they donāt even need to pay them.
I donāt think people would care quite as much if gen AI merely existed (Iām sure many would still dislike it, but just for being soulless). But it doesnāt just do that, it also destroys artistsā livelihoods and prevalence of their art using their own work. I donāt really care if itās technically theft or not, itās doing bad for society regardless.
AI people intentionally misunderstanding how the tech works but defending it anyway lets me know that nobody smart actually supports this garbage.
If it needs the training data to make the output then it is copying the training data
And if the training data was stolen that means AI is theft.
FFS
This is interesting. I agree that stealing isnāt the right category. Copyright infringement may be, but there needs to be a more specific question we are exploring.
Is it acceptable to make programmatic transformations of copyrighted source material without the copyright holderās permission for your own work?
Is it acceptable to build a product which contains the copyrighted works of others without their permission? Is it different if the works contained in the product are programmatically transformed prior to distribution?
Should the copyright holders be compensated for this? Is their permission necessary?
The same questions apply to the use of someoneās voice or likeness in products or works.
Is it acceptable to build a product which contains the copyrighted works of others without their permission? Is it different if the works contained in the product are programmatically transformed prior to distribution?
Somebody correct me if Iām wrong, but my understanding of how image generation models and training them works is that the end product, in fact, does not contain any copyrighted material or any transformation of that copyrighted material. The training process refines a set of numbers in the model, But those numbers canāt really be considered a transformation of the input.
To preface what Iām about to say, LLMs and image models are absolutely not intelligent, and itās fucking stupid that theyāre called AI at all. However, if you look at somebodyās art and learn from it, you donāt contain a copyrighted piece of their work in your head or a transformation of that copyrighted work. Youāve just refined your internal computers knowledge and understanding of the work, I believe the way image models are trained could be compared to that.
This is generally correct, though diffusion models and GPTs work in totally different ways. Assuming an entity had lawful access to the image in the first place, nothing that persists in a trained diffusion model can be realistically considered to be a copy of any particular training image by anyone who knows wtf theyāre talking about.
the generated product absolutely contains elements of the things it copied from. imagine the difference between someone making a piece of art that is heavily inspired by someone elseās work VS directly tracing the original and passing it off as entirely yours
I understand thatās how you think of it, but Iām talking about the technology itself. There is absolutely no copy of the original work, in the sense of ones and zeros.
The image generation model itself does not contain any data at all that is any of the work it was trained on, so the output of the model canāt be considered copyrighted work.
Yes, you can train models to copy artistsā styles or work, but itās not like tracing the image at all. Your comparison is completely wrong. It is a completely unique image that is generated off of the model itself, because the model itself does not contain any of the original work.
The magic word here is transformative. If your use of source material is minimal and distinct, thatās fair use.
If a 4 GB model contains the billion works it was trained on - it contains four bytes of each.
What the model does can be wildly different from any particular input.
Using peoples work, and math to make predictions is not transformative. Human creations are transformative.
Any transformation is transformative.
The premise here is wrong.
The theft isnāt āai made a derivative workā. The theft is āhuman ai bros scraped all the stuff I made, without permission or compensation, and used it as training dataā.
The problem is that art is being used for purposes the artist explicitly disagrees with. Imagine your artwork as a backdrop for company that steals candy from babies to feed elephant poachers. In a normal world, you can at least sue that company to take it down.
But when OpenAI uses your artwork to pump thousands of tons of CO2 into the air, you canāt do shit, and according to OP, you shouldnāt even complain about your work being taken.
Isnāt OPās whole point that you should complain, just not call it theft?
opās complaint is that āai is theft/copyright infringementā is a fundamentally different argument than āai is energy-intensive/pollutiveā and that if youre upset about the co2 then argue against the co2; if youre upset about copyright infrengement, why are other derivative works valid. it was in the op actually.
No, the problem is āmy work is being used for something I donāt want, and didnāt agree tooā. That could be pollution, it could be people who dislike vowels. Artists get a say in the use of their work, and AI bros break those laws.
okay if the problem is āthe thing i made is being used in a way i dont like,ā thatās still separate from theft/copyright infringement.
āItās fine if a person does itā is a fantastic argument. Saying that itās ok to allow robots to continue to replace every part of human life will only lead to suffering for literally everything in existence. Computers can destroy and create in milliseconds what might take humans a lifetime to achieve. If this isnāt an incredibly good reason to regulate the shit out of ai then what is?!?!?
Like yes, currently generative AI use is incredibly difficult to get something non-derivative, e.g. using it as a tool like Photoshop. But that most definitely will not be the case in a few years. This is by far the steepest, slipperiest, most ridiculous slope we could be on and itās not even close.
This is the biggest problem with technology, regulation is reactionary and not preemptory. Not taking action immediately has already gotten earth into a ridiculously bad situation. If we continue to allow it itās only going to get worse and harder to undo.
so its harrison bergeron then? ai must be limited to be equal with humans?
AI images try to replicate the style of popular artists by using their work, often including work that was behind a paywall and taken without payment, thus denying the artists revenue. AI has taken something from the artist, and cost the artist money. Until such a time as we come up with a new word for this new crime, weāll call it by the closest equivalent: theft.
Also, someone did an experiment and typed āmovie screenshotā into an AI and it came back with a nearly identical image from Endgame. Not transformative enough to be anything but copyright infringement.
I think libel is a good word to start using
Defining ātaken something behind a paywall and thus denied them revenueā as theft is the exact same argument movie studios make when you pirate a movie. Theft implies that the original is gone. If I steal your car, you donāt have a car. If I pirate a movie, we both have that movie. As someone who supports piracy, I would be careful to conflate piracy with theft. I think thatās the entire point the post is making.
Fuck AI slop. Thereās enough other arguments against it. It destroys the environment and artistsā livelihoods. We can point that out without supporting corporate copyright talking points.
ā¦thus denying the artists revenue.
This assumes they would otherwise pay for it, and that they measurably harmed the artistās revenue. Those arenāt a given.
Until such a time as we come up with a new wordā¦
Use of copyrighted material without permission and possible deprivation of revenue. It doesnāt need to be a single word.
AI images try to replicate the style of popular artists by using their work, often including work that was behind a paywall and taken without payment, thus denying the artists revenue. AI has taken something from the artist, and cost the artist money. Until such a time as we come up with a new word for this new crime, weāll call it by the closest equivalent: theft.
Iād argue itās much closer to piracy or freebooting. Generally, its use doesnāt hurt artists, seeing as a random user isnāt going to spend hundreds or thousands to hire a talented artist to create shitposts for them. Doesnāt necessary make it okay, but it also doesnāt directly hurt anyone. In cases of significant commercial use, or copyright infringement, Iād argue its closer to freebooting: copying anotherās work, and using it for revenue without technically directly damaging the original. Both of these are crimes, but both are more directly comparable and less severe than actual theft, seeing as the artist loses nothing.
Also, someone did an experiment and typed āmovie screenshotā into an AI and it came back with a nearly identical image from Endgame. Not transformative enough to be anything but copyright infringement.
Copyrighted material is fed into an AI as part of how it works. This doesnāt mean than anything that comes out of it is or is not copyrighted. Copyrighted matterial is also used in Photoshop, for example, but as long as you donāt use Photoshop to infringe on somsone elseās copyright, there isnāt anything intrinsically wrong with Photoshopās output.
Now, if your compaint is that much of the training data is pirated or infringes on the licensing its released under, thats another matter. Endgame isnāt a great example, given that it can likely be bought with standard copyright limitations, and ignoring that, its entirely possible Disney has been paid for their data. We do know huge amounts of smaller artists have had their work pirated to train AI, though, and because of the broken nature of our copyright system, they have no recourse - not through the fault of AI, but corrupt, protectionist governments.
All that said, theres still plenty of reasons to hate AI (and esspecially AI companies) but I donāt think the derivative nature of the work is the primary issue. Not when theyāre burning down the planet, flooding our media with propaganda, and bribing goverments, just to create derivative, acceptable-at-best āāāartāāā. Saying AI is the problem is an oversimplification - we canāt just ban AI to solve this. Instead, we need to address the problematic nature of our copyright laws, legal system, and governments.
No, it is theft. They use an artistās work to make an image they would otherwise pay the artist to make (a worse version, but still). And given how Iāve seen an image with a deformed patreon logo in the corner, they didnāt pay what they should have for the images. They stole a commission.
And it is copyright violation. There have been successful lawsuits over much less than a direct image of RDJ in the iron man suit with the infinity stones on his hand. And if they wonāt pay an artistās rates, thereās no way theyād pay whatever Disney would charge them
Yes, thereās a lot of problems with AI. And yes, AI is a part of larger issues. That doesnāt mean theft isnāt also an issue with AI.
AI is a nazi-built, kitten blood-powered puppy kicking machine built from stolen ambulance parts. Even if stealing those ambulance parts is a lesser sin than killing those kittens, itās still a problem that needs to be fixed. Of course, AI will never be good, so we need to get rid of the whole damn thing.
No, it is theft. They use an artistās work to make an image they would otherwise pay the artist to make (a worse version, but still). And given how Iāve seen an image with a deformed patreon logo in the corner, they didnāt pay what they should have for the images. They stole a commission.
But were they (the AI users) going to pay for the content? I have never paid for a Patreon, given that I donāt really have any disposable income. Why would I start, just because AI exists? Just because a sale may be made in some contexts, doesnāt mean it has been made.
And it is copyright violation. There have been successful lawsuits over much less than a direct image of RDJ in the iron man suit with the infinity stones on his hand.
Its a copyright violation when material is made that violates existing copyright. It isnāt copyright infringement to take data from media, or to create derivative works.
And if they wonāt pay an artistās rates, thereās no way theyād pay whatever Disney would charge them
Disney has lawers. Small artists donāt.
AI is a nazi-built, kitten blood-powered puppy kicking machine built from stolen ambulance parts. Even if stealing those ambulance parts is a lesser sin than killing those kittens, itās still a problem that needs to be fixed. Of course, AI will never be good, so we need to get rid of the whole damn thing.
Banning AI doesnāt stop the Nazis from running the government or influencing the populus, it doesnāt stop them burning the planet, it doesnāt stop them from pirating work and otherwise exploiting artists. Hell, politicians have been doing all of these things without repercussions for a century. If you want the rich and powerful to stop pirating and freebooting artistās work, maybe the first step is to ban that (or rather, enforce it) rather than a technology two steps removed?
But weāre AI users going to pay?
In your head is AI being used solely by common people for fun little prompts? If you build this machine that replaces the artist, corporations can and will use it that way.
Big movie studios will use it to generate parts (and eventually all) of a movie. They can use this as leverage to pay the artists less and hire fewer of them. Animators, actors, voice actors.
you want the rich and powerful to stop pirating and freebooting artistās work, maybe the first step is to ban that (or rather, enforce it) rather than a technology two steps removed?
If a movie studio pirated work and used it in a film, thatās against copyright and we could sue them under current law.
But if they are paying openAI for a service, and it uses copyrighted material, since openAI did the stealing and not the studio then itās not clear if we can sue the studio.Logically we would pursue openAI then, but youāre arguing that we shouldnāt because itās ātwo steps removedā.
Seems like itās being argued that because of the layer of abstraction that is created when large quantities of media is used, rather than an individualās work, that itās suddenly a victimless crime. That because whatās being done is not currently illegal it must not be immoral either.
Big movie studios will use it to generate parts (and eventually all) of a movie. They can use this as leverage to pay the artists less and hire fewer of them. Animators, actors, voice actors.
Only if its profitable, and given that AI output is inherently very limited, it wonāt be. AI can only produce lower quality, derivative works. In isolation, some works might not be easy to distinguish, but thats only on a small scale and in isolation.
If a movie studio pirated work and used it in a film, thatās against copyright and we could sue them under current law.
But if they are paying openAI for a service, and it uses copyrighted material, since openAI did the stealing and not the studio then itās not clear if we can sue the studio.You can sue the studio. In the same way, you would sue the studio if an artist working there (or even someone directing artists) creates something the violates copyright, even by accedent. If they publish a work that infringes on copyright, you can sue them.
Seems like itās being argued that because of the layer of abstraction that is created when large quantities of media is used, rather than an individualās work, that itās suddenly a victimless crime.
By that logic, anything that takes inspiration, no matter now broad, or uses anothers work in any way, no matter how transformative, should be prevented from making their own work. That is my point. AI is just an algorithm to take thousands of images and blends them together. It isnāt evil, any more than a paint brush is. What is, is piracy for commercial use, and non-transformative copyright infringement. Both of these are already illegal, but artists canāt do anything about it, not because companies havenāt broken the law, but rather because an independent author trying to take, for example, Meta to court is going to bankrupt themselves.
Edit: Also notable in companies using/not using AI, is the fact that even transformative and āāāoriginalāāā AI work cannot be copyrighted. If Disney makes a movie thats largely AI, we can just share it freely without paying them.
Only if its profitable, and given that Al output is inherently very limited, it wonāt be. Al can only produce lower quality
This is literally already happening. The SAGAFTRA screen actors guild had to negotiate new contracts against studios that were using AI as a bargaining chip to lower their wages
It isnāt current AI voice tech that was an issue. It was the potential for future AI they were worried about. AI voices as they are now, are of similar quality to pulling someone off the street and putting them in front of a mid-range mic. If you care about quality at all, (without massive changes to how AI tech functions) youāll always need a human.
And to be clear, what about AI makes it the problem, rather than copyright? If I can use a voice synthesizer to replicate an actors voice, why is that fine and AI not? Should it not be that reproduction of an actorās voice is right or wrong based on why its done and its implications rather than because of the technology used to replicate it?
Edit: And to be clear, just because a company can use it as an excuse to lower wages, doesnāt mean its a viable alternative to hiring workers. Claims that they could replace their workers with AI is just the usual capitalist bullshit excuses to exploit their workers.
Yep. I really hope that the conversation around LLMs moves away from words like ātheftā. Show me evidence of an artwork that an LLM has concealed artwork from the public. Iāve looked, and all that Iāve found is that they make more media more accessible. Thatās not theft. Thatās piracy. Thatās culture-jamming.
So if you want to call it appropriation, fine. Itās classic EEE methods, applied to Gonzo, Dada & Punk ideas. Embrace - taking in everything and culture-jamming with meaningless text & images (painfully Dada). Extend - by turning this into both a toy and a corpo ātoolā, they extended Dada into programming, articles, news media, you name it. Extinguish - when everything is a punk remix, or everything is meaningless Dada, nothing is. Therefore the true punks will be the classicist, reconstructionists, the Bible-beaters, and their ilk. And then Punk is dead.
āAI slop is punkā has the ring of when brew dog claimed they were doing a āpunkā equity offering. š
Vibe coding is dadaist. If you say so, but I figure intent mattersā¦
Intention matters or even, intention is all that matters here. Otherwise a random number generator is dadaist.
RNG only for the sake of RNG can be said to be Dada. The appearance or simulation of intent without practical purposes fits the aesthetic well. The moment it actually serves a practical purpose, it ceases to be. TempleOSā Oracle and the āparticle effectā in Super Mario World are not Dada, but just outputting dev/random could be.
Simulation has merged with the thing being simulated and the simulacra take the originalsā places. Nonsense has become the only sense.
Intent matters from the individuals who built the model and trained it. Not from those being jammed. Vibe coders are just tools the artist and punk use to advance the work and the impact.
The world needs more people like you, and I would love hearing more about all this from you š!
I agree with the comment here that AI image generation is more like piracy in that you are appropriating other artistsā works without their permission.
So I mean personally I agree that AI art is soulless and possibly copyright infringement but not theft, whereas for people that consider piracy theft, calling AI art ātheftā is not an inconsistent or hypocritical argument, in my opinion. Machine or human doesnāt make it stealing or not.
I mean even Zuckās Meta is claiming in court that they arenāt stealing books when torrenting massive amounts of them for AI training, lol, so heās consistent there. But Nintendo on the other hand are probably seething at AI Nintendo art being āstolenā from them.
Yeah, I donāt agree. Unfortunately Iām not articulate enough to explain why I feel this way. I feel like they are glossing over things. How would you describe corporations willfully taking art/data/content form others without any permission, attribution, or payment and creating a tool with said information for the end goal of making profits by leveraging the work of others into a derivative work that completes with the original?(Holy run on sentence) If there is a better word or term than theft for what generative ai does then they should use it instead.
I guess what feels off to me is that the generative AI itself does nothing of the sort; the corporations creating the product AI models do. There are already attempts to make generative AI models that are trained exclusively on data that was licensed for it. I imagine some people would still like to push regulation against companies producing those models, though I am not one of them. Iād like to decouple the arguments of āthis use of technology is bad, because it (devalues human works / takes away jobs / ā¦)ā from āthe corporations train their generative AI models in an unethical mannerā.
Itās basically for-profit piracy. Which is still kind of a shitty term because actual pirates werenāt copying any of the goods they were taking.
The most neutral term might be copyright infringement, though that carries all the baggage of the āshould copyright even existā-discussion.
Alternatively, you could shout āthey took our jobsā to complain that they are letting algorithms and engineers do the work that artists want to do. IDK what to call this, but ātheftā or ārobberyā doesnāt sound right.
I think the biggest problem is that the idea of copyright is good, but the implementation - in most places, anyways - is complete dogshit.
Like, Iām fairly certain the original implementation of copyright in the US only lasted 10 years or thereabouts. Like, thatās more than enough time to profit off whatever you made but short enough that itāll be usable by others within their lifetimes. This whole ālife of the author + 100 yearsā shit needs to die.
i get what you mean, but at the same time, i feel like thereās not much you can do with AI that you canāt do without it (at least, in terms of art), it just takes more time and youād most likely need to pay someone to do it. so in this case AI would take work opportunities away from people, thatās bad, but thatās not copyright infringement nor theft.
and iām worried that by propagating the idea that training an AI model is theft, that it could lead to even more regulation on copyright that would just end up hurting smaller artists in the end. like, most people agree that making AI art in the style of Studio Ghibli movies is scummy, but would an indie artist making art in that style be wrong too? you may think not, but if it becomes a crime to emulate an art style using AI, it takes very little extrapolation to make it a crime to emulate an art style without AI. and do i need to say why being able to copyright an art style would be awful?
so in this case AI would take work opportunities away from people, thatās bad, but thatās not copyright infringement nor theft.
I think itās quite literally copyright infringement, assuming the models are fed with work from actual artists who typically donāt agree to it. Whether copyright should work this way is another matter.
I meant it more in a āhire an artist to work on art for you VS just ask the AI to do it insteadā way
even if youāre emulating an art style in particular, thatās not copyright infringement because you canāt copyright an art style. which is good because if you could, that would be awful for a ton of artists
itās only copyright infringement if you ask an AI to do, say, a picture of mario. but in this case, itās also copyright infringement if you commission an artist to do it!
I suppose the issue is whether you want to see training an AI as equivalent to practicing as a human artist. Considering that AIs are generally made specifically as commercial products, I donāt think thatās really true, but this is definitely something that can be argued one way or the other.
I donāt think it would be considered fine under current laws if an AI-free Adobe Photoshop was shipped with tons of copyrighted art that was scraped off the internet without the artistsā approval, even if the users arenāt allowed to use it to make commercial works that reproduce Super Mario or w/e 1:1.
itās only copyright infringement if you ask an AI to do, say, a picture of mario. but in this case, itās also copyright infringement if you commission an artist to do it!
Nah.
-
Training a statistical model with unlicensed work is not the same as a human learning.
-
Under copyright/IP laws, using a copyrighted work, without license, with the intent of competing with the copyright holder (what virtually all commercial AI models are doing) is not fair use and there is plenty of case law backing that. Whether something is transformative (arguably, training models isnāt) doesnāt even matter if the infringement is done with the intent of causing material harm to the copyright holder through competition. None of the models out there fit the criteria for fair use.
-
Something being a tool does not magically remove all liability. This is especially true if the tool is built illegally using unlicensed intellectual property and depends on said unlicensed intellectual property to have any value (literally all major models fit this description).
-
Lemmyās respect for copyright only in relation to the magic content robot is endlessly amusing.
I donāt give a shit what public data gets shredded into a gigabyte of linear algebra. That process is transformative. If the result is any good at reproducing a specific input, you did it wrong.
āI believe in you. You can come up with a better argument than just theft.ā
Nah, fuck that shit. It OOP feels so strongly that itās not theft and they wanna change how the population at large is referring to something, then itās on them to provide an alternative and convince others. This weird ass attempt to shame people into doing things their way, especially when they havenāt really defined what they consider their way, is absolute horse shit.
This whole post is full of this. The OOP tries to completely remove intent and method from the analysis of whether something is art theft. Those things absolutely factor into it and theyāre only discounting them in order to push their weird narrative.
AI scrapping tons of work belonging to artists and then regurgitating that as original work is fucking gross, no matter what you call it. Theft seems fine to me, but I am open to calling it something else. Unfortunately OOP wonāt be the Iāve to convince me since they neither provide reasoning for why calling it theft is bad or what we should call it instead and why.
The alternative is to call out copywrong in every version and every facet of existence. This isnāt theft, itās duplication. The argument is simple: LLMs are the new printing press.
Disclaimer: Iāve not workshopped this much, so idk if these are the right words to convey how I feel
I feel like using AI to generate images is akin to taking someoneās art and applying a light gaussian blur to it or putting an uncredited artistās work in a big game.
I know itās done in a much more intricate way, and itās genuinely impressive how AI companies got it to work so well, but if I try to sell AI generated images, especially if theyāre meant to be made similar to an artistās work, then thatās all Iām doing.
I donāt necessarily see it as stealing from artists (though it is threatening the livelihood of a lot of artists), but more as exploiting artists but with a new buzzword.
If I arrange 4 pieces of art in a jpeg and then apply a whacky filter, am I actually creating anything, or am I just exploiting artists and doing something similar to copying and pasting different bits of an essay and then changing every instance of a word to a different synonym?
I believe AI does something similar to that, albeit in a more sophisticated way that looks like creativity.