Riskable

Father, Hacker (Information Security Professional), Open Source Software Developer, Inventor, and 3D printing enthusiast

  • 38 Posts
  • 1.43K Comments
Joined 2 years ago
cake
Cake day: June 23rd, 2023

help-circle




  • If you studied loads of classic art then started making your own would that be a derivative work? Because that’s how AI works.

    The presence of watermarks in output images is just a side effect of the prompt and its similarity to training data. If you ask for a picture of an Olympic swimmer wearing a purple bathing suit and it turns out that only a hundred or so images in the training match that sort of image–and most of them included a watermark–you can end up with a kinda-sorta similar watermark in the output.

    It is absolutely 100% evidence that they used watermarked images in their training. Is that a problem, though? I wouldn’t think so since they’re not distributing those exact images. Just images that are “kinda sorta” similar.

    If you try to get an AI to output an image that matches someone else’s image nearly exactly… is that the fault of the AI or the end user, specifically asking for something that would violate another’s copyright (with a derivative work)?




  • I wasn’t being pedantic. It’s a very fucking important distinction.

    If you want to say “unethical” you say that. Law is an orthogonal concept to ethics. As anyone who’s studied the history of racism and sexism would understand.

    Furthermore, it’s not clear that what Meta did actually was unethical. Ethics is all about how human behavior impacts other humans (or other animals). If a behavior has a direct negative impact that’s considered unethical. If it has no impact or positive impact that’s an ethical behavior.

    What impact did OpenAI, Meta, et al have when they downloaded these copyrighted works? They were not read by humans–they were read by machines.

    From an ethics standpoint that behavior is moot. It’s the ethical equivalent of trying to measure the environmental impact of a bit traveling across a wire. You can go deep down the rabbit hole and calculate the damage caused by mining copper and laying cables but that’s largely a waste of time because it completely loses the narrative that copying a billion books/images/whatever into a machine somehow negatively impacts humans.

    It is not the copying of this information that matters. It’s the impact of the technologies they’re creating with it!

    That’s why I think it’s very important to point out that copyright violation isn’t the problem in these threads. It’s a path that leads nowhere.




  • This image has me thinking about what a post-apocalyptic world would look like if the society weren’t industrial but instead went with the fantasy, “we live in custom-grown enormous trees” sort of environment. Would it even look post-apocalyptic? 🤔

    Maybe magic-reinforced stuff would still exist but be covered in weeds/vines. Like a really overgrown suspension bridge that looks sketchy AF but is actually super sturdy because of the enchantment.





  • Yeah it’s probably just a client side issue but the OP mentioned Element, specifically 🤷

    I just wanted to point out that Element is no fun! No fun at all!

    It works and it works great for what it does. Even voice and streaming are great with Element. It’s just got a terrible, no-fun interface and pointless limitations on things like looping videos. You can’t even configure it to make them play properly (as in, automatic and endlessly, the way they were meant to be played! 😤).

    Looping videos and animated emojis are super fun ways to chat with people. Even in professional settings! It really breaks up the humdrum and can motivate people to chat and share more.

    Element is all serious all the time and going into a chat channel there feels like a chore.


  • The person making claims is the one that needs to provide the evidence. Anyone spreading unverified rumors should be viewed like The Boy Who Cried Wolf.

    In order for a rumor or something like naked pictures to be effective in its purpose of bringing you down or blackmailing you it must cause emotional distress. If you don’t take it seriously–or even better, don’t give a shit about what random strangers say–it can’t have power over you.

    We don’t live in tiny, closely-tied villages anymore. A rumor about you may be on the Internet forever but it’ll be about as believable as a “hot singles in your area” ad and the likelihood of someone even finding it goes down over time as it gets drowned out by all the other bullshit that exists.

    Even celebrities that are famous for doing stupid shit are finding that nobody remembers their past failings anymore. And they’ll have hundreds to thousands of real news articles about their idiocy out there, ready for anyone to read.




  • They’re not illegally harvesting anything. Copyright law is all about distribution. As much as everyone loves to think that when you copy something without permission you’re breaking the law the truth is that you’re not. It’s only when you distribute said copy that you’re breaking the law (aka violating copyright).

    All those old school notices (e.g. “FBI Warning”) are 100% bullshit. Same for the warning the NFL spits out before games. You absolutely can record it! You just can’t share it (or show it to more than a handful of people but that’s a different set of laws regarding broadcasting).

    I download AI (image generation) models all the time. They range in size from 2GB to 12GB. You cannot fit the petabytes of data they used to train the model into that space. No compression algorithm is that good.

    The same is true for LLM, RVC (audio models) and similar models/checkpoints. I mean, think about it: If AI is illegally distributing millions of copyrighted works to end users they’d have to be including it all in those files somehow.

    Instead of thinking of an AI model like a collection of copyrighted works think of it more like a rough sketch of a mashup of copyrighted works. Like if you asked a person to make a Godzilla-themed My Little Pony and what you got was that person’s interpretation of what Godzilla combined with MLP would look like. Every artist would draw it differently. Every author would describe it differently. Every voice actor would voice it differently.

    Those differences are the equivalent of the random seed provided to AI models. If you throw something at a random number generator enough times you could–in theory–get the works of Shakespeare. Especially if you ask it to write something just like Shakespeare. However, that doesn’t meant the AI model literally copied his works. It’s just doing it’s best guess (it’s literally guessing! That’s how work!).