• rottingleaf@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    15
    ·
    4 hours ago

    Creating and distributing anything should be legal if no real person suffers during its creation and if it’s not intended at defamation, forgery, such things.

    • AstralPath@lemmy.ca
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      1
      ·
      3 hours ago

      Bruh how is creating and distributing a non-consensual nude-ified picture of a young girl not a cause for suffering for the victim? Please, explain that to the class.

      Did you just not go to school as a kid? If so, that would explain your absolute ineptitude on this topic. Your opinion is some real “your body, my choice” kind of energy.

      • krashmo@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 hours ago

        There’s a legitimate discussion to be had about harm reduction here. You’re approaching this topic from an all-or-nothing mindset but there’s quite a bit of research indicating that’s not really how it works in practice. Specifically as it relates to child pornography the argument goes that not allowing artificial material to be created leads to an increase in production of actual child pornography which obviously means more real children are being harmed than would be if other forms were not controlled in the same fashion. The same sort of logic could be applied to revenge porn, stolen selfies, or whatever else we’re calling the kind of thing this article is referring to. It may not be an identical scenario but I still think it would be fair to say that an AI generated image is not as damaging as a real one.

        That is not to say that nothing should be done in these situations. I haven’t decided what I think the right move is given the options in front of us but I think there’s quite a bit more nuance here than your comment would indicate.

        • MagicShel@lemmy.zip
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 hours ago

          I think this is probably a really good point. I have no issue with AI generated images, although obviously if they are used to do an illegal thing such has harassment or defamation, those things are still illegal.

          I’m of two minds when it comes to AI nudes of minors. The first is that if someone wants that and no actual person is harmed, I really don’t care. Let me caveat that here: I suspect there are people out there who, if inundated with fake CP, will then be driven to ideation about actual child abuse. And I think there is real harm done to that person and potentially the children if they go on to enact those fantasies. However I think it needs more data before I am willing to draw a firm conclusion.

          But the second is that a proliferation of AI CP means it will be very difficult to tell fakes from actual child abuse. And for that reason alone, I think it’s important that any distribution of CP, whether real or just realistic, must be illegal. Because at a minimum it wastes resources that could be used to assist actual children and find their abusers.

          So, absent further information, I think whatever a person whats to generate for themselves in private is just fine, but as soon as it starts to be distributed, I think that it must be illegal.

      • rottingleaf@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        3
        ·
        3 hours ago

        Read my comment again.

        Your opinion is some real “your body, my choice” kind of energy.

        My advice to you would be to improve your reading comprehension before judging this way.

        In particular, the word “defamation”.

    • Alphane Moon@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 hours ago

      You would be fine with AI-gen porn images of your teenage daughter being distributed around the internet?

          • rottingleaf@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            4
            ·
            3 hours ago

            Making forged pics of someone else falls under defamation.

            It’s very clearly not rape, sexual abuse, child pornography or non-consensual pornography.

      • droporain@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        11
        ·
        4 hours ago

        Meanwhile in reality check out what she is distributing through Snapchat and only fans… Maybe pursuing the actual crimes first then if there’s spare resources go after fiction.

        • Todd Bonzalez@lemm.ee
          link
          fedilink
          English
          arrow-up
          8
          ·
          3 hours ago

          Big “but what was she wearing?” energy here.

          I don’t give a shit if she’s doing Shein bikini hauls on Youtube. If you use AI to nudify her pictures, you’re manufacturing child pornography, and deserve the full consequences for doing that.

          As for OnlyFans, they are quite strict about age requirements. Children aren’t running OF accounts. You just hate women and needed to bring up OF to slut-shame.

          • rottingleaf@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            3
            ·
            3 hours ago

            If you use AI to nudify her pictures, you’re manufacturing child pornography, and deserve the full consequences for doing that.

            No, equating this to an actual child being raped is incorrect. These are not crimes of remotely equal magnitude.

            Comparing a person who raped a child, made photos and distributed them to a person who used Photoshop or an AI tool is, other than just evil, reducing the meaning of the former.

            • Arkouda@lemmy.ca
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              2 hours ago

              It is weird how hard you have been defending the production of child pornography in this thread.

              • rottingleaf@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                44 minutes ago

                What this conversation is about has as much to do with child pornography as hentai with loli characters.

                You just can’t argue without unsubstantiated accusations, can you?

                When real living people are being murdered and abused in droves, you are still worried more about glorified automated Photoshop and accusing its users of being the same as actual rapists.

                • Arkouda@lemmy.ca
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  4 minutes ago

                  What this conversation is about has as much to do with child pornography as hentai with loli characters.

                  Creating sexually explicit images of minors is child pornography.

                  You just can’t argue without unsubstantiated accusations, can you?

                  You literally confirmed my claim in your first sentence, and your last.

                  When real living people are being murdered and abused in droves, you are still worried more about glorified automated Photoshop and accusing its users of being the same as actual rapists.

                  Production of child pornography is production of child pornography. It does not need to involve rape. Producing child pornography is a separate crime.

                  Its users are pedophiles because they are producing child pornography. You are defending them.

                  These are the facts.