Rep. Joe Morelle, D.-N.Y., appeared with a New Jersey high school victim of nonconsensual sexually explicit deepfakes to discuss a bill stalled in the House.

  • aesthelete@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    edit-2
    11 months ago

    Aguing that since you do a crime with a tool, outlawing the crime outlaws the tool is a bad argument. Outlawing murder doesn’t outlaw knives.

    As far as enforcement, it may be enforced with varying degrees of success but the argument that someone may get away with the crime also isn’t a reason not to make it a crime.

    If someone created deep fakes using locally run models, rubbed one out and then deleted everything they probably wouldn’t be caught…but largely who cares that they didn’t? It’s the harm to others that it causes that you would largely like to prevent, and if a person didn’t distribute the image at all them “getting away with it” doesn’t matter much.

    Edit: I think the argument that existing laws already cover this is more compelling than any of the above arguments as far as why this new law shouldn’t be passed.

    • curiousaur@reddthat.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 months ago

      You conceded that no one cares if someone makes images locally then deletes them. But that’s how they’re all going to be made shortly.

      Currently folks are sharing them because not everyone has the means to create them, some folks do, and share what they’ve made.

      Once litterally every can just make them the moment they want to, no one will be sharing. Everyone will fall under that use case that you admitted no one would care about, which is exactly what I’ve been saying. It’s 1. futile to try to stop, and 2. going to become so wide spread that we as a society will stop caring about it.

      • aesthelete@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        11 months ago

        Once litterally every can just make them the moment they want to, no one will be sharing.

        I do not think this is true. There are reasons to generate and distribute these other than to have a personal wank off gallery.

          • aesthelete@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            11 months ago

            I’m curious as to why you cannot come up with any yourself, but here are a few from the top of my head: to pass them off as authentic (likely for clout purposes), to have a laugh with the boys about it, to collaborate with others on them, and to distribute them to harass, ridicule, or disparage the target of them.

            Degenerates exist in lots of shapes and forms, and not all degenerates will have enough of a sense of shame to be degenerates privately or to even know they are being degenerates at all.

            • curiousaur@reddthat.com
              link
              fedilink
              English
              arrow-up
              1
              ·
              11 months ago

              I don’t think you’re properly understanding the paradigm shift that’s coming with these models being open source and widely available while wearable AR smart glasses get better.

              “You know Sharon is HR, look at this scandalous photo of her.”

              “Uh, I’m seeing a live generated porno of everyone in this room right now, why would I care about that.”

              • aesthelete@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                ·
                edit-2
                11 months ago

                And I don’t think you’re fully understanding that the above is some type of fantasy you have, and will not actually be what the future is like at all.

                • curiousaur@reddthat.com
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  11 months ago

                  It’s probably a bit of an exaggeration, but my point stands. It’s going to be so easy for anyone to see ai gen material of anyone else, no one is going to care anymore.

                  • aesthelete@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    2
                    ·
                    edit-2
                    11 months ago

                    I don’t even think that’s necessarily true. If you make it illegal and/or platforms ban it, you’re already taking a step toward making it more difficult to do.

                    I think throughout this thread you’re mistaking the technically possible for the probable or likely.

                    By making it illegal, you essentially eliminate the commercial incentive for making it easy. Every barrier to doing something makes it more unlikely that people will do it. I understand that there is an inherent motive for people to do it anyway, but, every hoop they have to jump through (e.g. setting up their “own, local AI”) reduces the likelihood of them doing it.

                    People don’t even run their own email servers, music servers, video servers, etc. etc. etc…most people don’t even “jail break” devices…many don’t even store a local cache of regular porn…why the hell would most people bother themselves with setting up a local generative AI instance for this purpose?

                    Outlawing it and banning it from platforms makes it much more within the realm of the creepy basement weirdo rather than something that is as inevitably ubiquitous as you’re saying it will be.

                    Policy is very often about reduction of harms rather than elimination of harms. It’s not the black and white realm that you’re trying to make it out to be.