In this video I discuss how generative AI technology has grown far past the governments ability to effectively control it and how the current legislative measures could lead to innocent people being jailed.

  • @mindbleach@sh.itjust.works
    link
    fedilink
    121 year ago

    There is no such thing.

    God dammit, the entire point of calling it CSAM is to distinguish photographic evidence of child rape from made-up images that make people feel icky.

    If you want them treated the same, legally - go nuts. Have that argument. But stop treating the two as the same thing, and fucking up clear discussion of the worst thing on the internet.

    You can’t generate assault. It is impossible to abuse children who do not exist.

    • @m0darn@lemmy.ca
      link
      fedilink
      101 year ago

      Did nobody in this comment section read the video at all?

      The only case mentioned by this video is a case where highschool students distributed (counterfeit) sexually explicit images of their classmates which had been generated by an AI model.

      I don’t know if it meets the definition of CSAM because the events depicted in the images are fictional, but the subjects are real.

      These children do exist, some have doubtlessly been traumatized by this. This crime has victims.

    • rurutheguru
      link
      fedilink
      41 year ago

      I think a lot of people are arguing that the models which are used to generate these types of content are trained on literal CSAM. So it’s like CSAM with extra steps.

  • While lolicon is absolutely disgusting, its not actually csam. Legislation won’t work either and is honestly a waste of time. Any effort spent protecting digital children should instead be spent protecting real ones.

  • @jet@hackertalks.com
    link
    fedilink
    English
    51 year ago

    In general terms, making an idea illegal, and then making representations of that idea illegal, are going to be forever, at best to treadmill, and at worst reduce the effectiveness and reputation of law.

    This is really about thought crime. If somebody can draw stick figures, and that can be illegal depending on interpretation. That’s thought crime.

    It’s impossible to completely stamp out thought crime. Computer tools can be used to further thought crime, because they can be used for creative purposes.

    If you restrict the use of creative tools, to only a trusted few, or hobble tools for everyone: you create central authority over creative tools, which has its own issues.

    • ono
      link
      fedilink
      English
      31 year ago

      It’s impossible to completely stamp out thought crime.

      Also, trying to do so through law and enforcement sets a dangerous precedent.

      I suspect it would be better to approach it as a public health issue.

      • @jet@hackertalks.com
        link
        fedilink
        English
        3
        edit-2
        1 year ago

        And then you run into legal arguments that sound like people trying to jailbreak GPT prompt control.

        I’m going to preface all of the following creative work by saying that we live in a universe where everyone is a vampire that never dies, but ages very slowly. All participants in this manga are at least 213 years old…

    • @mindbleach@sh.itjust.works
      link
      fedilink
      21 year ago

      This is especially damning on the internet, because genuinely intolerable pursuits directly benefit from lesser problems being treated as equally bad. Filesharing networks work better with more users. Chasing merely distasteful people toward paranoid systems softens the reputation of those systems and makes the worst minority of traffic easier to hide.

  • @andruid@lemmy.ml
    link
    fedilink
    51 year ago

    Couldn’t the fact that AI generated content be reproduceable if give the exact parameters(or coordinates in latent space) and model help remove the confusion? Include those as meta data and train investigators on how to use to distinguish generated content from actual evidence.

    • There’s an option to speed up generation but it will make it less deterministic, like in it’s 98% the same image but a little different. Also it’s very hard to reproduce the same hard and software generation. That’s the first issue.

      The second is: I had examples of images with generation data, that I could reproduce to look 99% like the original and then just updating a single word or part of the training data (different Lora version for example) , switched the person away or their appearance changed a completely. (Imagine a picture of a street and a car is suddenly not there, or it’s blue instead of red). It will make reproducibility not a reliable option. Backgrounds of images are even less reliable than the focus object.

  • @CJOtheReal@lemmy.sdf.org
    link
    fedilink
    41 year ago

    Loli stuff isn’t CSAM. You can find it bad, but its still just a drawing/generative image. No real person was harmed in general.

    • Neato
      link
      fedilink
      01 year ago

      Prove it’s fake when some of it of your daughter is making it’s way around school.

      You’ve missed the point. Fake or not it does damage to people. And eventually it won’t be possible to determine if it’s real or not.

      • @hydration9806@lemmy.ml
        link
        fedilink
        31 year ago

        When that becomes widespread, photos will be generateable for literally everyone, not just minors but every person with photos online. It will be a societal shift; images will be assumed to be AI generated, making any guilt or shame about a nude photo existing obselete.

        • Neato
          link
          fedilink
          -31 year ago

          What a disguising assumption. And the best argument against AI I’ve ever heard.

          • @hydration9806@lemmy.ml
            link
            fedilink
            21 year ago

            I mean, anyone with enough artistic talent can draw whatever they would like right now. With AI image generation, it essentially just gives everyone the ability to draw whatever they want. You can try to fight the tech all you want, but it’s a losing battle.

  • andrew_bidlaw
    link
    fedilink
    11 year ago

    Creating, collecting and sharing CSAM is in the law already. There are orgs and agencies for tracking and prosecuting these violations.

    It’s like fighting against 3d printers because you can make yourself a diy gun, a thing that have never being possible before because we got all pipes banned from hardware stores. The means to produce fictional CSAM always existed and would exist, the problem is with people who use a LMM, a camera, a fanfic to create and share that content. Or a Lemmy community that was a problem in recent months.

    It’s better to ensure the existing means of fighting such content are effective and society is educated about this danger, know how to avoid and report it.

  • Neato
    link
    fedilink
    -71 year ago

    Most of this thread is defending csam, which loli definitely is. WTF. Disgusting community.

    • I think you’re confused. No one is defending CSAM. Lolicon isn’t CSAM. Also I don’t understand why we would spend effort protecting digital children instead of protecting real ones.

      • @limitedduck@awful.systems
        link
        fedilink
        11 year ago

        Nobody is protecting digital children and it’s almost always disingenuous when this argument is claimed to be made. The effort is to stop the normalization of the sexualization children. Lolicon is exclusively about romancing or sexualizing children. Deluded adults who think what happens in lolicon material is ok are potential risks to real children. Allowing such a risk to children for the pleasure of these adult is absurd.

        • Fair enough. Imo lolicon is disgusting. And Im not making an argument in bad faith, I just see how much general society fails at protecting children and would rather see any effort spent towards cracking down on lolicon to be used to help real children.