It seems crazy to me but ive seen this concept floated on several different post. There seems to be a number of users here that think there is some way AI generated CSAM will reduce Real life child victims.

Like the comments on this post here.

https://sh.itjust.works/post/6220815

I find this argument crazy. I don’t even know where to begin to talk about how many ways this will go wrong.

My views ( which are apprently not based in fact) are that AI CSAM is not really that different than “Actual” CSAM. It will still cause harm when viewing. And is still based in the further victimization of the children involved.

Further the ( ridiculous) idea that making it legal will some how reduce the number of predators by giving predators an outlet that doesnt involve real living victims, completely ignores the reality of the how AI Content is created.

Some have compared pedophilia and child sexual assault to a drug addiction. Which is dubious at best. And pretty offensive imo.

Using drugs has no inherent victim. And it is not predatory.

I could go on but im not an expert or a social worker of any kind.

Can anyone link me articles talking about this?

  • Surdon@lemm.ee
    link
    fedilink
    arrow-up
    12
    arrow-down
    3
    ·
    1 year ago

    Except there is a good bit of evidence to show that consuming porn is actively changing how we behave related to sex. By creating CSAM by AI, you create the depiction of a child that is mere object for the use of sexual gratification. That fosters a lack of empathy and an ego centric, self gratifying viewpoint. I think that can be said of all porn, honestly. The more I learn about what porn does to our brains the more problematic I see it

    • Killing_Spark@feddit.de
      link
      fedilink
      arrow-up
      6
      arrow-down
      2
      ·
      edit-2
      1 year ago

      I agree with this.

      The more I learn about what porn does to our brains the more problematic I see it

      And I agree with this especially. Turns out a brain that was/is at least in part there to get us to procreate isn’t meant to get this itch scratched 24/7.

      But to answer your concern: I will draw another comparison with addiction: Giving addicitive drugs out like candy isn’t wise just as it wouldn’t be wise to give access to generated csam to everyone. You’d need a control mechanism so that only people that need access get access. Admitedly this will deter a few people from getting their fix from the controlled instances compared to the completely free access. With drugs this seems to lead to a decrease of the amount of street-sold drugs though, so I see no reason this wouldn’t be true, at least to some extent, for csam.

      • Surdon@lemm.ee
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        I’m an advocate of safe injection sites, so I will agree somewhat here. Safe injection sites work because they identify addicts and aggressively supply them with resources to counteract the need for the addiction in the first place, all while encouraging less and less use. This is an approach that could have merit for pedophiles, but there are some issues that pop up with it as well that are unique- to consume a drug, the drug must enter the body somehow, where it is metabolized.

        CSAM on the other hand, is taken in simply by looking at it. There is no “gloves on” approach to generating or handing the content without absorbing it- the best that can be hoped for is have it generated by someone completely ‘immune’ to it, which raises questions about how “sexy” they could make the content- if it doesn’t “scratch the itch” the addicts will simply turn back to the real stuff.

        There is a slim argument to be made that you could actually create MORE pedophiles through classical conditioning by exposing nonpedophilic people to erotic content paired with what looks like children. You could of course have it produced and handled by recovering/in treatment pedophiles, but that sounds like it defeats the point of limited access entirely and is therefore still bad, at least to the ones in charge of distribution.

        Additionally, digital content isn’t destroyed upon consumption like a drug, and you have a more minor but still real problem of content diversion, where content made for the program is spread to those not getting the help that was meant to be paired with it. This is an issue, of course, but could be rationalized as worth it so long as at least some pedophiles were being treated.

        • Killing_Spark@feddit.de
          link
          fedilink
          arrow-up
          3
          ·
          edit-2
          1 year ago

          Yes there are a lot of open questions around this, especially about the who and how of generation, and tbh it makes me a bit uncomfortable to think about a system like this in detail, because it will have to include rating these materials on a “sexyness” scale which feels revolting.

    • Moira_Mayhem@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      10 months ago

      AI CSAM will not create new pedophiles, but it may keep existing pedophiles from encouraging a disgusting market of child exploiters.