Stories
Slash Boxes
Comments

SoylentNews is people

SoylentNews is powered by your submissions, so send in your scoop. Only 14 submissions in the queue.
posted by hubie on Saturday September 09 2023, @01:42AM   Printer-friendly
from the think-of-the-AI-generated-children dept.

https://arstechnica.com/information-technology/2023/09/ai-generated-child-sex-imagery-has-every-us-attorney-general-calling-for-action/

On Wednesday, American attorneys general from all 50 states and four territories sent a letter to Congress urging lawmakers to establish an expert commission to study how generative AI can be used to exploit children through child sexual abuse material (CSAM). They also call for expanding existing laws against CSAM to explicitly cover AI-generated materials.

"As Attorneys General of our respective States and territories, we have a deep and grave concern for the safety of the children within our respective jurisdictions," the letter reads. "And while Internet crimes against children are already being actively prosecuted, we are concerned that AI is creating a new frontier for abuse that makes such prosecution more difficult."

In particular, open source image synthesis technologies such as Stable Diffusion allow the creation of AI-generated pornography with ease, and a large community has formed around tools and add-ons that enhance this ability. Since these AI models are openly available and often run locally, there are sometimes no guardrails preventing someone from creating sexualized images of children, and that has rung alarm bells among the nation's top prosecutors. (It's worth noting that Midjourney, DALL-E, and Adobe Firefly all have built-in filters that bar the creation of pornographic content.)

"Creating these images is easier than ever," the letter reads, "as anyone can download the AI tools to their computer and create images by simply typing in a short description of what the user wants to see. And because many of these AI tools are 'open source,' the tools can be run in an unrestricted and unpoliced way."

As we have previously covered, it has also become relatively easy to create AI-generated deepfakes of people without their consent using social media photos.


Original Submission

 
This discussion was created by hubie (1068) for logged-in users only, but now has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 2) by istartedi on Tuesday September 12 2023, @03:57PM (1 child)

    by istartedi (123) on Tuesday September 12 2023, @03:57PM (#1324234) Journal

    False dichotomy? You can teach children to avoid suspicious strangers *and* teach children what's appropriate from adults they know *and* teach school staff what signs to look for.

    --
    Appended to the end of comments you post. Max: 120 chars.
    Starting Score:    1  point
    Karma-Bonus Modifier   +1  

    Total Score:   2  
  • (Score: 2) by TheReaperD on Sunday September 17 2023, @09:01AM

    by TheReaperD (5556) on Sunday September 17 2023, @09:01AM (#1325016)

    That would be great, if it were possible. Sadly, at the national policy level, if you expect people to learn a multi-prong nuanced policy, everyone will ignore you and do what some idiot on YouTube or Ticktock says instead. Even when you keep policies simple and focused to a couple of main points, you may still lose out to these people! But, to have any chance of getting people to listen to you, you have to keep it stupid simple and the main points can be remembered in a sentence or two. So, you have to pick your battles and choose the topics that will do the most good for the masses, even though you want to do more. It sucks, but that's reality. So, when I'm talking about points like this, it's not that I wouldn't be for dealing with it all, I'm just realistic to know that we could never make it work as a national policy, as vexing as that is.

    --
    Ad eundum quo nemo ante iit