Content-Length: 10386 | pFad | http://www.ic3.gov/PSA/2024/PSA240329

s Internet Crime Complaint Center (IC3) | Child Sexual Abuse Material Created by Generative AI and Similar Online Tools is Illegal

Alert Number: I-032924-PSA

Child Sexual Abuse Material Created by Generative AI and Similar Online Tools is Illegal


The FBI is warning the public that child sexual abuse material (CSAM) created with content manipulation technologies, to include generative artificial intelligence (AI), is illegal. Federal law prohibits the production, advertisement, transportation, distribution, receipt, sale, access with intent to view, and possession of any CSAM,1 including realistic computer-generated images.

Background

Individuals have been known to use content manipulation technologies and services to create sexually explicit photos and videos that appear true-to-life. One such technology is generative AI, which can create content — including text, images, audio, or video — with prompts by a user. Generative AI models create responses using sophisticated machine learning algorithms and statistical models that are trained often on open-source information, such as text and images from the internet. Generative AI models learn patterns and relationships from massive amounts of data, which enables them to generate new content that may be similar, but not identical, to the underlying training data. Recent advances in generative AI have led to expansive research and development as well as widespread accessibility, and now even the least technical users can generate realistic artwork, images, and videos — including CSAM — from text prompts.

Examples

Recent cases involving individuals having altered images into CSAM include a child psychiatrist and a convicted sex offender:

  • In November 2023, a child psychiatrist in Charlotte, North Carolina, was sentenced to 40 years in prison, followed by 30 years of supervised release, for sexual exploitation of a minor and using AI to create CSAM images of minors. Regarding the use of AI, the evidence showed the psychiatrist used a web-based AI application to alter images of actual, clothed minors into CSAM2.
  • In November 2023, a federal jury convicted a Pittsburgh, Pennsylvania registered sex offender of possessing modified CSAM of child celebrities. The Pittsburgh man possessed pictures that digitally superimposed the faces of child actors onto nude bodies and bodies engaged in sex acts3.

There are also incidents of teenagers using AI technology to create CSAM by altering ordinary clothed pictures of their classmates to make them appear nude.

Recommendations

  • For more information on altered images, see the FBI June 2023 PSA titled "Malicious Actors Manipulating Photos and Videos to Create Explicit Content and Sextortion Schemes" at https://www.ic3.gov/Media/Y2023/PSA230605.
  • If you are aware of CSAM production, including AI generated material, please report it to the following:
    1. National Center for Missing and Exploited Children [1-800-THE LOST or www.cybertipline.org]
    2. FBI Internet Crime Complaint Center [www.ic3.gov]

References










ApplySandwichStrip

pFad - (p)hone/(F)rame/(a)nonymizer/(d)eclutterfier!      Saves Data!


--- a PPN by Garber Painting Akron. With Image Size Reduction included!

Fetched URL: http://www.ic3.gov/PSA/2024/PSA240329

Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy