AI Image Training Dataset Contaminated with Child Sexual Abuse Images

Stanford researchers have discovered over 1,000 images of child sexual abuse in a popular open-source database used to train AI image-generating models. The presence of these illegal images raises concerns that AI tools may be learning to create hyper-realistic fake images of child exploitation. AI image generators have been increasingly promoted on pedophile forums, enabling the creation of uncensored explicit images of children. The inclusion of child abuse photos in training data allows AI models to produce content resembling real-life child exploitation. The researchers suggest implementing protocols to screen and remove abusive content from databases, increasing transparency in training data sets, and teaching image models to "forget" how to create explicit imagery. The illegal images are in the process of being removed from the training database.
- Child sexual abuse images have been used to train AI image generators The Washington Post
- AI image training dataset found to include child sexual abuse imagery The Verge
- Stanford study finds child abuse images in AI training data Axios
- Large AI Dataset Has Over 1000 Child Abuse Images, Researchers Find Bloomberg
- AI Image Dataset is Pulled After Child Sex Abuse Pictures Discovered PetaPixel
Reading Insights
0
1
2 min
vs 3 min read
78%
579 → 126 words
Want the full story? Read the original article
Read on The Washington Post