AI Image Training Dataset Contaminated with Child Sexual Abuse Images

1 min read
Source: The Washington Post
AI Image Training Dataset Contaminated with Child Sexual Abuse Images
Photo: The Washington Post
TL;DR Summary

Stanford researchers have discovered over 1,000 images of child sexual abuse in a popular open-source database used to train AI image-generating models. The presence of these illegal images raises concerns that AI tools may be learning to create hyper-realistic fake images of child exploitation. AI image generators have been increasingly promoted on pedophile forums, enabling the creation of uncensored explicit images of children. The inclusion of child abuse photos in training data allows AI models to produce content resembling real-life child exploitation. The researchers suggest implementing protocols to screen and remove abusive content from databases, increasing transparency in training data sets, and teaching image models to "forget" how to create explicit imagery. The illegal images are in the process of being removed from the training database.

Share this article

Reading Insights

Total Reads

0

Unique Readers

1

Time Saved

2 min

vs 3 min read

Condensed

78%

579126 words

Want the full story? Read the original article

Read on The Washington Post