this post was submitted on 06 Jan 2026
595 points (98.1% liked)

Technology

78511 readers
3254 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Allero@lemmy.today 1 points 3 days ago* (last edited 3 days ago) (1 children)

That would be true if children were abused specifically to obtain the training data. But what I'm talking about is using the data that already exists, taken from police investigations and other sources. Of course, it also requires victim's consent (as they grow old enough), as not everyone will agree to have materials of their abuse proliferate in any way.

Police has already used CSAM with victim's consent to better impersonate CSAM platform admins in investigative operations, leading to arrests of more child abusers and those sharing the materials around. While controversial, this came as a net benefit as it allowed to reduce the amount of avenues for CSAM sharing and the amount of people able to do so.

The case with AI is milder, as it requires minimum human interaction, so no one will need to re-watch the materials as long as victims are already identified. It's enough for the police to contact victims, get the agreement, and feed the data into AI without releasing the source. With enough data, AI could improve image and video generation, driving more watches away from real CSAM and reducing rates of abuse.

That is, if it works this way. There's a glaring research hole in this area, and I believe it is paramount to figure out if it helps. Then, we could decide whether to include already produced CSAM into the data, or if adult data is sufficient to make it good enough for the intended audience to make a switch.

[–] bookmeat@lemmynsfw.com 1 points 3 days ago (1 children)

You're going to tell me that there's no corporation out there that won't pay to improve their model with fresh data and not ask questions about where that data came from?

[–] Allero@lemmy.today 1 points 3 days ago* (last edited 3 days ago)

I think such matters should be kept strictly out of corporate hands, or be completed with total oversight.