doesn’t it follow that AI-generated CSAM can only be generated if the AI has been trained on CSAM?
This article even explicitely says as much.
My question is: why aren’t OpenAI, Google, Microsoft, Anthropic… sued for possession of CSAM? It’s clearly in their training datasets.
hendrik@palaver.p3x.de 4 weeks ago
Well, it can draw an astronaut on a horse, and I doubt it had seen lots of astronauts on horses...
ExtremeDullard@lemmy.sdf.org 4 weeks ago
Yeah but the article suggests that pedos train their local AI on existing CSAM, which would indicate that it’s somehow needed to generate AI-generated CSAM. Otherwise why would they bother? They’d just feed images of children in innocent settings and images of ordinary porn to get their local AI to generate CSAM.
rikudou@lemmings.world 4 weeks ago
How do they know that? Did the pedos text them to let them know? Sounds very made up.
hendrik@palaver.p3x.de 4 weeks ago
It's certainly technically possible. I suspect these AI models just aren't good at it. So the pedophiles need to train them on actual images.
I can imagine for example AI doesn't know what puberty is since it has in fact not seen a lot of naked children. It would try to infer from all the internet porn it's seen, and draw any female with big breasts, disregarding age. And that's not how children actually look.
I haven't tried, since it's illegal where I live. But that's my suspicion why pedophiles bother with training models.
GBU_28@lemm.ee 4 weeks ago
Training an existing model on a specific set of new data is known as “fine tuning”.
A base model has broad world knowledge and the ability to generate outputs of things it hasn’t specifically seen, but a tuned model will provide “better” (fucking yuck to even write it) results.
The closer your training data is to your desired result, the better.
Deceptichum@quokk.au 4 weeks ago
That’s not exactly how it works.
It can “understand” different concepts and mix them, without having to see the combination before hand.
As for the training thing, that would probably be more LORA. They’re like add-ons you can put on your AI to draw certain things better like a character, a pose, etc. not needed for the base model.
MolochAlter@lemmy.world 4 weeks ago
Why wouldn’t they? They have it on hand and it would obviously yield “better” results for their intended use case.
If you’re going as far as trying to generate AI csam you’re probably quite deep in that hole already, why settle for less?
AnAmericanPotato@programming.dev 4 weeks ago
This is not strictly true in general. Generative AI is able to produce output that is not in the training data, by learning a broad range of concepts and applying them in novel ways. I can generate an image of a rollerskating astronaut even if there are no rollerskating astronauts in the training data.
It is true that some training sets include CSAM, at least in the past. Back in 2023, researches found a few thousand such images in the LAION-5B dataset (roughly one per million images). 404 Media has an excellent article with details: www.404media.co/laion-datasets-removed-stanford-c…
On learning of this, LAION took down their database until it could properly cleaned. Source: laion.ai/notes/laion-maintenance/
Those images were collected from the public web. LAION took steps to avoid linking to illicit content (details in the link above), but clearly it’s an imperfect system. God only knows what closed companies (OpenAI, Google, etc.) are doing. With open data sets, at least any interested parties can review, verify, and report this stuff. With closed data sets, who knows?