Comment on What is the attraction to kids?
skullgiver@popplesburger.hilciferous.nl 1 year agoThat’s putting a lot of faith into CLIP, though. The thing is, to get CLIP to detect things like child porn reliably, you do need to train it to make the distinction. In my experience, CLIP tends to make up keywords, or at least misunderstand the situation surprisingly common.
If it weren’t super illegal and super unethical, AI could easily distinguish normal porn from illegal porn if you feed it enough tagged data of both. Categorisation is something these models are very good at, after all. That’s never ever going to happen (imagine the poor shmuck being hired to tag child rape for a dollar a day, horrific) but it’s the only way I trust AI to come up with something like this.
I think we need more research into this field. I’m also at least a little mad that AI companies release these models into the wild before the science is ready to prevent it from becoming a child rape image generator for the mentally ill. Companies just seem to throw their hands in the air and go “well we didn’t program it to do that, not our fault!” and deny any responsibility for what they’ve created.
Adalast@lemmy.world 1 year ago
You are correct, CLIP can misinterpret things, which is where human intelligence comes in. Having CLIP process the probabilities for the terminology that you describe what you are looking for then utilizing a bit of heuristics can go a long way. You don’t need to train it to recognize a nude child because it has been trained to recognize a child, and it has been trained to recognize nudity, so if an image scores high in “nude” and “child” just throw it out. Granted, it might be a picture of a woman breastfeeding while a toddler looks on, which is inherently not child pornography, but unless that is the specific image that is being prompted for, it is not that big of a deal to just toss it. We understand the conceptual linking so we can set the threshold parameters and adjust as needed.
As for the companies, it is a tough world surrounding it. The argument of a company that produced a piece of software being culpable for the misuse of said software is a very tenuous one. There have been attempts to make gun manufacturers liable for gun deaths (especially handguns since they really only have the purpose of killing humans). This one I can see, as the firearm killing a person is not a “misuse”, indeed, it is the express purpose for it’s creation. But what this would be would be more akin to wanting to hold Adobe liable for the child pornography that is edited in Lightroom, or Dropbox liable for someone using Dropbox API to set up a private distribution network for illicit materials. In reality, as long as the company did not design a product with the illegal activity expressly in mind, then they really shouldn’t be culpable for how people use it once it is in the wild.
I do feel like more needs to be done to make public the training data for public inspection, as well as forensic methods for interrogating the end products to figure out if they are lying and hiding materials that were used for training. That is just a general issue though that covers many of the ethical and legal issues surrounding AI training.