You can compress multiple TB nothing with the occasional meme down to a few MB.
Comment on Black Mirror AI
essteeyou@lemmy.world 16 hours ago
This is surely trivial to detect. If the number of pages on the site is greater than some insanely high number then just drop all data from that site from the training data.
It’s not like I can afford to compete with OpenAI on bandwidth, and they’re burning through money with no cares already.
Korhaka@sopuli.xyz 11 hours ago
essteeyou@lemmy.world 4 hours ago
When I deliver it as a response to a request I have to deliver the gzipped version if nothing else. To get to a point where I’m poisoning an AI I’m assuming it’s going to require gigabytes of data transfer that I pay for.
At best I’m adding to the power consumption of AI.
MonkeMischief@lemmy.today 44 minutes ago
I wonder, can I serve it ads and get paid?
…and it’s just bouncing around and around and around in circles before its handler figures out what’s up…
Heehee I like where your head’s at!
bane_killgrind@slrpnk.net 15 hours ago
Yeah sure, but when do you stop gathering regularly constructed data, when your goal is to grab as much as possible?
Markov chains are an amazingly simple way to generate data like this, and a little bit of stacked logic it’s going to be indistinguishable from real large data sets.
Valmond@lemmy.world 13 hours ago
Imagine the staff meeting:
You: we didn’t gather any data because it was poisoned
Corposhill: we collected 120TB only from harry-potter-fantasy-club.il !!
Boss: hmm who am I going to keep…
yetAnotherUser@lemmy.ca 12 hours ago
The boss fires both, “replaces” them for AI, and tries to sell the corposhill’s dataset to AI companies that make fantasy novels