Up next: LAPD starts covering their faces and requests an emergency budget boost for masks.
‘FuckLAPD.com’ Lets Anyone Use Facial Recognition to Instantly Identify Cops
Submitted 2 days ago by remington@beehaw.org to technology@beehaw.org
https://www.404media.co/fucklapd-com-lets-anyone-use-facial-recognition-to-instantly-identify-cops/
Comments
Powderhorn@beehaw.org 2 days ago
Quexotic@beehaw.org 19 hours ago
Might wanna access that site with tor.
CanadaPlus@lemmy.sdf.org 23 hours ago
Oh how the tables have turned, lol.
Boomkop3@reddthat.com 1 day ago
Is this not doxxing? Posted by a mod no less
Gaywallet@beehaw.org 1 day ago
I would love to hear what has you concerned about a tool which provides a piece of information which is, by law (California Penal Code Section 830.10), supposed to be accessible to all individuals interacting with the officer - their name and/or badge number.
Boomkop3@reddthat.com 1 day ago
The concern was the lack of knowledge that this was public. I noticed it’s in the article, I may have read over it
Vodulas@beehaw.org 1 day ago
A. No, this is an article talking about the tool. B. Cops are public figures. Name and badge number are public information. Hence why the first sentence in the article states it uses public records. It does not give their address and phone number. It is not doxxing
Boomkop3@reddthat.com 1 day ago
A. A gun is a tool as well, doesn’t mean you should make them public available
B. That makes a lot of sense. I’m not from around there, sorry for the misunderstanding
CanadaPlus@lemmy.sdf.org 23 hours ago
If it was identifying Lemmy users, it definitely would be. But, it’s a tool that reveals identities of a small, supposedly accountable group, and we’re just mentioning it, so it seems like there’s at least an argument to allow it.
icelimit@lemmy.ml 23 hours ago
Not supposedly, a group that is duty bound to be accountable.
Megaman_EXE@beehaw.org 1 day ago
I find this funny because the police have been doing this with civilians. My main concern is that this tech is not 100% accurate. I feel like it shouldn’t be used on its own.
I guess if it is used as a supplementary tool and not the main piece of evidence, it could maybe be okay? But I would be scared it would target an innocent individual, which could cause very negative or dangerous consequences. The main thing would be accuracy. I don’t know if it was addressed as part of the article is paywalled.
BurningRiver@beehaw.org 1 day ago
I find this funny because this
Is generally the least of their concerns.
Megaman_EXE@beehaw.org 1 day ago
Heh yeah you’re right
scrubbles@poptalk.scrubbles.tech 1 day ago
That’s the nuance of AI that anyone who has done any actual work with ML has known for decades now. ML is amazing. It’s not perfect. It’s actually pretty far from perfect. So you should never ever use it as a solo check, but it can be great for a double check.
Such as with cancer. AI can be a wonderful choice to detecting a melanoma, if used correctly. Such as:
a doctor has already cleared a mole, but if you want to know if it warrants a second opinion by another doctor. You could have the model to have a confidence of say, 80% sure that the first doctor is correct in that it is fine.
if you do not have access to a doctor immediately, it can be a fine check, again only to a certain percentage. Say that in this case in the future you are worried but cannot access a doctor easily. A patient could snap a photo and in this case a very high confidence rating would say that it is probably fine, with a disclaimer that it is just an AI model and if it changes or you are still worried, get it checked.
Unfortunately, all of that nuance in that it is all just probabilities is completely lost on both the creators of all of these AI tools, and the risks are not actually passed to the users so blind trust is the number one problem.
We see it here with police too. “It said it’s them”. No, it only said to a specific confidence that it might be them. That’s a very different thing. You should never use it to find someone, only to verify someone.
I actually really like how airport security implemented it because it’s actually using it well. Here’s an ID, it has a photo of a person. Compare it to the photo taken there in person, and it should verify to a very high confidence that they are the same person. If in doubt, there’s a human there to also verify it. That’s good ML usage.
Powderhorn@beehaw.org 1 day ago
I’m miles away from AI, so this may be me talking out of my ass, but shouldn’t a smaller database (thousands) be more accurate than anything orders of magnitude larger?
Megaman_EXE@beehaw.org 1 day ago
You know, I honestly don’t know! That’s a good question