phutatorius@lemmy.zip
on 05 Feb 2026 17:39
nextcollapse
This shit needs to be heavily regulated, with massive fines for anyone who inconveniences a customer due to a false positive or a staff error such as what happened in this case.
chillhelm@lemmy.world
on 05 Feb 2026 18:03
nextcollapse
Honestly don’t care too much about random super markets. This shit gets really fun when governments use it (like eg ICE is doing right now in the US).
Strider@lemmy.world
on 05 Feb 2026 18:34
nextcollapse
When you’ll notice it, it will be too late.
ultranaut@lemmy.world
on 05 Feb 2026 19:23
nextcollapse
Random super markets can license access to their data. I could easily imagine a company like Palantir or Flock leveraging systems like these in their government contracting. Whether or not these things are privately owned by creepy corporations or under the direct control of a government agency feels like a distinction without a difference, either way the infrastructure of totalitarianism is being constructed around us with these technologies.
ButteryMonkey@piefed.social
on 05 Feb 2026 19:24
collapse
I care about random supermarkets. People need to eat and obtain basic necessities even if some random database with no oversight says they might be dangerous in some way that we don’t get to know about.
AmidFuror@fedia.io
on 05 Feb 2026 18:37
nextcollapse
If you read the story, you'll see that it was face recognition by humans that was at fault, not automated face recognition. It would be like if the store had a picture posted in the staff room that said "Do not let this person shop here," and the staff had thought this shopper was the guy in error.
NABDad@lemmy.world
on 05 Feb 2026 20:14
nextcollapse
“Idiot Store Staff Mistake Someone For Someone Else” doesn’t get the same clicks.
Zamboni_Driver@lemmy.ca
on 05 Feb 2026 21:00
collapse
That’s a bit of a stretch to say the system was not at fault. The system pops up an alert and says he this brown guy should not be in your store and shows a picture of a brown guy, staff go out and find a different brown guy and kick him out of the store. It’s still the system which is the issue, it scanned faces, sent and alert, but wasn’t able to accurately communicate to the staff which specific person they should be worried about. The staff aren’t facial recognition experts, the shitty system led to this issue occuring.
Corey Doctrow calls the humans in this loop reverse centaurs..
Edit: not unicorns
MagnificentSteiner@lemmy.zip
on 05 Feb 2026 23:46
collapse
Instructions unclear… the government has now created regulations to further entrench the technology and make sure that the companies providing it have no responsibility or accountability to anyone whatsoever.
earlstilt@feddit.uk
on 05 Feb 2026 17:46
nextcollapse
Fuck Sainsburys
Maeve@kbin.earth
on 05 Feb 2026 19:01
nextcollapse
He said supermarket staff were unable to explain why he was being told to leave, and would only direct him to a QR code leading to the website of the firm Facewatch, which the retailer has hired to run facial recognition in some of its stores. He said when he contacted Facewatch, he was told to send in a picture of himself and a photograph of his passport before the firm confirmed it had no record of him on its database.
“One of the reasons I was angry was because I shouldn’t have to prove I am innocent,” Rajah said. “I shouldn’t have to prove I’m wrongly identified as a criminal.” He described the incident as feeling “quite like Minority Report, Orwellian”.
He said while doing his normal shop, he was approached by three members of the store’s staff, one of whom appeared to affirm that he was the person pictured on a device they had. It is understood the Facewatch system flagged someone else who had entered the store, and staff mistook Rajah for him.
Rajah was concerned some form of permanent record implying he had been involved in criminality might have been created on Facewatch’s system. Eventually, the firm told him he was not on its database and referred him back to Sainsbury’s.
BassTurd@lemmy.world
on 05 Feb 2026 21:31
collapse
I remember reading something last year about how some not most facial recognition software has racist biases because the models are primarily trained on Caucasian people. Based on just the name, I’m going to assume Rajeh is not white, and may be a victim of that bias. Granted, it shouldn’t be used at all, so it’s splitting hairs more than anything.
ChairmanMeow@programming.dev
on 05 Feb 2026 22:52
collapse
IIRC they found that even with balanced training data facial recognition models just do worse on darker-skinned people. Something about cameras picking up less contrast on the skin, meaning there are fewer easily-identifiable facial features it can pick up from an image.
Corporal_Punishment@feddit.uk
on 05 Feb 2026 19:58
nextcollapse
I have a senior data protection role in a public sector organisation. I actively do everything I can to prevent AI and biometrics being used except in the most innocuous of cases (eg letting people use chatgpt to help write stuff).
ZombieCyborgFromOuterSpace@piefed.ca
on 05 Feb 2026 20:30
nextcollapse
What the fuck kind of business does a grocery store have to store people’s biometric data? Fuck that.
Quexotic@infosec.pub
on 05 Feb 2026 22:03
nextcollapse
“sorry, you can’t buy food because of surveillance capitalism’s inherent violence”
Time to boycott.
Pappabosley@lemmy.world
on 05 Feb 2026 23:01
collapse
I think the main point is being missed, what does it matter even if did have a criminal past? We’re meant to have a rehabilitation focused justice system, which doesn’t work if you exclude ex-offenders from the spaces they need to be in to live. If it didn’t detect him actively committing a crime, then it shouldn’t have been an issue.
threaded - newest
This shit needs to be heavily regulated, with massive fines for anyone who inconveniences a customer due to a false positive or a staff error such as what happened in this case.
Honestly don’t care too much about random super markets. This shit gets really fun when governments use it (like eg ICE is doing right now in the US).
When you’ll notice it, it will be too late.
Random super markets can license access to their data. I could easily imagine a company like Palantir or Flock leveraging systems like these in their government contracting. Whether or not these things are privately owned by creepy corporations or under the direct control of a government agency feels like a distinction without a difference, either way the infrastructure of totalitarianism is being constructed around us with these technologies.
I care about random supermarkets. People need to eat and obtain basic necessities even if some random database with no oversight says they might be dangerous in some way that we don’t get to know about.
If you read the story, you'll see that it was face recognition by humans that was at fault, not automated face recognition. It would be like if the store had a picture posted in the staff room that said "Do not let this person shop here," and the staff had thought this shopper was the guy in error.
“Idiot Store Staff Mistake Someone For Someone Else” doesn’t get the same clicks.
That’s a bit of a stretch to say the system was not at fault. The system pops up an alert and says he this brown guy should not be in your store and shows a picture of a brown guy, staff go out and find a different brown guy and kick him out of the store. It’s still the system which is the issue, it scanned faces, sent and alert, but wasn’t able to accurately communicate to the staff which specific person they should be worried about. The staff aren’t facial recognition experts, the shitty system led to this issue occuring.
Corey Doctrow calls the humans in this loop reverse centaurs..
Edit: not unicorns
Instructions unclear… the government has now created regulations to further entrench the technology and make sure that the companies providing it have no responsibility or accountability to anyone whatsoever.
Fuck Sainsburys
I remember reading something last year about how some not most facial recognition software has racist biases because the models are primarily trained on Caucasian people. Based on just the name, I’m going to assume Rajeh is not white, and may be a victim of that bias. Granted, it shouldn’t be used at all, so it’s splitting hairs more than anything.
IIRC they found that even with balanced training data facial recognition models just do worse on darker-skinned people. Something about cameras picking up less contrast on the skin, meaning there are fewer easily-identifiable facial features it can pick up from an image.
I have a senior data protection role in a public sector organisation. I actively do everything I can to prevent AI and biometrics being used except in the most innocuous of cases (eg letting people use chatgpt to help write stuff).
What the fuck kind of business does a grocery store have to store people’s biometric data? Fuck that.
“sorry, you can’t buy food because of surveillance capitalism’s inherent violence”
Time to boycott.
I think the main point is being missed, what does it matter even if did have a criminal past? We’re meant to have a rehabilitation focused justice system, which doesn’t work if you exclude ex-offenders from the spaces they need to be in to live. If it didn’t detect him actively committing a crime, then it shouldn’t have been an issue.