A big biometric security company in the UK, Facewatch, is in hot water after their facial recognition system caused a major snafu - the system wrongly identified a 19-year-old girl as a shoplifter.

  • PseudorandomNoise@lemmy.world
    link
    fedilink
    English
    arrow-up
    271
    arrow-down
    2
    ·
    1 month ago

    Despite concerns about accuracy and potential misuse, facial recognition technology seems poised for a surge in popularity. California-based restaurant CaliExpress by Flippy now allows customers to pay for their meals with a simple scan of their face, showcasing the potential of facial payment technology.

    Oh boy, I can’t wait to be charged for someone else’s meal because they look just enough like me to trigger a payment.

    • Cethin@lemmy.zip
      link
      fedilink
      English
      arrow-up
      184
      ·
      1 month ago

      I have an identical twin. This stuff is going to cause so many issues even if it worked perfectly.

      • Maggoty@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        edit-2
        1 month ago

        Not uh! All you and your twin have to do is write the word Twin on your forehead every morning. Just make sure to never commit a crime with it written where your twin puts their sign. Or else, you know… You might get away with it.

        Nope no obvious problems here at all!

      • CeeBee@lemmy.world
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        31
        ·
        1 month ago

        Ok, some context here from someone who built and worked with this kind tech for a while.

        Twins are no issue. I’m not even joking, we tried for multiple months in a live test environment to get the system to trip over itself, but it just wouldn’t. Each twin was detected perfectly every time. In fact, I myself could only tell them apart by their clothes. They had very different styles.

        The reality with this tech is that, just like everything else, it can’t be perfect (at least not yet). For all the false detections you hear about, there have been millions upon millions of correct ones.

          • CeeBee@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            19
            ·
            1 month ago

            Yes, because like I said, nothing is ever perfect. There can always be a billion little things affecting each and every detection.

            A better statement would be “only one false detection out of 10 million”

            • Zron@lemmy.world
              link
              fedilink
              English
              arrow-up
              41
              ·
              1 month ago

              You want to know a better system?

              What if each person had some kind of physical passkey that linked them to their money, and they used that to pay for food?

              We could even have a bunch of security put around this passkey that makes it’s really easy to disable it if it gets lost or stolen.

              As for shoplifting, what if we had some kind of societal system that levied punishments against people by providing a place where the victim and accused can show evidence for and against the infraction, and an impartial pool of people decides if they need to be punished or not.

              • CeeBee@lemmy.world
                link
                fedilink
                English
                arrow-up
                12
                arrow-down
                2
                ·
                1 month ago

                100%

                I don’t disagree with a word you said.

                FR for a payment system is dumb.

            • fishpen0@lemmy.world
              link
              fedilink
              English
              arrow-up
              18
              arrow-down
              1
              ·
              edit-2
              1 month ago

              Another way to look at that is ~810 people having an issue with a different 810 people every single day assuming only one scan per day. That’s 891,000 people having a huge fucking problem at least once every single year.

              I have this problem with my face in the TSA pre and passport system and every time I fly it gets worse because their confidence it is correct keeps going up and their trust in my actual fucking ID keeps going down

              • CeeBee@lemmy.world
                link
                fedilink
                English
                arrow-up
                6
                arrow-down
                6
                ·
                edit-2
                1 month ago

                I have this problem with my face in the TSA pre and passport system

                Interesting. Can you elaborate on this?

                Edit: downvotes for asking an honest question. People are dumb

        • MonkderDritte@feddit.de
          link
          fedilink
          English
          arrow-up
          18
          arrow-down
          1
          ·
          1 month ago

          it can’t be perfect (at least not yet).

          Or ever, because it locks you out after a drunken night otherwise.

          • CeeBee@lemmy.world
            link
            fedilink
            English
            arrow-up
            8
            arrow-down
            4
            ·
            1 month ago

            Or ever because there is no such thing as 100% in reality. You can only add more digits at the end of your accuracy, but it will never reach 100.

        • boatswain@infosec.pub
          link
          fedilink
          English
          arrow-up
          15
          ·
          1 month ago

          In fact, I myself could only tell them apart by their clothes. They had very different styles.

          This makes it sound like you only tried one particular set of twins–unless there were multiple sets, and in each set the two had very different styles? I’m no statistician, but a single set doesn’t seem statistically significant.

          • CeeBee@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            4
            ·
            1 month ago

            What I’m saying is we had a deployment in a large facility. It was a partnership with the org that owned the facility to allow us to use their location as a real-world testing area. We’re talking about multiple buildings, multiple locations, and thousands of people (all aware of the system being used).

            Two of the employees were twins. It wasn’t planned, but it did give us a chance to see if twins were a weak point.

            That’s all I’m saying. It’s mostly anecdotal, as I can’t share details or numbers.

            • boatswain@infosec.pub
              link
              fedilink
              English
              arrow-up
              9
              arrow-down
              2
              ·
              1 month ago

              Two of the employees were twins. It wasn’t planned, but it did give us a chance to see if twins were a weak point.

              No, it gave you a chance to see if that particular set of twins was a weak point.

              • CeeBee@lemmy.world
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                1
                ·
                1 month ago

                With that logic we would need to test the system on every living person to see where it fails.

                The system had been tested ad nauseum in a variety of scenarios (including with twins and every other combination you can think of, and many you can’t). In this particular situation, a real-world test in a large facility with many hundreds of cameras everywhere, there happened to be twins.

                It’s a strong data point regardless of your opinion. If it was the only one then you’d have a point. But like I said, it was an anecdotal example.

        • techt@lemmy.world
          link
          fedilink
          English
          arrow-up
          8
          ·
          edit-2
          1 month ago

          Can you please start linking studies? I think that might actually turn the conversation in your favor. I found a NIST study (pdf link), on page 32, in the discussion portion of 4.2 “False match rates under demographic pairing”:

          The results above show that false match rates for imposter pairings in likely real-world scenarios are much higher than those from measured when imposters are paired with zero-effort.

          This seems to say that the false match rate gets higher and higher as the subjects are more demographically similar; the highest error rate on the heat map below that is roughly 0.02.

          Something else no one here has talked about yet – no one is actively trying to get identified as someone else by facial recognition algorithms yet. This study was done on public mugshots, so no effort to fool the algorithm, and the error rates between similar demographics is atrocious.

          And my opinion: Entities using facial recognition are going to choose the lowest bidder for their system unless there’s a higher security need than, say, a grocery store. So, we have to look at the weakest performing algorithms.

          • CeeBee@lemmy.world
            link
            fedilink
            English
            arrow-up
            7
            ·
            1 month ago

            My references are the NIST tests.

            https://pages.nist.gov/frvt/reports/1N/frvt_1N_report.pdf

            That might be the one you’re looking at.

            Another thing to remember about the NIST tests is that they try to use a standardized threshold across all vendors. The point is to compare the results in a fair manner across systems.

            The system I worked on was tested by NIST with an FMR of 1e-5. But we never used that threshold and always used a threshold that equated to 1e-7, which is orders of magnitude more accurate.

            And my opinion: Entities using facial recognition are going to choose the lowest bidder for their system unless there’s a higher security need than, say, a grocery store. So, we have to look at the weakest performing algorithms.

            This definitely is a massive problem and likely does contribute to poor public perception.

            • techt@lemmy.world
              link
              fedilink
              English
              arrow-up
              4
              ·
              1 month ago

              Thanks for the response! It sounds like you had access to a higher quality system than the worst, to be sure. Based on your comments I feel that you’re projecting the confidence in that system onto the broader topic of facial recognition in general; you’re looking at a good example and people here are (perhaps cynically) pointing at the worst ones. Can you offer any perspective from your career experience that might bridge the gap? Why shouldn’t we treat all facial recognition implementations as unacceptable if only the best – and presumably most expensive – ones are?

              A rhetorical question aside from that: is determining one’s identity an application where anything below the unachievable success rate of 100% is acceptable?

              • CeeBee@lemmy.world
                link
                fedilink
                English
                arrow-up
                5
                ·
                1 month ago

                Based on your comments I feel that you’re projecting the confidence in that system onto the broader topic of facial recognition in general; you’re looking at a good example and people here are (perhaps cynically) pointing at the worst ones. Can you offer any perspective from your career experience that might bridge the gap? Why shouldn’t we treat all facial recognition implementations as unacceptable if only the best – and presumably most expensive – ones are?

                It’s a good question, and I don’t have the answer to it. But a good example I like to point at is the ACLU’s announcement of their test on Amazon’s Rekognition system.

                They tested the system using the default value of 80% confidence, and their test resulted in 20% false identification. They then boldly claimed that FR systems are all flawed and no one should ever use them.

                Amazon even responded saying that the ACLU’s test with the default values was irresponsible, and Amazon’s right. This was before such public backlash against FR, and the reasoning for a default of 80% confidence was the expectation that most people using it would do silly stuff like celebrity lookalikes. That being said, it was stupid to set the default to 80%, but that’s just hindsight speaking.

                My point here is that, while FR tech isn’t perfect, the public perception is highly skewed. If there was a daily news report detailing the number of correct matches across all systems, these few showing a false match would seem ridiculous. The overwhelming vast majority of news reports on FR are about failure cases. No wonder most people think the tech is fundamentally broken.

                A rhetorical question aside from that: is determining one’s identity an application where anything below the unachievable success rate of 100% is acceptable?

                I think most systems in use today are fine in terms of accuracy. The consideration becomes “how is it being used?” That isn’t to say that improvements aren’t welcome, but in some cases it’s like trying to use the hook on the back of a hammer as a screw driver. I’m sure it can be made to work, but fundamentally it’s the wrong tool for the job.

                FR in a payment system is just all wrong. It’s literally forcing the use of a tech where it shouldn’t be used. FR can be used for validation if increased security is needed, like accessing a bank account. But never as the sole means of authentication. You should still require a bank card + pin, then the system can do FR as a kind of 2FA. The trick here would be to first, use a good system, and then second, lower the threshold that borders on “fairly lenient”. That way you eliminate any false rejections while still maintaining an incredibly high level of security. In that case the chances of your bank card AND pin being stolen by someone who looks so much like you that it tricks FR is effectively impossible (but it can never be truly zero). And if that person is being targeted by a threat actor who can coordinate such things then they’d have the resources to just get around the cyber security of the bank from the comfort of anywhere in the world.

                Security in every single circumstance is a trade-off with convenience. Always, and in every scenario.

                FR works well with existing access control systems. Swipe your badge card, then it scans you to verify you’re the person identified by the badge.

                FR also works well in surveillance, with the incredibly important addition of human-in-the-loop. For example, the system I worked on simply reported detections to a SoC (with all the general info about the detection including the live photo and the reference photo). Then the operator would have to look at the details and manually confirm or reject the detection. The system made no decisions, it simply presented the info to an authorized person.

                This is the key portion that seems to be missing in all news reports about false arrests and whatnot. I’ve looked into all the FR related false arrests and from what I could determine none of those cases were handled properly. The detection results were simply taken as gospel truth and no critical thinking was applied. In some of those cases the detection photo and reference (database) photo looked nothing alike. It’s just the people operating those systems are either idiots or just don’t care. Both of those are policy issues entirely unrelated to the accuracy of the tech.

                • hazeebabee@slrpnk.net
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  1 month ago

                  Super interesting to read your more technical perspective. I also think facial recognition (and honestly most AI use cases) are best when used to supplement an existing system. Such as flagging a potential shoplifter to human security.

                  Sadly most people don’t really understand the tech they use for work. If the computer tells them something they just kind of blindly believe it. Especially in a work environment where they have been trained to do what the machine says.

                  My guess is that the people were trained on how to use the system at a very basic level. Troubleshooting and understanding the potential for error typically isn’t covered in 30min corporate instructional meetings. They just get a little notice saying a shoplifter is in the store and act on that without thinking.

                • techt@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  1 month ago

                  The mishandling is indeed what I’m concerned about most. I now understand far better where you’re coming from, sincere thanks for taking the time to explain. Cheers

        • Cethin@lemmy.zip
          link
          fedilink
          English
          arrow-up
          5
          ·
          1 month ago

          This tech (AI detection) or purpose built facial recognition algorithms?

      • Telodzrum@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        28
        ·
        1 month ago

        If it works anything like Apple’s Face ID twins don’t actually map all that similar. In the general population the probability of matching mapping of the underlying facial structure is approximately 1:1,000,000. It is slightly higher for identical twins and then higher again for prepubescent identical twins.

        • MonkderDritte@feddit.de
          link
          fedilink
          English
          arrow-up
          39
          ·
          edit-2
          1 month ago

          Meaning, 8’000 potential false positives per user globally. About 300 in US, 80 in Germany, 7 in Switzerland.

          Might be enough for Iceland.

          • ramjambamalam@lemmy.ca
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            1 month ago

            I can already imagine the Tom Clancy thriller where some Joe Nobody gets roped into helping crack a terrorist’s locked phone because his face looks just like the terrorist’s.

          • Telodzrum@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            36
            ·
            1 month ago

            Yeah, which is a really good number and allows for near complete elimination of false matches along this vector.

            • 4am@lemm.ee
              link
              fedilink
              English
              arrow-up
              26
              arrow-down
              1
              ·
              1 month ago

              I promise bro it’ll only starve like 400 people please bro I need this

              • Telodzrum@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                16
                ·
                1 month ago

                No you misunderstood. That is a reduction in commonality by a literal factor of one million. Any secondary verification point is sufficient to reduce the false positive rate to effectively zero.

                • AwesomeLowlander@lemmy.dbzer0.com
                  link
                  fedilink
                  English
                  arrow-up
                  20
                  ·
                  1 month ago

                  secondary verification point

                  Like, running a card sized piece of plastic across a reader?

                  It’d be nice if they were implementing this to combat credit card fraud or something similar, but that’s not how this is being deployed.

                • BassTurd@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  16
                  ·
                  1 month ago

                  Which means the face recognition was never necessary. It’s a way for companies to build a database that will eventually get exploited. 100% guarantee.

        • Cethin@lemmy.zip
          link
          fedilink
          English
          arrow-up
          22
          arrow-down
          1
          ·
          1 month ago

          Yeah, people with totally different facial structures get identified as the same person all the time with the “AI” facial recognition, especially if your darker skinned. Luckily (or unluckily) I’m white as can be.

          I’m assuming Apple’s software is a purpose built algorithm that detects facial features and compares them, rather than the black box AI where you feed in data and it returns a result. Thats the smart way to do it, but it takes more effort.

          • CeeBee@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 month ago

            people with totally different facial structures get identified as the same person all the time with the “AI” facial recognition

            All the time, eh? Gonna need a citation on that. And I’m not talking about just one news article that pops up every six months. And nothing that links back to the UCLA’s 2018 misleading “report”.

            I’m assuming Apple’s software is a purpose built algorithm that detects facial features and compares them, rather than the black box AI where you feed in data and it returns a result.

            You assume a lot here. People have this conception that all FR systems are trained blackbox models. This is true for some systems, but not all.

            The system I worked with, which ranked near the top of the NIST FRVT reports, did not use a trained AI algorithm for matching.

            • Cethin@lemmy.zip
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              1 month ago

              I’m not doing a bunch of research to prove the point. I’ve been hearing about them being wrong fairly frequently, especially on darker skinned people, for a long time now. It doesn’t matter how often it is. It sounds like you have made up your mind already.

              I’m assuming that of apple because it’s been around for a few years longer than the current AI craze has been going on. We’ve been doing facial recognition for decades now, with purpose built algorithms. It’s not mucb of leap to assume that’s what they’re using.

              • CeeBee@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                ·
                1 month ago

                I’ve been hearing about them being wrong fairly frequently, especially on darker skinned people, for a long time now.

                I can guarantee you haven’t. I’ve worked in the FR industry for a decade and I’m up to speed on all the news. There’s a story about a false arrest from FR at most once every 5 or 6 months.

                You don’t see any reports from the millions upon millions of correct detections that happen every single day. You just see the one off failure cases that the cops completely mishandled.

                I’m assuming that of apple because it’s been around for a few years longer than the current AI craze has been going on.

                No it hasn’t. FR systems have been around a lot longer than Apple devices doing FR. The current AI craze is mostly centered around LLMs, object detection and FR systems have been evolving for more than 2 decades.

                We’ve been doing facial recognition for decades now, with purpose built algorithms. It’s not mucb of leap to assume that’s what they’re using.

                Then why would you assume companies doing FR longer than the recent “AI craze” would be doing it with “black boxes”?

                I’m not doing a bunch of research to prove the point.

                At least you proved my point.

                • Cethin@lemmy.zip
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  1 month ago

                  You don’t see any reports from the millions upon millions of correct detections that happen every single day. You just see the one off failure cases that the cops completely mishandled.

                  Obviously. I don’t have much of an issue with it when it’s working properly (although I do still absolutely have an issue with it still). It being wrong and causing issues fairly frequently, and every 5 or 6 months is frequent (this is a low number, just the frequency of it causing newsworthy issues) with it not being deployed widely yet, is a pretty big issue. Scale that up by several orders of magnitude if it’s widely adopted and the errors will be constant.

                  No it hasn’t. FR systems have been around a lot longer than Apple devices doing FR. The current AI craze is mostly centered around LLMs, object detection and FR systems have been evolving for more than 2 decades… Then why would you assume companies doing FR longer than the recent “AI craze” would be doing it with “black boxes”?

                  You’re repeating what I said. Apples FR tech is a few years older than the machine learning tech that we have now. FR in general is several decades old, and it’s not ML based. It’s not a black box. You can actually know what it’s doing. I specifically said they weren’t doing it with black boxes. I said the AI models are. Please read again before you reply.

                  At least you proved my point.

                  You wrongly assuming what I said, which is actually the opposite of what I said, is the reason I’m not putting in the effort. You’ve made up your mind. I’m not going to change it, so I’m not putting in the effort it would take to gather the data, just to throw it into the wind. It sounds like you are already aware of some of it, but somehow think it’s not bad.

        • 4am@lemm.ee
          link
          fedilink
          English
          arrow-up
          21
          arrow-down
          2
          ·
          1 month ago

          And yet this woman was mistaken for a 19-year-old 🤔

          • Telodzrum@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            20
            ·
            1 month ago

            Shitty implementation doesn’t mean shitty concept, you’d think a site full of tech nerds would understand such a basic concept.

            • Hawk@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 month ago

              Pretty much everyone here agrees that it’s a shitty concept. Doesn’t solve anything and it’s a privacy nightmare.

        • chiisana@lemmy.chiisana.net
          link
          fedilink
          English
          arrow-up
          8
          ·
          1 month ago

          I think from a purely technical point of view, you’re not going to get FaceID kind of accuracy on theft prevention systems. Primarily because FaceID uses IR array scanning within arm’s reach from the user, whereas theft prevention is usually scanned from much further away. The distance makes it much harder to get the fidelity of data required for an accurate reading.

          • sugar_in_your_tea@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            5
            ·
            1 month ago

            Yup, it turns out if you have millions of pixels to work with, you have a better shot at correctly identifying someone than if you have dozens.

          • CeeBee@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 month ago

            I think from a purely technical point of view, you’re not going to get FaceID kind of accuracy on theft prevention systems. Primarily because FaceID uses IR array scanning within arm’s reach from the user, whereas theft prevention is usually scanned from much further away. The distance makes it much harder to get the fidelity of data required for an accurate reading.

            This is true. The distance definitely makes a difference, but there are systems out there that get incredibly high accuracy even with surveillance footage.

    • Liz@midwest.social
      link
      fedilink
      English
      arrow-up
      9
      ·
      1 month ago

      I have come across a stranger online who looks exactly like me. We even share the same first name. We even live in the same area. I’m so excited for this wonderful new technology…

      ಠ⁠_⁠ಠ

    • TrickDacy@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 month ago

      Are we assuming there is no pin or any other auth method? That would be unlike any other payment system I’m aware of. I have to fingerprint scan on my phone to use my credit cards even though I just unlocked my phone to attempt it

      • MajorHavoc@programming.dev
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 month ago

        I’ve got an extremely common-looking face in a major city.

        Indeed, it’s likely to be a problem, if you stick with committing few or no crimes.

        The good news is that, should you choose to commit an above-average number of crimes, then the system will be prone to under-report you.

        So that’s nice. /sarcasm, I’m not actually advocating for more crimes. Though I am pointing out that maybe the folks installing these things aren’t incentivising the behavior they want.

  • Jackthelad@lemmy.world
    link
    fedilink
    English
    arrow-up
    184
    arrow-down
    1
    ·
    1 month ago

    Well, this blows the “if you’ve not done anything wrong, you have nothing to worry about” argument out of the water.

  • MentalEdge@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    179
    arrow-down
    2
    ·
    1 month ago

    Even if someone did steal a mars-bar… Banning them from all food-selling establishments seems… Disproportional.

    Like if you steal out of necessity, and get caught once, you then just starve?

    Obviously not all grocers/chains/restaurants are that networked yet, but are we gonna get to a point where hungry people are turned away at every business that provides food, once they are on “the list”?

    • DivineDev@kbin.run
      link
      fedilink
      arrow-up
      49
      ·
      1 month ago

      No no, that would be absurd. You’ll also be turned away if you are not on the list if you’re unlucky.

    • FuryMaker@lemmy.world
      link
      fedilink
      English
      arrow-up
      32
      ·
      1 month ago

      get caught once, you then just starve?

      Maybe they send you to Australia again?

      The world hasn’t changed has it.

    • lolcatnip@reddthat.com
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      2
      ·
      edit-2
      1 month ago

      They’ve essentially created their own privatized law enforcement system. They aren’t allowed to enforce their rules the same way a government would be, but punishment like banning a person from huge swaths of economic life can still be severe. The worst part is that private legal systems almost never have any concept of rights or due process, so there is absolutely nothing stopping them from being completely arbitrary in how they apply their punishments.

      I see this kind of thing as being closely aligned with right wingers’ desire to privatize everything, abolish human rights, and just generally turn the world into a dystopian hellscape for anyone who isn’t rich and well connected.

    • MacN'Cheezus@lemmy.today
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 month ago

      Like if you steal out of necessity, and get caught once, you then just starve?

      I mean… you could try getting on food stamps or whatever sort of government assistance is available in your country for this purpose?

      In pretty much all civilized western countries, you don’t HAVE to resort to becoming a criminal simply to get enough food to survive. It’s really more of a sign of antisocial behavior, i.e. a complete rejection of the system combined with a desire to actively cause harm to it.

      Or it could be a pride issue, i.e. people not wanting to admit to themselves that they are incapable of taking care of themselves on their own and having to go to a government office in order to “beg” for help (or panhandle outside the supermarket instead).

    • intensely_human@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      5
      ·
      1 month ago

      If that case ever does exist (god forbid), I hope that there’s something like a free-entry market so they can set up their own food solutions instead of being forced to starve.

      If it’s a free market, and every existing business is coordinating to refuse to sell food to this person, then there’s a profit opportunity in getting food to them. You could even charge them double for food, and make higher profits selling to the grocery-banned class, while saving their lives.

      That may sound cold-hearted, but what I’m trying to point out is that in this scenario, the profit motive is pulling that food to those people who need it. It’s incentivizing people who otherwise wouldn’t care, and enabling people who do care, to feed those hungry people by channeling money toward the solution.

      And that doesn’t require anything specific about food to be in the code that runs the marketplace. All you need is a policy that new entrants to the market are allowed, and without any lengthy waiting process for a permit or whatever. You need a rule that says “You can’t stop other players from coming in an competing with you”, which is the kind of rule you need to run a free market, and then the rest of the problem is solved by people’s natural inclinations.

      I know I’m piggybacking here. I’m just saying that a situation in which only some finite cartel of providers gets to decide who can buy food, is an example of a massive violation of free market principals.

      People think “free market” means “the market is free to do evil”. No. “Free market” just means the people inside it are free to buy and sell what they please, when they please.

      Yes it means stores can ban people. But it also means other people can start stores that do serve those people. It means “I don’t have to deal with you if I don’t want to, but I also can’t control your access to other people”.

      A pricing cartel or a blacklisting cartel is a form of market disease. The best prevention and cure is to ensure the market is a free one - one which new players can enter at will - which means you can’t enforce that cartel reliably since there’s always someone outside the blacklisting cartel who could benefit from defecting from the blacklist.

      • MentalEdge@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        7
        ·
        edit-2
        1 month ago

        That is some serious “capitalism can solve anything and therefore will, if only we let it”-type brain rot.

        This “solution” relies on so many assumptions that don’t even begin to hold water.

        Of course any utopian framework for society could deal with every conceivable problem… But in practice they don’t, and always require intentional regulation to a greater or lesser extent in order to prevent harm, because humans are humans.

        This particular potential problem is almost certainly not the kind that simply “solves itself” if you let it.

        And IMO suggesting otherwise is an irresponsible perpetuation of the kind of thinking that has led human civilization to the current reality of millions starving in the next few decades, due to the predictable environmental destruction of arable land in the near future.

      • Zagorath@aussie.zone
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 month ago

        Wtf kind of Randian hellscape nonsense is this? They should be allowed to charge double to exploit people who are already disadvantaged by the way other companies are treating them? Fuck this nonsense.

        Go create your own bear-infested village somewhere nobody with any morals has to live near you. But this time do it from scratch rather than ruining things for the people already living there.

    • orrk@lemmy.world
      link
      fedilink
      English
      arrow-up
      51
      arrow-down
      2
      ·
      1 month ago

      you will sit down and be quiet, all you parasites stifling innovation, the market will solve this, because it is the most rational thing in existence, like trains, oh god how I love trains, I want to be f***ed by trains.

      ~~Rand

      • Jtotheb@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 month ago

        The whole wide world of authors who have written about the difficulties of this new technological age and you choose the one who had to pretend her work was unpopular

        • orrk@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 month ago

          go read rand, her books literally advocate for an anti-social world order, where the rich and powerful have the ability to do whatever they want without impediment as the “workers” are described as parasites that should get in line or die

    • Alexstarfire@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      18
      ·
      1 month ago

      Are you suggesting they shouldn’t be allowed to ban people from stores? The only problem I see here is misused tech. If they can’t verify the person, they shouldn’t be allowed to use the tech.

      I do think there need to be reprocussions for situations like this.

      • Queen HawlSera@lemm.ee
        link
        fedilink
        English
        arrow-up
        18
        ·
        1 month ago

        Well there should be a limited amount of ability to do so. I mean there should be police reports or something at the very least. I mean, what if Facial Recognition AI catches on in grocery stores? Is this woman just banned from all grocery stores now? How the fuck is she going to eat?

        • Alexstarfire@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          1 month ago

          That’s why I said this was a misuse of tech. Because that’s extremely problematic. But there’s nothing to stop these same corps from doing this to a person even if the tech isn’t used. This tech just makes it easier to fuck up.

          I’m against the use of this tech to begin with but I’m having a hard time figuring out if people are more upset about the use of the tech or about the person being banned from a lot of stores because of it. Cause they are separate problems and the latter seems more of an issue than the former. But it also makes fucking up the former matter a lot more as a result.

          • TonyOstrich@lemmy.world
            link
            fedilink
            English
            arrow-up
            8
            ·
            1 month ago

            I wish I could remember where I saw it, but years ago I read something in relation to policing that said a certain amount of human inefficiency in a process is actually a good thing to help balance bias and over reach that could occur when technology could technically do in seconds what would take a human days or months.

            In this case if a person is enough of a problem that their face becomes known at certain branches of a store it’s entirely reasonable for that store to post a sign with their face saying they are aren’t allowed. In my mind it would essentially create a certain equilibrium in terms of consequences and results. In addition to getting in trouble for stealing itself, that individual person also has a certain amount of hardship placed on them that may require they travel 40 minutes to do their shopping instead of 5 minutes to the store nearby. A sign and people’s memory also aren’t permanent, so it’s likely that after a certain amount of time that person would probably be able to go back to that store if they had actually grown out of it.

            Or something to that effect. If they steal so much that they become known to the legal system there should be processes in place to address it.

            And even with all that said, I’m just not that concerned with theft at large corporate retailers considering wage theft dwarfs thefts by individuals by at least an order of magnitude.

          • Queen HawlSera@lemm.ee
            link
            fedilink
            English
            arrow-up
            7
            ·
            1 month ago

            Reddit became a ban-happy wasteland, and if the tides swing a similar way, we’ll see a society where Big Tech gates people out of the very foundation of Modern Society. It’s exclusion that I’m against.

  • retrospectology@lemmy.world
    link
    fedilink
    English
    arrow-up
    135
    arrow-down
    2
    ·
    edit-2
    1 month ago

    This can’t be true. I was told that if she has nothing to hide she has nothing to worry about!

  • HiramFromTheChi@lemmy.world
    link
    fedilink
    English
    arrow-up
    110
    arrow-down
    1
    ·
    1 month ago

    Not the first time facial recognition tech has been misused, and certainly won’t be the last. The UK in particular has caught a lotta flak around this.

    We seem to have a hard time connecting the digital world to the physical world and realizing just how interwoven they are at this point.

    Therefore, I made an open source website called idcaboutprivacy to demonstrate the importance—and dangers—of tech like this.

    It’s a list of news articles that demonstrate real-life situations where people are impacted.

    If you wanna contribute to the project, please do. I made it simple enough to where you don’t need to know Git or anything advanced to contribute to it. (I don’t even really know Git.)

  • Buffalox@lemmy.world
    link
    fedilink
    English
    arrow-up
    92
    arrow-down
    4
    ·
    1 month ago

    This is why some UK leaders wanted out of EU, to make their own rules with way less regard for civil rights.

    • Lad@reddthat.com
      link
      fedilink
      English
      arrow-up
      11
      ·
      1 month ago

      It’s the Tory way. Authoritarianism, culture wars, fucking over society’s poorest.

    • ᕙ(⇀‸↼‶)ᕗ@lemm.ee
      link
      fedilink
      English
      arrow-up
      14
      arrow-down
      10
      ·
      1 month ago

      nah i think main thing was a super fragile identity. i mean they have been shit all the time since before EU. when talks between france,germany and uk took place the insisted to take control of EU.

      if you live on an island for generations with limited new genetic input…well, thats where you end up.

      • sailingbythelee@lemmy.world
        link
        fedilink
        English
        arrow-up
        28
        arrow-down
        4
        ·
        1 month ago

        We humans have these things called “boats” that have enabled the British Isles to receive regular inputs of new genetic material. Pretty useful things, these boats, and somewhat pivotal in the history of the UK.

      • FooBarrington@lemmy.world
        link
        fedilink
        English
        arrow-up
        17
        ·
        1 month ago

        I don’t understand the tendency to attribute harmful behaviours of the rich and powerful to these strange, irrational reasons. No, UK leaders didn’t spend millions upon millions on propaganda because they have a fragile identity. They did it because they’ll make money off of it, and will be able to move the legislation towards their own goals.

        It’s the same when people say Putin invaded Ukraine because he wants to restore the glory of the Soviet Union. No, he doesn’t care about any of that, he cares about staying in power and becoming more powerful. One of the best ways to do so is to invade other countries, as long as you don’t lose.

        • uis@lemm.ee
          link
          fedilink
          English
          arrow-up
          3
          ·
          1 month ago

          It’s the same when people say Putin invaded Ukraine because he wants to restore the glory of the Soviet Union. No, he doesn’t care about any of that, he cares about staying in power and becoming more powerful. One of the best ways to do so is to invade other countries, as long as you don’t lose.

          Thank you. I see so many people who don’t get it. I’m happy some people understand it without sending them link to one of few Ekaterina Shulman’s lectures in English.

          • FooBarrington@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 month ago

            Thank you for the validation, sometimes I feel like I’m going crazy with how often these things are repeated.

            But those lectures do sound interesting - would you mind linking them when you have the time?

            • uis@lemm.ee
              link
              fedilink
              English
              arrow-up
              3
              ·
              1 month ago

              This is not the lecture I originally intended to post. Also small correction for 1:00:02 first answer in poll should be translated as “social fainess”.

              If you find lecture where she says about “dealing with internal problems by external means” and “dropping concrete slab on nation’s head” - that is one I intended to link, but still searching which one it is.

      • TheGrandNagus@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 month ago

        if you live on an island for generations with limited new genetic input…well, thats where you end up.

        Literally the most diverse country in Europe lol

      • Maggoty@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 month ago

        Nah the core drivers wanted their own little neoliberal haven where they didn’t have to listen to the EU. They’d have been rich either way, but this way they get more power.

  • The Menemen!@lemmy.world
    link
    fedilink
    English
    arrow-up
    78
    ·
    edit-2
    1 month ago

    Even if she were the shoplifter, how would that work? “Sorry mate, you shoplifted when you were 16, now you can never buy food again.”?

  • Sarmyth@lemmy.world
    link
    fedilink
    English
    arrow-up
    78
    ·
    edit-2
    1 month ago

    Facial recognition still struggles with really bad mistakes that are always bad optics for the business that uses it. I’m amazed anyone is still willing to use it in its current form.

    It’s been the norm that these systems can’t tell the difference between people of dark pigmentation if it even acknowledges it’s seeing a person at all.

    Running a system with a decade long history or racist looking mistakes is bonkers in the current climate.

    • UnderpantsWeevil@lemmy.world
      link
      fedilink
      English
      arrow-up
      13
      ·
      edit-2
      1 month ago

      The catch is that its only really a problem for the people getting flagged. Then you’re guilty until proven innocent, and the only person to blame is a soulless machine with a big button that reads “For customer support, go fuck yourself”.

      As security theater, its cheap and easy to implement. As a passive income stream for tech bros, its a blank check. As a buzzword for politicians who can pretend they’re forward-thinking by rolling out some vaporware solution to a non-existent problem, its on the tip of everyone’s tongue.

      I’m old enough to remember when sitting Senator Ted Kennedy got flagged by the Bush Admin’s No Fly List and people were convinced this is the sort of shit that would reform the intrusive, incompetent, ill-conceived TSA. 20 years later… nope, it didn’t. Enjoy techno-hell, Brits.

    • Buddahriffic@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      1 month ago

      I’m curious how well the systems can differentiate doppelgangers and identical twins. Or if some makeup is enough to fool it.

      • PM_Your_Nudes_Please@lemmy.world
        link
        fedilink
        English
        arrow-up
        23
        ·
        1 month ago

        Facial recognition uses a few key elements of the face to hone in on matches, and traditional makeup doesn’t obscure any of those areas. In order to fool facial recognition, the goal is often to avoid face detection in the first place; Asymmetry, large contrasting colors, obscuring one (or both) eyes, hiding the oval head shape and jawline, and rhinestones (which sparkle and reflect light nearly randomly, making videos more confusing) seem to work well. But as neural nets improve, they also get harder to fool, so what works for one system may not work for every system.

        CV Dazzle (originally inspired by dazzle camouflage used on some warships) is a makeup style that tries to fool the most popular facial recognition systems.

        Note that those tend to obscure the bridge of the nose, the brow line, the jawline, etc… Because those are key identification areas for facial recognition.

        • Buddahriffic@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          1 month ago

          Yeah, if we can still recognize those as faces, it’s possible for a neural net to do so as well.

          But I’m talking more about differentiating faces than hiding entirely from such systems. Like makeup can be used to give the illusion that the shape of the face is different with false contour shading. You can’t really change the jawline (I think… I’m not skilled in makeup myself but have an awareness of what it can do) but you can change where the cheekbones appear to be, as well as the overall shape of the cheeks, and you can adjust the nose, too (unless it’s a profile angle).

          I think the danger in trying to hide that you have a face entirely is that if it gets detected, there’s a good chance that it will be flagged for attempting to fool the system because those examples you gave are pretty obvious, once you know what’s going on.

          It would be like going in to a bank with a ski mask to avoid being recognized vs going in as Mrs Doubtfire. Even if they are just trying to do banking that one time, the ski mask will attract unwanted attention while using a different face would accomplish the goal of avoiding attention.

  • Chaotic Entropy@feddit.uk
    link
    fedilink
    English
    arrow-up
    69
    ·
    edit-2
    1 month ago

    Oh good… janky oversold systems that do a lot of automation on a very shaky basis are also having high impacts when screwing up.

    Also “Facewatch” is such an awful sounding company.

  • Vipsu@lemmy.world
    link
    fedilink
    English
    arrow-up
    68
    ·
    1 month ago

    Can’t wait for something like this get hacked. There’ll be a lot explaining to do.

    • Jayjader@jlai.lu
      link
      fedilink
      English
      arrow-up
      28
      ·
      edit-2
      1 month ago

      Still, I think the only way that would result in change is if the hack specifically went after someone powerful like the mayor or one of the richest business owners in town.

  • afraid_of_zombies@lemmy.world
    link
    fedilink
    English
    arrow-up
    68
    arrow-down
    11
    ·
    1 month ago

    The developers should be looking at jail time as they falsely accused someone of commiting a crime. This should be treated exactly like if I SWATed someone.

    • Guest_User@lemmy.world
      link
      fedilink
      English
      arrow-up
      38
      arrow-down
      6
      ·
      1 month ago

      I get your point but totally disagree this is the same as SWATing. People can die from that. While this is bad, she was excluded from stores, not murdered

      • cheesepotatoes@lemmy.world
        link
        fedilink
        English
        arrow-up
        28
        arrow-down
        3
        ·
        1 month ago

        You lack imagination. What happens when the system mistakenly identifies someone as a violent offender and they get tackled by a bunch of cops, likely resulting in bodily injury.

          • PM_Your_Nudes_Please@lemmy.world
            link
            fedilink
            English
            arrow-up
            20
            ·
            1 month ago

            I mean, the article points out that the lady in the headline isn’t the only one who has been affected; A dude was falsely detained by cops after they parked a facial recognition van on a street corner, and grabbed anyone who was flagged.

          • blusterydayve26@midwest.social
            link
            fedilink
            English
            arrow-up
            18
            ·
            edit-2
            1 month ago

            That’s not very reassuring, we’re still only one computer bug away from that situation.

            Presumably she wasn’t identified as a violent criminal because the facial recognition system didn’t associate her duplicate with that particular crime. The system would be capable of associating any set of crimes with a face. It’s not like you get a whole new face for each different possible crime. So, we’re still one computer bug away from seeing that outcome.

          • cheesepotatoes@lemmy.world
            link
            fedilink
            English
            arrow-up
            7
            ·
            edit-2
            1 month ago

            No, it wouldn’t be. The base circumstance is the same, the software misidentifying a subject. The severity and context will vary from incident to incident, but the root cause is the same - false positives.

            There’s no process in place to prevent something like this going very very bad. It’s random chance that this time was just a false positive for theft. Until there’s some legislative obligation (such as legal liability) in place to force the company to create procedures and processes for identifying and reviewing false positives, then it’s only a matter of time before someone gets hurt.

            You don’t wait for someone to die before you start implementing safety nets. Or rather, you shouldn’t.

      • FiniteBanjo@lemmy.today
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        2
        ·
        1 month ago

        In the UK at least a SWATing would be many many times more deadly and violent than a normal police interaction. Can’t make the same argument for the USA or Russia, though.

    • ipkpjersi@lemmy.ml
      link
      fedilink
      English
      arrow-up
      21
      arrow-down
      1
      ·
      edit-2
      1 month ago

      I’m not so sure the blame should solely be placed on the developers - unless you’re using that term colloquially.

      • IllNess@infosec.pub
        link
        fedilink
        English
        arrow-up
        19
        arrow-down
        1
        ·
        1 month ago

        Developers were probably the first people to say that it isn’t ready. Blame the sales people that will say anything for money.

        • yetAnotherUser@feddit.de
          link
          fedilink
          English
          arrow-up
          3
          ·
          1 month ago

          It’s impossible to have a 0% false positive rate, it will never be ready and innocent people will always be affected. The only way to have a 0% false positive rate is with the following algorithm:

          def is_shoplifter(face_scan):
          return False

          • Zagorath@aussie.zone
            link
            fedilink
            English
            arrow-up
            5
            ·
            1 month ago
            line 2
                return False
                ^^^^^^
            IndentationError: expected an indented block after function definition on line 1
            
            • yetAnotherUser@feddit.de
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 month ago

              Weird, for me the indentation renders correctly. Maybe because I used Jerboa and single ticks instead of triple ticks?

              • Zagorath@aussie.zone
                link
                fedilink
                English
                arrow-up
                2
                ·
                edit-2
                1 month ago

                Interesting. This is certainly not the first time there have been markdown parsing inconsistencies between clients on Lemmy, the most obvious example being subscript and superscript, especially when ~multiple words~ ^get used^ or you use ^reddit ^style ^(superscript text).

                But yeah, checking just now on Jerboa you’re right, it does display correctly the way you did it. I first saw it on the web in lemmy-ui, which doesn’t display it properly, unless you use the triple backticks.

        • afraid_of_zombies@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          6
          ·
          1 month ago

          They worked on it, they knew what could happen. I could face criminal charges if I do certain things at work that harm the public.

          • IllNess@infosec.pub
            link
            fedilink
            English
            arrow-up
            7
            arrow-down
            2
            ·
            1 month ago

            I have no idea where Facewatch got their software from. The developers of this software could’ve been told their software will be used to find missing kids. Not really fair to blame developers. Blame the people on top.

    • Lith@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      8
      ·
      1 month ago

      Actually this one feels pretty similar to watch_dogs. Wasn’t this the plot to watch_dogs 2?

      • sfxrlz@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        edit-2
        1 month ago

        Now I’m interested in the plot of watch dogs 2…

        Edit: it’s indeed the plot of watch dogs 2

        https://en.m.wikipedia.org/wiki/Watch_Dogs_2

        In 2016, three years after the events of Chicago, San Francisco becomes the first city to install the next generation of ctOS – a computing network connecting every device together into a single system, developed by technology company Blume. Hacker Marcus Holloway, punished for a crime he did not commit through ctOS 2.0 …

    • erwan@lemmy.ml
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      2
      ·
      1 month ago

      If she’s been flagged as shoplifter she’s probably black!

      • feedum_sneedson@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        edit-2
        1 month ago

        Why, because all shoplifters are black? I don’t understand. She’s being mistaken for another person, a real person on the system.

        I used to know a smackhead that would steal things to order, I wonder if he’s still alive and whether he’s on this database. Never bought anything off him but I did buy him a drink occasionally. He’d had a bit of a difficult childhood.

        • Dudewitbow@lemmy.zip
          link
          fedilink
          English
          arrow-up
          4
          ·
          1 month ago

          i think its more that all facial recognition systems have a harder time picking up faces with darker complexion (same idea why phone cameras are bad at it for example). this failure leads to a bunch of false flags. its hilariously evident when you see police jurisdictions using it and aresting innocent people.

          not saying the womans of colored background, but would explain why it might be triggering.

        • erwan@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          1 month ago

          Because the algorithm trying to identify shoplifters is probably trained on biased dataset

          • feedum_sneedson@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 month ago

            That’s not how it works! It has mistakenly identified her as a specific individual who was already convicted of theft.

      • Konala Koala@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        2
        ·
        1 month ago

        Congratulations, you just identified yourself as a racist and need to understand you can’t just judge someone without first getting to know them.