• SenorBolsa@beehaw.org
        link
        fedilink
        English
        arrow-up
        14
        ·
        edit-2
        1 year ago

        it will just hand out random DUIs to RAM owners based on statistics and reckless driving tickets to black altimas with tinted windows and bedazzled plate frames.

      • Quokka@quokk.au
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        That would be great.

        Imagine an AI that’s morally blind, pulling over expensive cars because the owners are more likely to break traffic laws knowing they can afford the fine.

      • lemillionsocks@beehaw.org
        link
        fedilink
        arrow-up
        26
        ·
        1 year ago

        We’ve already got numerous examples of how these ai models and face recognition models tend to have biases or are fed data that accidentally has a racial bias. Its not a stretch of the imagination to see how this can go wrong.

        • Scrubbles@poptalk.scrubbles.tech
          link
          fedilink
          arrow-up
          7
          ·
          1 year ago

          Yep, the age old “garbage in garbage out”. If we had a perfect track record we could just send in all the cop data, but we know for a fact the poor and PoC are stopped more than others. You send that into AI it will learn those same biases

      • boonhet@lemm.ee
        link
        fedilink
        arrow-up
        12
        ·
        1 year ago

        No, but they’re disproportionately affected by fines. For rich people a fine is just the cost of a privilege.

        Exception being something like Finnish speeding fines which are income dependent.

      • MyTurtleSwimsUpsideDown@kbin.social
        link
        fedilink
        arrow-up
        7
        ·
        edit-2
        1 year ago

        Not that I’m aware of.
        But we know the criminal justice system currently has biases. If the data the “AI” is trained on was affected by these biases, or others that we don’t realize, then it will produce biased results.

        • Pigeon@beehaw.org
          link
          fedilink
          arrow-up
          4
          ·
          1 year ago

          One of the worst parts about all this to me is that the AI and the dataset used to trained it are kept secret as proprietary information, and the police and governments buy it anyway despite that nobody can even try to check the code or dataset to see what biases or errors it might have (and definitely does).

  • lemillionsocks@beehaw.org
    link
    fedilink
    English
    arrow-up
    27
    ·
    1 year ago

    Wonder how long before these things start automatically mailing you a ticket because they calculated your speed between camera stops and determined you were going 3-5mph over the limit on the highway and if the “well it was used to catch a criminal” crowd will be as accepting of this technology then

    • Xkdrxodrixkr@feddit.de
      link
      fedilink
      English
      arrow-up
      19
      ·
      1 year ago

      The section control technology youre refering to is already being used in some countries, and they ofcourse automatically send you your speeding tickets. 3-5 mph would only be a marginal fine tho, if there even would be a fine at all. I dont understand though how that would be any different from normal speed checks, except for the fact that it might be more accurate.

      • jmp242@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        6
        ·
        1 year ago

        The problem in the US is they can’t give moving violations without someone there to testify. Usually that’s the officer. If the officer doesn’t show up, the ticket is tossed. I’m not really sure why they can give redlight tickets (unclear if that held up or not), but some of it had to do with if it was something that affected your license, or was a “violation” instead of a crime like a parking ticket.

  • money_loo@1337lemmy.com
    link
    fedilink
    English
    arrow-up
    13
    ·
    1 year ago

    Oh no, the A.I. identified someone as a drug trafficker, and the police pulled that person over on suspicion of being a drug trafficker, and found out that he was indeed a drug trafficker, and now he’s upset he got caught by a robot dragnet.

    I don’t think drugs should be criminalized, but are we supposed to be upset that A.I. is going to finally help parse data and solve crimes?

        • lobelia581@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          24
          ·
          1 year ago

          There’s a difference between using AI as a tool and using it as a solution. Though knowing how this society works, it’ll start off as a tool like now, and soon enough the higher ups will wonder why humans are even necessary in the process, especially when they need to be paid, and against everyone else’s objections they’ll get rid of the human verification part and use only AI, and when things go wrong the people in charge will go “who could have seen it coming?”

          • money_loo@1337lemmy.com
            link
            fedilink
            English
            arrow-up
            9
            ·
            1 year ago

            Sounds like oversight, transparency, and regulation are in order. There’s no putting this genie back in the bottle, unfortunately.

          • SSUPII@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 year ago

            Wasn’t there recently the news of the suicide outline that fired its staff for generative AI chat bots until not even a few days later it started giving dangerous responses?

      • money_loo@1337lemmy.com
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        McDonalds and White Castle have already begun using ALPR to tailor drive-through experiences, detecting returning customers and using past orders to guide them through the ordering process

        Yeah you’re right, helping people order lunch is literally 1984.

        • Programmer Belch@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          20
          ·
          1 year ago

          Uh oh, the things you are buying look pretty suspicious, we are going to wiretap your house to make sure you are not doing any no-nos.

          I really dislike people using technology to analyze my habits, I prefer just stumbling onto things I like because they were around the places I usually look. Yes, I also don’t like algorithmic content, it just makes people try to appease the algorithm, meaning less effort into the thing they do

    • webghost0101@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      44
      ·
      edit-2
      1 year ago

      This time they where right because it was indeed a drugsdealer but just look at what it took to get this data

      “in this case it was used to examine the driving patterns of anyone passing one of Westchester County’s 480 cameras over a two-year period.”

      “the AI determined that Zayas’ car was on a journey typical of a drug trafficker. According to a Department of Justice prosecutor filing, it made nine trips from Massachusetts to different parts of New York between October 2020 and August 2021 following routes known to be used by narcotics pushers and for conspicuously short stays.”

      So apparently making long trips with short stays is now enough proof to be searched by police. And if they can extrapolate that into “this guys a dealer” how much other data and possible extrapolations got caught in the crossfire off all those cameras. How long till someone in power decides selling some of that info to corporations is a good way to line state/government/own pockets?

      Maybe we should place cameras in everyones house and listening bugs in every single phone? Criminality solved? Or hear me out, the real criminals will adept, find new and novel ways while the common citizen is kept in line with fines for even the smallest offense.

      Then the police state will want to escalate the tools again, even more suppressing technology. Good thing were spend so much resources continuing to bully normal citizens into generating cash flow from fines. Money and resources well spend?

      Or maybe the world needs some actually intelligent people that can find the root causes of criminal behavior and restructure society to improve well being and chances so people want to belong and maintain it rather then feeling like the system of opportunities is rigged against them so they should cheat to survive.

    • Catoblepas@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      34
      ·
      1 year ago

      Acknowledging this is a dragnet, a practice generally considered unconstitutional since the 1950s, actually illustrates pretty well why people are upset about it. Even if it would result in more easy prosecutions for cops, it doesn’t change that it’s mass surveillance and an unconstitutional practice.

    • MagicShel@programming.dev
      link
      fedilink
      English
      arrow-up
      33
      ·
      1 year ago

      This pattern might indicate drugs. Or adultery, which isn’t illegal. It could be a straight job such as a mobile MRI technician. It might be a landlord.

      In short this is likely to affect innocent people. It’s like if you’ve got a name that happens to be on the no fly list, right? Your travel is fucked and you haven’t done anything wrong.

      • money_loo@1337lemmy.com
        link
        fedilink
        English
        arrow-up
        5
        ·
        1 year ago

        Worry, what…that you’ll get pulled over for suspicions while being innocent, and then the cops would be forced to find out you’re innocent?

        Yeah I can see that being inconvenient to downright dangerous depending on the cop, personally I think it has the potential to do more good than bad.

        It’s just a shame the police have such a hard on for drugs when there’s so much worse stuff going on out there.

          • money_loo@1337lemmy.com
            link
            fedilink
            English
            arrow-up
            5
            ·
            1 year ago

            Wait until you grow up a bit and learn about probable cause, or people getting pulled over for nothing. This tech changes nothing about your idea of “innocent until proven guilty” vs what actually happens on the daily.

            • elfpie@beehaw.org
              link
              fedilink
              English
              arrow-up
              20
              ·
              1 year ago

              I think that’s exactly the point. The current situation is already bad, tools that reinforce the bad part of the system shouldn’t be accepted.

              • money_loo@1337lemmy.com
                link
                fedilink
                English
                arrow-up
                1
                ·
                1 year ago

                It’s not the point. You guys are discussing pipe dreams and impossible scenarios. I’m just trying to be pragmatic about what’s happening.

                Did you miss the part in the article about this tech being run by private companies? Or how it’s so seamless it can be installed on ANY existing camera system? No upgrades to hardware required.

                This surveillance genie is already out of the bottle so just hoping it will go away or be made suddenly illegal in a country that has had a hard-on for surveillance on its own people since at least 9/11 is foolish.

                So the best we can do is hope for proper laws regulating and controlling it, so it doesn’t turn into all of those evil things everyone always wants to jump to first.

                Because ACAB, but recognize that police work is also one of the areas that could really benefit from AI technology, they are constantly flooded with information from all sorts of sources and it leads to ridiculous backlogs that actually affect society.

                So yeah, this tech, like all tech, has the potential to do great harm to society if not reigned in, but it also has the potential to help find your child after they’ve been abducted, or locate your wife or mom after someone has attacked them.

                It could do legitimate good in society, too, if used correctly.

        • Dr Cog@mander.xyz
          link
          fedilink
          English
          arrow-up
          18
          ·
          1 year ago

          And a certain percentage of innocent people are found guilty. You don’t see how expanding the arrests of innocent people is a bad thing? It has the potential to ruin lives

        • 👁️🫦👁️@lemm.ee
          link
          fedilink
          English
          arrow-up
          7
          ·
          1 year ago

          Big “if you’re innocent, you’ve got nothing to hide” energy. I can be innocent and still not want unnecessary interaction with police and appreciate my privacy

  • Magnus@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    6
    ·
    1 year ago

    Wait so if you drive like a crazy person after a few beers and a good ol’smash up with the lads then you are suddenly a criminal?

    • nzodd@beehaw.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      I do my best driving after a few beers.

      Granted, that’s mostly because I have a much easier time hitting pedestrians when I’m sober.

  • nieceandtows@programming.dev
    link
    fedilink
    English
    arrow-up
    6
    ·
    edit-2
    1 year ago

    Wake me when the AI can identify people who are going to commit crimes before they commit the crimes.

    Edit: it’s a precog reference from Minority Report, btw.