top | item 44860000

Flock Now Using AI to Report to Police If Our Movement Patterns Are "Suspicious"

118 points| cyberphobe | 6 months ago |aclu.org

58 comments

order

username332211|6 months ago

I feel the ship has sailed here.

Once the police started to record every interaction with the public, along with their existing habit of placing traffic cameras left and right, they acquired enough data to track people.

Trying to restrict the analysis of existing data is never going to work. The police can always point to some death that wouldn't have happened, if they had ran Flock's software on their surveillance footage.

And even if by some miracle you manage to forbid plate recognition, cross referencing, etc, every ambitious (or lazy) detective would start doing it on the down low with OSS software.

FireBeyond|6 months ago

As an ex-Flock employee:

I was sold during the recruiting process on high ethics and morals and an idealistic vision.

The reality was a surveillance state, questionable policies on data sharing between agencies and private installations (HOA, etc.), and a CEO with a very literal belief that Flock should "eliminate all crime" - not "visionary" but far more literal. It was way too Minority Report for my liking.

They have a public "disclosure" site that supposedly shows the agencies using Flock that is absolutely inaccurate (there are three agencies in my County alone using it that are not listed there).

Any conversations about ethics and the other "should we even do this?" questions got consistently shorter and superficial during my time there.

potato3732842|6 months ago

> The police can always point to some death that wouldn't have happened

Why must it be this way?

Because we're a bunch of bitch-ass pansies unwilling to tell our fellow countrymen (and women) to shove it when they permit the use of such logic.

I don't care "how many children need to die" or whatever, the sum total of the affronts upon our freedom a is not worth it. What even is the point of caring about the children if we're giving them a totalitarian dump to inherit?

Ntrails|6 months ago

> Once the police started to record every interaction with the public

I don't think this is true? As far as I can tell any time the recording is mentioned in a complaint at the police behaviour the camera was off due to [battery life|maintenance|other].

Manuel_D|6 months ago

Courts have previously held that heuristics based determinations are not sufficient to serve as probable cause. E.g. "predictive policing" technologies can be used for e.g. scheduling officers to different areas, but aren't valid to conduct a search.

If this feature is used to make an arrest, there's a good chance the case would be thrown out.

genocidicbunny|6 months ago

The case can be thrown out, but it's still going to cause you massive disruptions. Everything from just being arrested in the first place and being held in custody for some amount of time, to having to hire a lawyer, to the social consequences of your name being tied to being arrested. It's going to cost you time, money, stress, family and social relationships. And there's a non-zero chance that if your life starts being investigated after such an arrest, something could be found to still affect you or your family and friends.

And once you're on their radar, you're probably going to also end up being marked for extra scrutiny. You might find yourself being pulled over more often, or getting the SSSS on your airplane boarding pass.

pjc50|6 months ago

https://www.msn.com/en-us/news/us/emiliano-s-agents-left-beh...

ICE are to a large extent above the law. Their entire purpose is to snatch people and move them to locations where they can be denied legal redress. A couple of high profile cases have only got redress due to very dedicated intervention by congresspeople, which does not scale.

I think people need to start reckoning with the underlying problem, which is that oppressive policing in America is popular provided it's happening to someone else.

ringeryless|6 months ago

I'm not sure such cases would be thrown out. See "parallel construction" for examples of illegally obtained data the DEA was advised to build an evidence chain NOT based on the illegally obtained info, but based upon evidence gleaned after the fact but built to show discovery during the course of investigation.

username332211|6 months ago

Who says the feature will be used to make an arrest?

The heuristics are clearly about who to pull over, etc. Evidence for arrest/search will be determined afterwards. And, as far as search is concerned, it could be as simple as getting a dog to bark.

garbagewoman|6 months ago

Why would the police even inform you that flock was used?

potato3732842|6 months ago

If this feature is used to make an arrest, there's a good chance the case would be thrown out.

They'll just parallel construct the crap out of it and get their arrest anyway.

ttemPumpinRary|6 months ago

But flock now has an Api for to Cause, the parallel construction AI. /s

So if they flock to the cause, all arrests are go. And there are always fallback crimes everyone in a modern society commits, that can be dragged in after a search .

ghssds|6 months ago

> With our new Multi-State Insights feature, law enforcement is alerted when suspect vehicles have been detected in multiple states

So, using our freedom of movement is now suspicious?

username332211|6 months ago

That's not what the sentence you cite is saying.

If they decide you are suspicious, they'll get an email alert about your location.

superultra|6 months ago

I live in the neighborhood where Flock started. The three Georgia Tech grads moved into a house in the West End in Atlanta. It’s a great neighborhood but like any urban neighborhood, you often deal with car break-in’s, so the roommates built a prototype security cam.

All fine so far. Except that the direction it was pointed at was the neighborhood middle school. Which means these three notably white college students started flock by surveilled predominantly black young kids.

The neighborhood was pissed - but what are you going to do?

Eventually Flock took off and they moved out.

My point is that if your product started as surveillance on not just another age demographic but a racial and class demographic, is it any surprise that all of this is fundamentally in the DNA of the company?

potato3732842|6 months ago

Even if you leave the race baiting out it's still wholly unacceptable.

Change the race of the parties up all you want and it doesn't change a thing.

saubeidl|6 months ago

Big Tech is building a dystopia and we all here are complicit. Become aware of the consequences of your actions and try to minimize the harm.

kentm|6 months ago

But, if we don’t build the Torment Nexus then someone else will anyway!

etiennebausson|6 months ago

Certainly not all no, maybe not even most.

I do not feel guilty of the unethical actions of others.

kotaKat|6 months ago

Grab a pole saw, cut em down, take a shit ton of free nice Tenergy battery packs and a SIM card with free data.

It’s time to snip the flock.

chung8123|6 months ago

Are there IR lights you can put on your license plate to block cameras from taking pictures of it?

FirmwareBurner|6 months ago

There are passive ways too on AliExpress like IR reflective sprays, coatings and films, but in my country, and I suspect in most of Europe, any intentional tampering with the legibility of your license plate is illegal and can land you hefty fines or even jail if caught.

troyvit|6 months ago

What would happen if you stenciled paintings of license-plate-like patterns all over the back of your car? Then you're not tampering with the plate itself, but I guess you end up with a goofy-lookin' car.

fennecfoxy|6 months ago

Sure, if you're doing a going to prison any% glitchless.

g42gregory|6 months ago

And you think OpenAI, Anthropic, and Co. will not do exactly the same, when you engage in a wrongthink?

Larry Ellison said it best a few months ago: "AI will keep you on your best behavior"[1].

[1] Meeting with Sam Altman, President Trump, and Masayoshi Son, announcing $500 million infrastructure deal.

rossant|6 months ago

Welcome to the future. Minority Report's coming. How terrifying.

collingreen|6 months ago

We're getting to the point where the thing making Minority report look unreasonable is that they had three (extremely drugged, troubled) humans making decisions as a consensus instead of just using an algorithm or a police officer's vibes. Nobody really doing these things today would slow down enough for that or risk someone having empathy and derailing the whole operation.

The last ten years have stuck out as a continuous loop of "this is so messed up it wouldn't be a believable/good movie plot".

blitzar|6 months ago

"They 'trust me'. Dumb fucks." - big tech / small tech / startups (every year)

thrown-0825|6 months ago

People ITT talking about probable cause like the rule of law still exists in the US