High-Tech Surveillance Amplifies Police Bias and Overreach

Today, in the midst of renewed outrage against structural racism and police brutality, and in the shadow of an even deeper economic recession, law enforcement organizations face the same temptation to adopt a technology-based solution to deep societal problems. Police chiefs are likely to want to turn the page from the current levels of community anger and distrust.

The Dangers of High-Tech Surveillance
Instead of repeating the mistakes of the past 12 years or so, communities have an opportunity to reject the expansion of big data policing. The dangers have only increased, the harms made plain by experience.

Those small startup companies that initially rushed into the policing business have been replaced by big technology companies with deep pockets and big ambitions.

Axon capitalized on the demands for police accountability after the protests in Ferguson and Baltimore to become a multimillion-dollar company providing digital services for police-worn body cameras. Amazon has been expanding partnerships with hundreds of police departments through its Ring cameras and Neighbors App. Other companies like BriefCam, Palantir and Shotspotter offer a host of video analyticssocial network analysis and other sensor technologies with the ability to sell technology cheaply in the short run with the hope for long term market advantage.

The technology itself is more powerful. The algorithmic models created a decade ago pale in comparison to machine learning capabilities today. Video camera streams have been digitized and augmented with analytics and facial recognition capabilities, turning static surveillance into a virtual time machine to find patterns in crowds. Adding to the data trap are smartphonessmart homes and smart cars, which now allow police to uncover individuals’ digital trails with relative ease.

The technology is more interconnected. One of the natural limiting factors of first-generation big data policing technology was the fact that it remained siloed. Databases could not communicate with one another. Data could not be easily shared. That limiting factor has shrunk as more aggregated data systems have been developed within government and by private vendors.

The promise of objective, unbiased technology didn’t pan out. Race bias in policing was not fixed by turning on a camera. Instead the technology created new problems, including highlighting the lack of accountability for high-profile instances of police violence.

Lessons for Reining in Police Spying
The harms of big data policing have been repeatedly exposed. Programs that attempted to predict individuals’ behaviors in Chicago and Los Angeles have been shut down after devastating audits cataloged their discriminatory impact and practical failure. Place-based predictive systems have been shut down in Los Angeles and other cities that initially had adopted the technology. Scandals involving facial recognitionsocial network analysis technology and large-scale sensor surveillance serve as a warning that technology cannot address the deeper issues of race, power and privacy that lie at the heart of modern-day policing.

The lesson of the first era of big data policing is that issues of race, transparency and constitutional rights must be at the forefront of design, regulation and use. Every mistake can be traced to a failure to see how the surveillance technology fits within the context of modern police power – a context that includes longstanding issues of racism and social control. Every solution points to addressing that power imbalance at the front end, through local oversight, community engagement and federal law, not after the technology has been adopted.

The debates about defundingdemilitarizing and reimagining existing law enforcement practices must include a discussion about police surveillance. There is a decade of missteps to learn from and era-defining privacy and racial justice challenges ahead. How police departments respond to the siren call of big data surveillance will reveal whether they’re on course to repeat the same mistakes.

Andrew Guthrie Ferguson is Professor of Law, American University. This article is published courtesy of The Conversation.