Most Organizations Don’t Care if Surveillance AI Is Biased

Nearly 2 out of 3 survey respondents say they either don’t care or aren’t sure if they care if their AI-powered security cameras are biased.

Most Organizations Don’t Care if Surveillance AI Is Biased

Photo via Adobe, by Alex

San Antonio, Texas — Pro-Vigil, a provider of remote video monitoring and crime deterrence solutions, has published a research report that found organizations are more concerned about their artificial intelligence (AI)-powered video surveillance system’s ability to deter crime than any potential bias issues.

Pro-Vigil surveyed 100 users of digital video surveillance across a variety of commercial vertical markets to gain an understanding of people’s knowledge of AI and how it’s being used in their video surveillance systems, as well as their opinions around AI bias.

The company found:

  • 62% of respondents said they either don’t care or aren’t sure if they care if their AI is biased.
  • When asked if they would do anything if their AI video system was doing a good job deterring crime, but was using unethical algorithms, more than one-third (37%) of respondents said they would do nothing.
  • Most survey respondents understood whether or not their video surveillance systems were using AI. Most (64%) indicated they weren’t using AI, while 21% said they were using AI. The rest were unsure.
  • 26% indicated there is a person in their organization who is responsible for understanding how AI is used. The rest either didn’t know or said there was no such person.
  • Nearly 90% said they would not know how to check to see if their AI video surveillance system was biased.

One specific area that has been been called out recently is facial recognition, which has increasingly come under fire over the past few years as the use of the technology has grown. A 2019 study found that Amazon’s facial-detection technology, Amazon Rekognition, was misidentifying women, especially those with darker skin. Overall, the technology was found to be racially bias.

This type of misidentification has led to major cities and states around the country banning the technology. Lawmakers have even introduced legislation to ban the federal government from utilizing facial recognition and similar technologies.

While AI and facial recognition have their benefits, as SIA likes to attest, Pro-Vigil’s survey reveals that users require more education on the technology. You can view the full research report here.


This article originally appeared in CS sister publication Security Sales & Integration. Steve Karantzoulidis is SSI’s web editor.

If you appreciated this article and want to receive more valuable industry content like this, click here to sign up for our FREE digital newsletters!

Leading in Turbulent Times: Effective Campus Public Safety Leadership for the 21st Century

This new webcast will discuss how campus public safety leaders can effectively incorporate Clery Act, Title IX, customer service, “helicopter” parents, emergency notification, town-gown relationships, brand management, Greek Life, student recruitment, faculty, and more into their roles and develop the necessary skills to successfully lead their departments. Register today to attend this free webcast!

Leave a Reply

Your email address will not be published. Required fields are marked *

Get Our Newsletters
Campus Safety Conference promo