Businesses Are Harvesting Our Biometric Data. The Public Needs Assurances on Security
One of the core challenges is the issue of consent. How can the public ever give informed consent if they are constantly monitored by cameras and unaware of who is storing and using their biometric data?
This fundamental problem underscores the difficulty in resolving privacy concerns. Businesses face the daunting task of obtaining clear, informed consent from people who might not even know they are being observed.
Without transparent practices and explicit consent mechanisms, it’s nearly impossible to ensure that the public is truly aware of and agrees to the use of their biometric data.
Think about your digital security. If your password gets stolen, you can change it. If your credit card is compromised, you can cancel it. But your face? That’s permanent. Biometric data is incredibly sensitive because it cannot be altered once it’s compromised. This makes it a high-stakes game when it comes to security.
If a database is breached, hackers could misuse this data for identity theft, fraud, or even harassment.
Another issue is algorithmic bias and discrimination. If data is used for decision-making, how can companies ensure that diverse and sufficient data is included to train the algorithm?
Companies might use biometric data for authentication, personalized marketing, employee monitoring and access control. There is a significant risk of gender and racial biases if the algorithm is primarily trained on data from a homogenous group, such as white males.
Companies should also be ensuring that digital bias is not perpetuated. Failing to do so may lead to societal inequalities.
Legislation and Awareness
As facial recognition becomes more common, the need for robust legislation becomes urgent. Laws must mandate clear consent before capturing anyone’s biometric data. They should also set strict standards for how this data is stored and secured to prevent breaches.
It’s equally crucial that the public becomes more aware of the issue. While people are becoming more conscious about data protection, facial recognition often flies under the radar. It’s invisible in our everyday lives, and many don’t realize the risks and ethical issues. Educating the public is vital.
Incorporating the principles of responsible AI into the deployment of facial recognition technology would be a good place to start. Responsible AI emphasizes fairness, accountability, transparency and ethics. This means that AI systems, including facial recognition, should be designed and used in ways that respect human rights, privacy and dignity.
However, businesses might not necessarily prioritize these principles if they are not being held accountable by regulatory bodies or the public.
Transparency is a cornerstone of responsible AI. If organizations using facial recognition remain secretive about their practices, we cannot trust them with our biometric data.
Companies armed with only your personal information can be very powerful in terms of manipulative marketing. It takes only “one like” for bespoke campaigns to target you very accurately.
But now, political parties such as the PTI in Pakistan have embraced vision-AI technology to allow leader Imran Khan to campaign despite serving a prison sentence.
Visual data capturing and analysis are particularly critical compared to non-visual data because they provide richer, more intimate and more immediate insights into human behavior and identity.
That’s why its growing use by businesses raises so many concerns about privacy and consent. While the public remains unaware of the extent to which their visual data is being captured and utilized, their information will be vulnerable to misuse or exploitation.
Kamran Mahroof isAssociate Professor, Supply Chain Analytics, University of Bradford. Amizan Omar is Associate Professor of Strategic Management, University of Bradford. Irfan Mehmood is Associate Professor in Business Analytics, University of Bradford. This article is published courtesy of The Conversation.