СÀ¶ÊÓƵ

Skip to content

Our casual use of facial analysis tools can lead to more sinister applications

Facial recognition technologies have become more popular through increasingly sophisticated devices and popular fads. Casual use of face scanning and analysis features has long-term social impacts.
womanonphone
Face scanning technologies are more everyday than we might think.

On Dec. 14, the governments of British Columbia, Alberta and Québec ordered facial recognition company Clearview AI to . Discussions about the risks of facial recognition systems that rely on automated face analysis technologies tend to focus on corporations, national governments and law enforcement. But what is of great concern are the ways in which facial reognition and analysis have become integrated into our everyday lives.

Amazon, Microsoft and IBM have after disproportionately misidentifying .

Facebook and Clearview AI have dealt with and for building databases of billions of face templates without people’s consent.

In the United Kingdom, police face scrutiny for their use of . The Chinese government .

And yet, to grasp the scope and consequences of these systems we must also pay attention to the casual practices of everyday users who apply face scans and analysis in routine ways that contribute to the erosion of privacy and increase social discrimination and racism.

As a researcher of and their , I regularly explore how user actions can build or change norms around matters like privacy and identity. In this regard, adoption and use of face analysis systems and products in our everyday lives may be reaching a dangerous tipping point.

Everyday face scans

Open-source make face analysis or recognition an easy add-on for app developers. We already use facial recognition to unlock our phones or . Video cameras incorporated into smart homes use facial recognition to identify visitors as well as personalize screen displays and audio reminders. The auto-focus feature on cellphone cameras includes face detection and tracking, while cloud photo storage generates albums and themed slideshows by matching and grouping faces it recognizes in the images we make.

Face analysis is used by many apps including social media filters and accessories that produce effects like artificially ageing and animating facial features. Self-improvement and forecasting apps for beauty, horoscopes or ethnicity detection also generate advice and conclusions based on facial scans.

But using face analysis systems for horoscopes, selfies or identifying who’s on our front steps can have long-term societal consequences: they can and tracking, while .

Casual risks

When repeated over time, such low-stakes and quick-reward uses can inure us to face scanning more generally, opening the door to . We have no control over — and little insight into — who runs those systems and how the data is used.

If we already subject our faces to automated scrutiny, not only with our consent but also with our active participation, then being subjected to similar scans and analysis as we move through public spaces or access services might not seem particularly intrusive.

In addition, our personal use of face analysis technologies contributes directly to the development and implementation of larger systems meant for tracking populations, ranking clients or developing suspect pools for investigations. Companies can collect and share data that connects our images to our identities, or for .

Even if the platform we use restricts such uses, partner products may not abide by the same restrictions. The development of new databases of private individuals can be lucrative, especially when these can comprise multiple face images of each user or can associate images with identifying information, such as account names.

Pseudoscientific digital profiling

But perhaps most troubling, our growing embrace of facial analysis technologies feeds into how they not only determine an individual’s identity, but also their background, character and social value.

Many predictive and diagnostic apps that scan our faces to determine our ethnicity, beauty, wellness, emotions and even our potential earning power build on the disturbing historical pseudosciences of , and .

These depended to varying degrees on face analysis to justify racial hierarchies, colonization, chattel slavery, forced sterilization and preventative incarceration.

Our use of face analysis technologies can , implying they have a legitimate place in society. This complicity can then justify for uses such as or .

Building better habits

Regulating how facial recognition systems collect, interpret and distribute biometric data of face scanning and analysis. There has been some policy progress in and , but greater regulation is needed.

In addition, we need to confront our own habits and assumptions. How might we be putting ourselves and others, especially marginalized populations, at risk by making such machine-based scrutiny commonplace?

A few simple adjustments may help us address the creeping assimilation of facial analysis systems in our everyday lives. A good start is to change app and device settings to minimize scanning and sharing. Before downloading apps, research them and .

Resist the short-lived thrill of the latest social media face-effect fad — do we really need to know how we’d look as Pixar characters? Reconsider smart devices equipped with facial recognition technologies. Be aware of the rights of those whose image might be captured on a smart home device — you should always get explicit consent from anyone passing before the lens.

These small changes, if multiplied across users, products and platforms, can protect our data and buy time for greater reflection on the risks, benefits and fair deployment of facial recognition technologies.

Stephen Monteiro does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

push icon
Be the first to read breaking stories. Enable push notifications on your device. Disable anytime.
No thanks