Study finds popular face ID systems may have racial bias


Tech giants have made some major strides in advancing facial recognition technology. 

It’s now popping up in smartphones, laptops and tablets, all with the goal of making our lives easier. 

But a new study, called ‘Gender Shades,’ has found that it may not be working for all users, especially those who aren’t white males. 

A researcher from the MIT Media Lab discovered that popular facial recognition services from Microsoft, IBM and Face++ vary in accuracy based on gender and race.

To illustrate this, researcher Joy Buolamwini created a data set using 1,270 photos of parliamentarians from three African nations and three Nordic countries. 

The faces were selected to represent a broad range of human skin tones, using a labeling system developed by dermatologists, called the Fitzpatrick scale. 

The scale is viewed as more objective and precise than classifying based on race, according to the New York Times.

Buolamwini found that when the person in a photo was a white man, the facial recognition software worked 99% of the time. 

But when the photo was of a darker skinned woman, there was a nearly 35% error rate. 

For darker skinned men, their gender was misidentified roughly 12% of the time, while in a set of lighter skinned females, gender was misidentified about 7% of the time. 

Microsoft’s facial recognition systems failed to identify darker skinned women in 21% of cases, while IBM and Face++’s rates were almost 35%. All had error rates that were less than 1% for white skinned males.   

Face++, which is made by Megvii, a Chinese facial recognition startup, is used at Alibaba’s offices for employee ID, as well as in some Chinese public transportation systems.  

Microsoft, IBM and Face++ were analyzed in the study because they offered gender classification features in their facial identification software and their code was publicly available for testing, the Times noted.

The findings shed a light on how artificial intelligence technology is trained and evaluated, as well as the diversity of those who build the systems. 

For example, researchers at a major U.S. technology company claimed their facial-recognition systems were accurate more than 97% of the time, but the data set used to assess its performance was strikingly more than 77% male and more than 83% white, according to the study. 

Put simply, this suggests if more photos and data from white men are used in a certain data set, that can affect how accurately the system is able to identify other genders and races.

‘What’s really important here is the method and how that method applies to other applications,’ Buolamwini said. 

‘The same data-centric techniques that can be used to try to determine sombody’s gender are also used to identify a person when you’re looking for a criminal suspect or to unlock your phone,’ she added. 

The study goes on to say that an error in the output of facial-recognition algorithms can have ‘serious consequences.’

‘It has recently been shown that algorithms trained with biased data have resulted in algorithms trained with biased data have resulted in algorithmic discrimination,’ it notes.   

Buolamwini has now started to advocate for ‘algorithmic accountability,’ which is working to make automated decisions more transparent, explainable and fair, the Times said. 

It’s not the first time that racial inaccuracies have been detected in popular image and facial recognition services. 

In 2015, Google came under scrutiny after it was discovered that its photo-organizing service labeled African-Americans as gorillas. 

Major errors such as these in facial recognition have to be addressed before ‘commercial companies are to build genuinely fair, transparent and accountable facial analysis algorithms,’ the study notes.  

IBM has said it will take steps to improve its facial recognition systems.

‘We have a new model now that we brought out that is much more balanced in terms of accuracy across the benchmark that Joy was looking at,’ said Ruchir Puri, chief architect of IBM’s Watson AI system, in a statement. 

‘The model isn’t specifically a response to her paper, but we took it upon ourselves to address the questions she had raised directly, including her benchmark,’ he added. 

Microsoft has also said it will take steps to improve the accuracy of the firm’s facial recognition technology.  

Interestingly, the study didn’t analyze one facial recognition technology that’s become increasingly popular over the last few months: Apple’s Face ID. 

The tech giant began offering its Face ID facial recognition service in its high-end iPhone X last year. 

Users have duped Apple’s Face ID technology on several occasions, most recently two friends who discovered an iPhone X would unlock after scanning both their faces. 

But that type of facial recognition technology is different from Microsoft, IBM and Face++ services because it works to verify/identify a user, not classify them based on race or gender. 

Face ID scans a user’s face with infrared lasers to create a sample. Each time a user tries to unlock their phone, those same lasers scan the person’s face to make sure that it matches up.  



Leave A Reply