Feature KELVIN CHAN/AP/SHUTTERSTOCK KELVIN The Metropolitan Police in London used facial-recognition cameras to scan for wanted people in February.

Fussey and Murray listed a number of ethical and privacy concerns with the dragnet, and questioned whether it was legal at all. And they queried the accuracy of the system, which is sold by Tokyo-based technology giant NEC. The software flagged 42 people over 6 trials BEATING that the researchers analysed; officers dis- missed 16 matches as ‘non-credible’ but rushed out to stop the others. They lost 4 people in the crowd, but still stopped 22: only 8 turned out to be correct matches. The police saw the issue differently. They BIOMETRIC BIAS said the system’s number of false positives was tiny, considering the many thousands of faces Facial recognition is improving — but the bigger that had been scanned. (They didn’t reply to issue is how it’s used. By Davide Castelvecchi Nature’s requests for comment for this article.) The accuracy of facial recognition has improved drastically since ‘’ tech- hen London’s Metropolitan extracted key features and compared them niques were introduced into the field about a Police tested real-time to those of suspects from a watch list. “If there decade ago. But whether that means it’s good facial-recognition technology is a match, it pulls an image from the live feed, enough to be used on lower-quality, ‘in the wild’ between 2016 and 2019, they together with the image from the watch list.” images is a hugely controversial issue. And invited Daragh Murray to moni- Officers then reviewed the match and decided questions remain about how to transparently tor some of the trials from a con- whether to rush out to stop the ‘suspect’ and, evaluate facial-recognition systems. trol room inside a police van. occasionally, arrest them. In 2018, a seminal paper by computer scien- “It’s like you see in the Scotland Yard, as the headquarters of the tists Timnit Gebru, then at Research Wmovies,” says Murray, a legal scholar at the Uni- London police force is sometimes known, had in New York City and now at in Moun- versity of Essex in Colchester, UK. As cameras commissioned Murray and his University of tain View, California, and at the scanned passers-by in shopping centres or Essex colleague Pete Fussey, a sociologist, to Massachusetts Institute of Technology in Cam- public squares, they fed images to a computer conduct an independent study of its dragnet. bridge found that leading facial-recognition inside the van. Murray and police officers saw But their results1, published in July 2019, might software packages performed much worse at the software draw rectangles around faces not have been quite what the law-enforcement identifying the gender of women and people as it identified them in the live feed. It then agency had hoped for. of colour than at classifying male, white faces2.

Nature | Vol 587 | 19 November 2020 | 347 ©2020 Spri nger Nature Li mited. All ri ghts reserved.

Feature Concerns over demographic bias have since passport control at national borders. small group of vendors where false positives been quoted frequently in calls for morato- One measure of progress is the Face Recogni- based on demographic differentials were unde- riums or bans of facial-recognition software. tion Vendor Test, an independent benchmark- tectable”, but that match rates could be compro- In June, the world’s largest scientific com- ing assessment that the US National Institute of mised by outdoor, poorly lit or grainy images. puting society, the Association for Com- Standards and Technology (NIST) in Gaithers- puting Machinery in New York City, urged a burg, Maryland, has been conducting for two False faces suspension of private and government use decades. Dozens of laboratories, both com- One-to-one verification, such as recognizing of facial-recognition technology, because mercial and academic, have voluntarily taken the rightful owner of a passport or smart- of “clear bias based on ethnic, racial, gender, part in the latest round of testing, which began phone, has become extremely accurate; here, and other human characteristics”, which it in 2018 and is ongoing. NIST measures the per- is as skilful as the sharp- said injured the rights of individuals in spe- formance of each lab’s software package on its est-eyed humans. In this field, cutting-edge cific demographic groups. Axon, a maker of own image data sets, which include frontal and research focuses on detecting malevolent body cameras worn by police officers across profile police mugshots, and pictures scraped attacks. The first facial-recognition systems the , has said that facial recogni- from the Internet. (The US technology giants for unlocking phones, for example, were easily tion isn’t accurate enough to be deployed in Amazon, Apple, Google and Facebook have fooled by showing the phone a photo of the its products. Some US cities have banned the not taken part in the test.) owner, Jain says; 3D face recognition does bet- use of the technology in policing, and US law- In reports released late last year, the NIST ter. “Now the biggest challenge is very-high- makers have proposed a federal moratorium. team described massive steps forward in the quality face masks.” In one project, Jain and his Companies say they’re working to fix the technology’s performance during 2018, both collaborators are working on detecting such biases in their facial-recognition systems, and for one-to-many searches3 and for one-to-one impersonators by looking for skin texture. some are claiming success. But many research- verification4 (see also go.nature.com/35pku9q). But one-to-many verification, as Murray ers and activists are deeply sceptical. They argue “We have seen a significant improvement in found, isn’t so simple. With a large enough that even if the technology surpasses some face-recognition accuracy,” says Craig Watson, watch list, the number of false positives benchmark in accuracy, that won’t assuage an electrical engineer who leads NIST’s image flagged up can easily outweigh the true hits. deeper concerns that facial-recognition tools group. “We know that’s largely because of con- This is a problem when police must make are used in discriminatory ways. volutional neural networks,” he adds, a type of quick decisions about stopping someone. But deep neural network that is especially efficient mistakes also occur in slower investigations. More accurate but still biased at recognizing images. In January, Robert Williams was arrested at his Facial-recognition systems are often propri- The best algorithms can now identify peo- house in Farmington Hills, Michigan, after a etary and swathed in secrecy, but specialists ple from a profile image taken in the wild police facial-recognition system misidenti- say that most involve a multi-stage process — matching it with a frontal view from the fied him as a watch thief on the basis of blurry (see ‘How facial recognition works’) using deep database — about as accurately as the best surveillance footage of a Black man, which learning to train massive neural networks on it matched to his driving licence. The Amer- large sets of data to recognize patterns. “Every- “Systems are being brought ican Civil Liberties Union (ACLU), a non-profit body who does face recognition now uses deep organization in New York City, filed a com- learning,” says Anil Jain, a computer scientist to the wild without a plaint about the incident to Detroit police in at Michigan State University in East Lansing. proper evaluation of their June, and produced a video in which Williams The first stage in a typical system locates performance.” recounts what happened when a detective one or more faces in an image. Faces in the feed showed him the surveillance photos on paper. from a surveillance camera might be viewed in a “I picked that paper up, held it next to my face range of lighting conditions and from different facial-recognition software from a decade and said, ‘This is not me. I hope y’all don’t think angles, making them harder to recognize than ago could recognize frontal images, NIST all Black people look alike.’ And then he said: in a standard passport photo, for instance. The found. Recognizing a face in profile “has been ‘The computer says it’s you,’” Williams said. He algorithm will have been trained on millions a long-sought milestone in face recognition was released after being detained for 30 hours. of photos to locate ‘landmarks’ on a face, such research”, the NIST researchers wrote. ACLU attorney Phil Mayor says the technology as the eyes, nose and mouth, and it distils the But NIST also confirmed what Buolamwini should be banned. “It doesn’t work, and even information into a compact file, ranging from and Gebru’s gender-classification work sug- when it does work, it remains too dangerous less than 100 bytes to a few kilobytes in size. gested: most packages tended to be more a tool for governments to use to surveil their The next task is to ‘normalize’ the face, artifi- accurate for white, male faces than for peo- own citizens for no compelling return,” he says. cially rotating it into a frontal, well-illuminated ple of colour or for women5. In particular, Shortly after the ACLU complaint, Detroit view. This produces a set of facial ‘features’ faces classified in NIST’s database as African police chief James Craig acknowledged that the that can be compared with those extracted American or Asian were 10–100 times more software, if used by itself, would misidentify from an existing database of faces. This will likely to be misidentified than those classified cases “96% of the time”. Citing concerns over typically consist of pictures taken under con- as white. False positives were also more likely racial bias and discrimination, at least 11 US cit- trolled conditions, such as police mugshots. for women than for men. ies have banned facial recognition by public Because the feature representations are com- This inaccuracy probably reflects imbal- authorities in the past 18 months. But Detroit pact, structured files, a computer can quickly ances in the composition of each company’s police still use the technology. In late 2019, the scan millions of them to find the closest match. training database, Watson says — a scourge force adopted policies to ban live camera sur- Matching faces to a large database — called that data scientists often describe as ‘gar- veillance and to use the software only on still one-to-many identification — is one of two bage in, garbage out’. Still, discrepancies var- images and as part of criminal investigations; main types of facial-recognition system. The ied between packages, indicating that some Williams was arrested before the policy went other is one-to-one verification, the relatively companies might have begun to address the into practice, Craig said in June. (He did not simple task of making sure that a person problem, he adds. respond to Nature’s requests for comment.) matches their own photo. It can be applied NEC, which supplies Scotland Yard’s software, Other aspects of facial analysis, such as to anything from unlocking a smartphone to noted that in NIST’s analysis, it was “among a trying to deduce someone’s personality on

348 | Nature | Vol 587 | 19 November 2020 ©2020 Spri nger Nature Li mited. All ri ghts reserved. ©2020 Spri nger Nature Li mited. All ri ghts reserved.

the basis of their facial expressions, are even HOW FACIAL RECOGNITION WORKS more controversial. Researchers have shown Facial-recognition systems analyse a face’s geometry to create a faceprint — a biometric marker that can be used this doesn’t work6 — even the best software can to recognize or identify a person. Another use is facial analysis, which tries to classify a face according to labels such as gender, age, ethnicity or emotion. only be trained on images tagged with other people’s guesses. But companies around the world are still buying unproven technology that assesses interview candidates’ person- alities on the basis of videos of them talking. Nuria Oliver, a computer scientist based in Alicante, Spain, says that governments should regulate the use of facial recognition and other Capture image Locate face Locate landmarks: Align into frontal, potentially useful technologies to prevent eyes, nose and mouth well-lit view. Extract abuses (see page 350). “Systems are being geometrical features brought to the wild without a proper evalua- tion of their performance, or a process of veri- fication and reproducibility,” says Oliver, who Facial recognition Facial analysis is co-founder and vice-president of a regional network called the European Laboratory for Verify Identify Classify Learning and Intelligent Systems. Confirm that faceprint Check against database Infer human-defined matches stored example to discover identity characteristics Persistent problems (one-to-one comparison) (one-to-many comparison) Some proposals for regulation have called for authorities to establish accuracy standards ✖ input: and require that humans review any algo- rithm’s conclusions. But a standard based on, say, passing NIST benchmarks is much too low ✖ analysing... result: a bar on its own to justify deploying the tech- { nology, says , a technology fellow ✖ age: 56 gender: female in Ottawa with the Internet foundation Mozilla sentiment: happy who specializes in auditing facial-recognition confidence: 8 systems. ✔ } This year, Raji, Buolamwini, Gebru and others Examples: Examples: Examples: published another paper on the performance of Unlock a smartphone Scan crowd until ‘hit’ found Assess person’s age or gender Travel through a passport gate against watch list Assess person’s emotional state commercial systems, and noted that although Verify school or work attendance Match person of interest against (tests are controversial and less some firms had improved at classifying gender vast database reliable than facial recognition). across lighter- and darker-skinned faces, they were still worse at guessing a person’s age from of how to prevent it from being misused. The lot more sceptical of claims that you need faces with darker skin7. “The assessment process technology will improve, Murray says, but ever-more-precise forms of public surveil- is incredibly immature. Every time we under- doubts will remain over the legitimacy of lance,” says Kate Crawford, a computer scien- stand a new dimension to evaluate, we find out operating a permanent dragnet on innocent tist at New York University and co-director of that the industry is not performing at the level people, and over the criteria by which people the AI Now Institute. In August 2019, Crawford that it thought it did,” Raji says. It is important, are put on a watch list. called for a moratorium on governments’ use she says, that companies disclose more about Concerns about privacy, ethics and human of facial-recognition algorithms (K. Crawford how they test and train their facial-recognition rights will grow. The world’s largest biometric Nature 572, 565; 2019). systems, and consult with the communities in programme, in India, involves using facial Meanwhile, having declared its pilot project which the technology will be used. recognition to build a giant national ID card sys- a success, Scotland Yard announced in January Technical standards cannot stop tem called Aadhaar. Anyone who lives in India that it would begin to deploy live facial recog- facial-recognition systems from being used can go to an Aadhaar centre and have their pic- nition across London. in discriminatory ways, says Amba Kak, a legal ture taken. The system compares the photo with scholar at New York University’s AI Now Insti- existing records on 1.3 billion people to make Davide Castelvecchi reports for Nature from tute. “Are these systems going to be another sure the applicant hasn’t already registered London. Additional reporting by Antoaneta tool to propagate endemic discriminatory under a different name. “It’s a mind-boggling Roussi and Richard Van Noorden. practices in policing?” Human operators often system,” says Jain, who has been a consultant 1. Fussey, P. & Murray, D. Independent Report on the end up confirming a system’s biases rather for it. “The beauty of it is, it ensures one person London Metropolitan Police Service’s Trial of Live Facial than correcting it, Kak adds. Studies such as has only one ID.” But critics say it turns non-card Recognition Technology (Univ. Essex, 2019). the Scotland Yard external review show that owners into second-class citizens, and some 2. Buolamwini, J. & Gebru, T. Proc. Mach. Learn. Res. 81, 77–91 (2018). humans tend to overestimate the technology’s allege it was used to purge legitimate citizens 3. Grother, P., Ngan, M. & Hanaoka, K. Face Recognition credibility, even when they see the computer’s from voter rolls ahead of elections. Vendor Test (FRVT) Part 2: Identification (NIST, 2019). false match next to the real face. “Just putting And the most notorious use of biometric 4. Grother, P., Ngan, M. & Hanaoka, K. Ongoing Face in a clause ‘make sure there is a human in the technology is the surveillance state set up by Recognition Vendor Test (FRVT) Part 1: Verification. Updated 10 September 2020 (NIST, 2020). loop’ is not enough,” she says. the Chinese government in the Xinjiang prov- 5. Grother, P., Ngan, M. & Hanaoka, K. Face Recognition Kak and others support a moratorium on ince, where facial-recognition algorithms are Vendor Test Part 3: Demographic Effects (NIST, 2019). any use of facial recognition, not just because used to help single out and persecute people 6. Feldman Barrett, L., Adolphs, R., Marsella, S., Martinez, A. M. & Pollak, S. D. Psychol. Sci. Public Interest 20, 1–68 (2019). the technology isn’t good enough yet, but also from religious minorities (see page 354). 7. Raji, I. D. et al Preprint at https://arxiv.org/abs/2001.00964 because there needs to be a broader discussion “At this point in history, we need to be a (2020).

Nature | Vol 587 | 19 November 2020 | 349 ©2020 Spri nger Nature Li mited. All ri ghts reserved. ©2020 Spri nger Nature Li mited. All ri ghts reserved.