Recent from talks
Nothing was collected or created yet.
Facial recognition system
View on WikipediaThis article's factual accuracy may be compromised due to out-of-date information. (June 2022) |


A facial recognition system[1] is a technology potentially capable of matching a human face from a digital image or a video frame against a database of faces. Such a system is typically employed to authenticate users through ID verification services, and works by pinpointing and measuring facial features from a given image.[2]
Development began on similar systems in the 1960s, beginning as a form of computer application. Since their inception, facial recognition systems have seen wider uses in recent times on smartphones and in other forms of technology, such as robotics. Because computerized facial recognition involves the measurement of a human's physiological characteristics, facial recognition systems are categorized as biometrics. Although the accuracy of facial recognition systems as a biometric technology is lower than iris recognition, fingerprint image acquisition, palm recognition or voice recognition, it is widely adopted due to its contactless process.[3] Facial recognition systems have been deployed in advanced human–computer interaction, video surveillance, law enforcement, passenger screening, decisions on employment and housing and automatic indexing of images.[4][5]
Facial recognition systems are employed throughout the world today by governments and private companies.[6] Their effectiveness varies, and some systems have previously been scrapped because of their ineffectiveness. The use of facial recognition systems has also raised controversy, with claims that the systems violate citizens' privacy, commonly make incorrect identifications, encourage gender norms[7][8] and racial profiling,[9] and do not protect important biometric data. The appearance of synthetic media such as deepfakes has also raised concerns about its security.[10] These claims have led to the ban of facial recognition systems in several cities in the United States.[11] Growing societal concerns led social networking company Meta Platforms to shut down its Facebook facial recognition system in 2021, deleting the face scan data of more than one billion users.[12][13] The change represented one of the largest shifts in facial recognition usage in the technology's history. IBM also stopped offering facial recognition technology due to similar concerns.[14]
History of facial recognition technology
[edit]Automated facial recognition was pioneered in the 1960s by Woody Bledsoe, Helen Chan Wolf, and Charles Bisson, whose work focused on teaching computers to recognize human faces.[15] Their early facial recognition project was dubbed "man-machine" because a human first needed to establish the coordinates of facial features in a photograph before they could be used by a computer for recognition. Using a graphics tablet, a human would pinpoint facial features coordinates, such as the pupil centers, the inside and outside corners of eyes, and the widows peak in the hairline. The coordinates were used to calculate 20 individual distances, including the width of the mouth and of the eyes. A human could process about 40 pictures an hour, building a database of these computed distances. A computer would then automatically compare the distances for each photograph, calculate the difference between the distances, and return the closed records as a possible match.[15]
In 1970, Takeo Kanade publicly demonstrated a face-matching system that located anatomical features such as the chin and calculated the distance ratio between facial features without human intervention. Later tests revealed that the system could not always reliably identify facial features. Nonetheless, interest in the subject grew and in 1977 Kanade published the first detailed book on facial recognition technology.[16]
In 1993, the Defense Advanced Research Project Agency (DARPA) and the Army Research Laboratory (ARL) established the face recognition technology program FERET to develop "automatic face recognition capabilities" that could be employed in a productive real life environment "to assist security, intelligence, and law enforcement personnel in the performance of their duties." Face recognition systems that had been trialled in research labs were evaluated. The FERET tests found that while the performance of existing automated facial recognition systems varied, a handful of existing methods could viably be used to recognize faces in still images taken in a controlled environment.[17] The FERET tests spawned three US companies that sold automated facial recognition systems. Vision Corporation and Miros Inc were founded in 1994, by researchers who used the results of the FERET tests as a selling point. Viisage Technology was established by an identification card defense contractor in 1996 to commercially exploit the rights to the facial recognition algorithm developed by Alex Pentland at MIT.[18]
Following the 1993 FERET face-recognition vendor test, the Department of Motor Vehicles (DMV) offices in West Virginia and New Mexico became the first DMV offices to use automated facial recognition systems to prevent people from obtaining multiple driving licenses using different names. Driver's licenses in the United States were at that point a commonly accepted form of photo identification. DMV offices across the United States were undergoing a technological upgrade and were in the process of establishing databases of digital ID photographs. This enabled DMV offices to deploy the facial recognition systems on the market to search photographs for new driving licenses against the existing DMV database.[19] DMV offices became one of the first major markets for automated facial recognition technology and introduced US citizens to facial recognition as a standard method of identification.[20] The increase of the US prison population in the 1990s prompted U.S. states to established connected and automated identification systems that incorporated digital biometric databases, in some instances this included facial recognition. In 1999, Minnesota incorporated the facial recognition system FaceIT by Visionics into a mug shot booking system that allowed police, judges and court officers to track criminals across the state.[21]


Until the 1990s, facial recognition systems were developed primarily by using photographic portraits of human faces. Research on face recognition to reliably locate a face in an image that contains other objects gained traction in the early 1990s with the principal component analysis (PCA). The PCA method of face detection is also known as Eigenface and was developed by Matthew Turk and Alex Pentland.[22] Turk and Pentland combined the conceptual approach of the Karhunen–Loève theorem and factor analysis, to develop a linear model. Eigenfaces are determined based on global and orthogonal features in human faces. A human face is calculated as a weighted combination of a number of Eigenfaces. Because few Eigenfaces were used to encode human faces of a given population, Turk and Pentland's PCA face detection method greatly reduced the amount of data that had to be processed to detect a face. Pentland in 1994 defined Eigenface features, including eigen eyes, eigen mouths and eigen noses, to advance the use of PCA in facial recognition. In 1997, the PCA Eigenface method of face recognition[23] was improved upon using linear discriminant analysis (LDA) to produce Fisherfaces.[24] LDA Fisherfaces became dominantly used in PCA feature based face recognition. While Eigenfaces were also used for face reconstruction. In these approaches no global structure of the face is calculated which links the facial features or parts.[25]
Purely feature based approaches to facial recognition were overtaken in the late 1990s by the Bochum system, which used Gabor filter to record the face features and computed a grid of the face structure to link the features.[26] Christoph von der Malsburg and his research team at the University of Bochum developed Elastic Bunch Graph Matching in the mid-1990s to extract a face out of an image using skin segmentation.[22] By 1997, the face detection method developed by Malsburg outperformed most other facial detection systems on the market. The so-called "Bochum system" of face detection was sold commercially on the market as ZN-Face to operators of airports and other busy locations. The software was "robust enough to make identifications from less-than-perfect face views. It can also often see through such impediments to identification as mustaches, beards, changed hairstyles and glasses—even sunglasses".[27]
Real-time face detection in video footage became possible in 2001 with the Viola–Jones object detection framework for faces.[28] Paul Viola and Michael Jones combined their face detection method with the Haar-like feature approach to object recognition in digital images to launch AdaBoost, the first real-time frontal-view face detector.[29] By 2015, the Viola–Jones algorithm had been implemented using small low power detectors on handheld devices and embedded systems. Therefore, the Viola–Jones algorithm has not only broadened the practical application of face recognition systems but has also been used to support new features in user interfaces and teleconferencing.[30]
Ukraine is using the US-based Clearview AI facial recognition software to identify dead Russian soldiers. Ukraine has conducted 8,600 searches and identified the families of 582 deceased Russian soldiers. The IT volunteer section of the Ukrainian army using the software is subsequently contacting the families of the deceased soldiers to raise awareness of Russian activities in Ukraine. The main goal is to destabilise the Russian government. It can be seen as a form of psychological warfare. About 340 Ukrainian government officials in five government ministries are using the technology. It is used to catch spies that might try to enter Ukraine.[31]
Clearview AI's facial recognition database is only available to government agencies who may only use the technology to assist in the course of law enforcement investigations or in connection with national security.[32]
The software was donated to Ukraine by Clearview AI. Russia is thought to be using it to find anti-war activists. Clearview AI was originally designed for US law enforcement. Using it in war raises new ethical concerns. One London based surveillance expert, Stephen Hare, is concerned it might make the Ukrainians appear inhuman: "Is it actually working? Or is it making [Russians] say: 'Look at these lawless, cruel Ukrainians, doing this to our boys'?"[33]
Techniques for face recognition
[edit]
While humans can recognize faces without much effort,[34] facial recognition is a challenging pattern recognition problem in computing. Facial recognition systems attempt to identify a human face, which is three-dimensional and changes in appearance with lighting and facial expression, based on its two-dimensional image. To accomplish this computational task, facial recognition systems perform four steps. First face detection is used to segment the face from the image background. In the second step the segmented face image is aligned to account for face pose, image size and photographic properties, such as illumination and grayscale. The purpose of the alignment process is to enable the accurate localization of facial features in the third step, the facial feature extraction. Features such as eyes, nose and mouth are pinpointed and measured in the image to represent the face. The so established feature vector of the face is then, in the fourth step, matched against a database of faces.[35]
Traditional
[edit]
Some face recognition algorithms identify facial features by extracting landmarks, or features, from an image of the subject's face. For example, an algorithm may analyze the relative position, size, and/or shape of the eyes, nose, cheekbones, and jaw.[36] These features are then used to search for other images with matching features.[37]
Other algorithms normalize a gallery of face images and then compress the face data, only saving the data in the image that is useful for face recognition. A probe image is then compared with the face data.[38] One of the earliest successful systems[39] is based on template matching techniques[40] applied to a set of salient facial features, providing a sort of compressed face representation.
Recognition algorithms can be divided into two main approaches: geometric, which looks at distinguishing features, or photo-metric, which is a statistical approach that distills an image into values and compares the values with templates to eliminate variances. Some classify these algorithms into two broad categories: holistic and feature-based models. The former attempts to recognize the face in its entirety while the feature-based subdivide into components such as according to features and analyze each as well as its spatial location with respect to other features.[41]
Popular recognition algorithms include principal component analysis using eigenfaces, linear discriminant analysis, elastic bunch graph matching using the Fisherface algorithm, the hidden Markov model, the multilinear subspace learning using tensor representation, and the neuronal motivated dynamic link matching.[citation needed][42] Modern facial recognition systems make increasing use of machine learning techniques such as deep learning.[43]
Human identification at a distance (HID)
[edit]To enable human identification at a distance (HID) low-resolution images of faces are enhanced using face hallucination. In CCTV imagery faces are often very small. But because facial recognition algorithms that identify and plot facial features require high resolution images, resolution enhancement techniques have been developed to enable facial recognition systems to work with imagery that has been captured in environments with a high signal-to-noise ratio. Face hallucination algorithms that are applied to images prior to those images being submitted to the facial recognition system use example-based machine learning with pixel substitution or nearest neighbour distribution indexes that may also incorporate demographic and age related facial characteristics. Use of face hallucination techniques improves the performance of high resolution facial recognition algorithms and may be used to overcome the inherent limitations of super-resolution algorithms. Face hallucination techniques are also used to pre-treat imagery where faces are disguised. Here the disguise, such as sunglasses, is removed and the face hallucination algorithm is applied to the image. Such face hallucination algorithms need to be trained on similar face images with and without disguise. To fill in the area uncovered by removing the disguise, face hallucination algorithms need to correctly map the entire state of the face, which may be not possible due to the momentary facial expression captured in the low resolution image.[44]
3-dimensional recognition
[edit]
Three-dimensional face recognition technique uses 3D sensors to capture information about the shape of a face. This information is then used to identify distinctive features on the surface of a face, such as the contour of the eye sockets, nose, and chin.[45] One advantage of 3D face recognition is that it is not affected by changes in lighting like other techniques. It can also identify a face from a range of viewing angles, including a profile view.[45][37] Three-dimensional data points from a face vastly improve the precision of face recognition. 3D-dimensional face recognition research is enabled by the development of sophisticated sensors that project structured light onto the face.[46] 3D matching technique are sensitive to expressions, therefore researchers at Technion applied tools from metric geometry to treat expressions as isometries.[47] A new method of capturing 3D images of faces uses three tracking cameras that point at different angles; one camera will be pointing at the front of the subject, second one to the side, and third one at an angle. All these cameras will work together so it can track a subject's face in real-time and be able to face detect and recognize.[48]
Thermal cameras
[edit]
A different form of taking input data for face recognition is by using thermal cameras, by this procedure the cameras will only detect the shape of the head and it will ignore the subject accessories such as glasses, hats, or makeup.[49] Unlike conventional cameras, thermal cameras can capture facial imagery even in low-light and nighttime conditions without using a flash and exposing the position of the camera.[50] However, the databases for face recognition are limited. Efforts to build databases of thermal face images date back to 2004.[49] By 2016, several databases existed, including the IIITD-PSE and the Notre Dame thermal face database.[51] Current thermal face recognition systems are not able to reliably detect a face in a thermal image that has been taken of an outdoor environment.[52]
In 2018, researchers from the U.S. Army Research Laboratory (ARL) developed a technique that would allow them to match facial imagery obtained using a thermal camera with those in databases that were captured using a conventional camera.[53] Known as a cross-spectrum synthesis method due to how it bridges facial recognition from two different imaging modalities, this method synthesize a single image by analyzing multiple facial regions and details.[54] It consists of a non-linear regression model that maps a specific thermal image into a corresponding visible facial image and an optimization issue that projects the latent projection back into the image space.[50] ARL scientists have noted that the approach works by combining global information (i.e. features across the entire face) with local information (i.e. features regarding the eyes, nose, and mouth).[55] According to performance tests conducted at ARL, the multi-region cross-spectrum synthesis model demonstrated a performance improvement of about 30% over baseline methods and about 5% over state-of-the-art methods.[54]
Application
[edit]Social media
[edit]Founded in 2013, Looksery went on to raise money for its face modification app on Kickstarter. After successful crowdfunding, Looksery launched in October 2014. The application allows video chat with others through a special filter for faces that modifies the look of users. Image augmenting applications already on the market, such as Facetune and Perfect365, were limited to static images, whereas Looksery allowed augmented reality to live videos. In late 2015 SnapChat purchased Looksery, which would then become its landmark lenses function.[56] Snapchat filter applications use face detection technology and on the basis of the facial features identified in an image a 3D mesh mask is layered over the face.[57] A variety of technologies attempt to fool facial recognition software by the use of anti-facial recognition masks.[58]
DeepFace is a deep learning facial recognition system created by a research group at Facebook. It identifies human faces in digital images. It employs a nine-layer neural net with over 120 million connection weights, and was trained on four million images uploaded by Facebook users.[59][60] The system is said to be 97% accurate, compared to 85% for the FBI's Next Generation Identification system.[61]
TikTok's algorithm has been regarded as especially effective, but many were left to wonder at the exact programming that caused the app to be so effective in guessing the user's desired content.[62] In June 2020, TikTok released a statement regarding the "For You" page, and how they recommended videos to users, which did not include facial recognition.[63] In February 2021, however, TikTok agreed to a $92 million settlement to a US lawsuit which alleged that the app had used facial recognition in both user videos and its algorithm to identify age, gender and ethnicity.[64]
ID verification
[edit]The emerging use of facial recognition is in the use of ID verification services. Many companies and others are working in the market now to provide these services to banks, ICOs, and other e-businesses.[65] Face recognition has been leveraged as a form of biometric authentication for various computing platforms and devices;[37] Android 4.0 "Ice Cream Sandwich" added facial recognition using a smartphone's front camera as a means of unlocking devices,[66][67] while Microsoft introduced face recognition login to its Xbox 360 video game console through its Kinect accessory,[68] as well as Windows 10 via its "Windows Hello" platform (which requires an infrared-illuminated camera).[69] In 2017, Apple's iPhone X smartphone introduced facial recognition to the product line with its "Face ID" platform, which uses an infrared illumination system.[70]
Face ID
[edit]Apple introduced Face ID on the flagship iPhone X as a biometric authentication successor to the Touch ID, a fingerprint based system. Face ID has a facial recognition sensor that consists of two parts: a "Romeo" module that projects more than 30,000 infrared dots onto the user's face, and a "Juliet" module that reads the pattern.[71] The pattern is sent to a local "Secure Enclave" in the device's central processing unit (CPU) to confirm a match with the phone owner's face.[72]
The facial pattern is not accessible by Apple. The system will not work with eyes closed, in an effort to prevent unauthorized access.[72] The technology learns from changes in a user's appearance, and therefore works with hats, scarves, glasses, and many sunglasses, beard and makeup.[73] It also works in the dark. This is done by using a "Flood Illuminator", which is a dedicated infrared flash that throws out invisible infrared light onto the user's face to get a 2d picture in addition to the 30,000 facial points.[74]
Healthcare
[edit]Facial recognition algorithms can help in diagnosing some diseases using specific features on the nose, cheeks and other part of the human face.[75] Relying on developed data sets, machine learning has been used to identify genetic abnormalities just based on facial dimensions.[76] FRT has also been used to verify patients before surgery procedures.
In March, 2022 according to a publication by Forbes, FDNA, an AI development company claimed that in the space of 10 years, they have worked with geneticists to develop a database of about 5,000 diseases and 1500 of them can be detected with facial recognition algorithms.[77]
Deployment of FRT for availing government services
[edit]India
[edit]In an interview, the National Health Authority chief Dr. R.S. Sharma said that facial recognition technology would be used in conjunction with Aadhaar to authenticate the identity of people seeking vaccines.[78] Ten human rights and digital rights organizations and more than 150 individuals signed a statement by the Internet Freedom Foundation that raised alarm against the deployment of facial recognition technology in the central government's vaccination drive process.[79] Implementation of an error-prone system without adequate legislation containing mandatory safeguards, would deprive citizens of essential services and linking this untested technology to the vaccination roll-out in India will only exclude persons from the vaccine delivery system.[80]
In July, 2021, a press release by the Government of Meghalaya stated that facial recognition technology (FRT) would be used to verify the identity of pensioners to issue a Digital Life Certificate using "Pensioner's Life Certification Verification" mobile application.[81] The notice, according to the press release, purports to offer pensioners "a secure, easy and hassle-free interface for verifying their liveness to the Pension Disbursing Authorities from the comfort of their homes using smart phones". Mr. Jade Jeremiah Lyngdoh, a law student, sent a legal notice to the relevant authorities highlighting that "The application has been rolled out without any anchoring legislation which governs the processing of personal data and thus, lacks lawfulness and the Government is not empowered to process data."[82]
Deployment in security services
[edit]
Commonwealth
[edit]The Australian Border Force and New Zealand Customs Service have set up an automated border processing system called SmartGate that uses face recognition, which compares the face of the traveller with the data in the e-passport microchip.[83][84] All Canadian international airports use facial recognition as part of the Primary Inspection Kiosk program that compares a traveler face to their photo stored on the ePassport. This program first came to Vancouver International Airport in early 2017 and was rolled up to all remaining international airports in 2018–2019.[85]
Police forces in the United Kingdom have been trialing live facial recognition technology at public events since 2015.[86] In May 2017, a man was arrested using an automatic facial recognition (AFR) system mounted on a van operated by the South Wales Police. Ars Technica reported that "this appears to be the first time [AFR] has led to an arrest".[87] However, a 2018 report by Big Brother Watch found that these systems were up to 98% inaccurate.[86] The report also revealed that two UK police forces, South Wales Police and the Metropolitan Police, were using live facial recognition at public events and in public spaces.[88] In September 2019, South Wales Police use of facial recognition was ruled lawful.[88] Live facial recognition has been trialled since 2016 in the streets of London and will be used on a regular basis from Metropolitan Police from beginning of 2020.[89] In August 2020 the Court of Appeal ruled that the way the facial recognition system had been used by the South Wales Police in 2017 and 2018 violated human rights.[90]
However, by 2024 the Metropolitan Police were using the technique with a database of 16,000 suspects, leading to over 360 arrests, including rapists and someone wanted for grievous bodily harm for 8 years. They claim a false positive rate of only 1 in 6,000. The photos of those not identified by the system are deleted immediately.[91]
United States
[edit]
The U.S. Department of State operates one of the largest face recognition systems in the world with a database of 117 million American adults, with photos typically drawn from driver's license photos.[92] Although it is still far from completion, it is being put to use in certain cities to give clues as to who was in the photo. The FBI uses the photos as an investigative tool, not for positive identification.[93] As of 2016,[update] facial recognition was being used to identify people in photos taken by police in San Diego and Los Angeles (not on real-time video, and only against booking photos)[94] and use was planned in West Virginia and Dallas.[95]
In recent years Maryland has used face recognition by comparing people's faces to their driver's license photos. The system drew controversy when it was used in Baltimore to arrest unruly protesters after the death of Freddie Gray in police custody.[96] Many other states are using or developing a similar system however some states have laws prohibiting its use.
The FBI has also instituted its Next Generation Identification program to include face recognition, as well as more traditional biometrics like fingerprints and iris scans, which can pull from both criminal and civil databases.[97] The federal Government Accountability Office criticized the FBI for not addressing various concerns related to privacy and accuracy.[98]
Starting in 2018, U.S. Customs and Border Protection deployed "biometric face scanners" at U.S. airports. Passengers taking outbound international flights can complete the check-in, security and the boarding process after getting facial images captured and verified by matching their ID photos stored on CBP's database. Images captured for travelers with U.S. citizenship will be deleted within up to 12-hours. The Transportation Security Administration (TSA) had expressed its intention to adopt a similar program for domestic air travel during the security check process in the future. The American Civil Liberties Union is one of the organizations against the program, concerning that the program will be used for surveillance purposes.[99]
In 2019, researchers reported that Immigration and Customs Enforcement (ICE) uses facial recognition software against state driver's license databases, including for some states that provide licenses to undocumented immigrants.[98]
In December 2022, 16 major domestic airports in the US started testing facial-recognition tech where kiosks with cameras are checking the photos on travelers' IDs to make sure that passengers are not impostors.[100] In 2025, it was revealed that the New Orleans Police Department had rolled out what the ACLU's Freed Wessler called "the first known widespread effort by police in a major US city to use AI to identify people in live camera feeds for the purpose of making immediate arrests." in defiance of a 2022 city ordinance limiting the use of the technology.[101]
China
[edit]In 2006, the "Skynet" (天網))Project was initiated by the Chinese government to implement CCTV surveillance nationwide and as of 2018,[update] there have been 20 million cameras, many of which are capable of real-time facial recognition, deployed across the country for this project.[102] Some official claim that the current Skynet system can scan the entire Chinese population in one second and the world population in two seconds.[103]

In 2017, the Qingdao police was able to identify twenty-five wanted suspects using facial recognition equipment at the Qingdao International Beer Festival, one of which had been on the run for 10 years.[104] The equipment works by recording a 15-second video clip and taking multiple snapshots of the subject. That data is compared and analyzed with images from the police department's database and within 20 minutes, the subject can be identified with a 98.1% accuracy.[105]
In 2018, Chinese police in Zhengzhou and Beijing were using smart glasses to take photos which are compared against a government database using facial recognition to identify suspects, retrieve an address, and track people moving beyond their home areas.[106][107]
As of late 2017,[update] China has deployed facial recognition and artificial intelligence technology in Xinjiang. Reporters visiting the region found surveillance cameras installed every hundred meters or so in several cities, as well as facial recognition checkpoints at areas like gas stations, shopping centers, and mosque entrances.[108][109] In May 2019, Human Rights Watch reported finding Face++ code in the Integrated Joint Operations Platform (IJOP), a police surveillance app used to collect data on, and track the Uighur community in Xinjiang.[110] Human Rights Watch released a correction to its report in June 2019 stating that the Chinese company Megvii did not appear to have collaborated on IJOP, and that the Face++ code in the app was inoperable.[111] In February 2020, following the Coronavirus outbreak, Megvii applied for a bank loan to optimize the body temperature screening system it had launched to help identify people with symptoms of a Coronavirus infection in crowds. In the loan application Megvii stated that it needed to improve the accuracy of identifying masked individuals.[112]
Many public places in China are implemented with facial recognition equipment, including railway stations, airports, tourist attractions, expos, and office buildings. In October 2019, a professor at Zhejiang Sci-Tech University sued the Hangzhou Safari Park for abusing private biometric information of customers. The safari park uses facial recognition technology to verify the identities of its Year Card holders. An estimated 300 tourist sites in China have installed facial recognition systems and use them to admit visitors. This case is reported to be the first on the use of facial recognition systems in China.[113] In August 2020, Radio Free Asia reported that in 2019 Geng Guanjun, a citizen of Taiyuan City who had used the WeChat app by Tencent to forward a video to a friend in the United States was subsequently convicted on the charge of the crime "picking quarrels and provoking troubles". The Court documents showed that the Chinese police used a facial recognition system to identify Geng Guanjun as an "overseas democracy activist" and that China's network management and propaganda departments directly monitor WeChat users.[114]
In 2019, Protestors in Hong Kong destroyed smart lampposts amid concerns they could contain cameras and facial recognition system used for surveillance by Chinese authorities.[115] Human rights groups have criticized the Chinese government for using artificial intelligence facial recognition technology in its suppression against Uyghurs,[116] Christians[117] and Falun Gong practitioners.[118][119]
India
[edit]Even though facial recognition technology (FRT) is not fully accurate,[120] it is being increasingly deployed for identification purposes by the police in India. FRT systems generate a probability match score, or a confidence score between the suspect who is to be identified and the database of identified criminals that is available with the police. The National Automated Facial Recognition System (AFRS)[121] is already being developed by the National Crime Records Bureau (NCRB), a body constituted under the Ministry of Home Affairs. The project seeks to develop and deploy a national database of photographs which would comport with a facial recognition technology system by the central and state security agencies. The Internet Freedom Foundation has flagged concerns regarding the project.[122] The NGO has highlighted that the accuracy of FRT systems are "routinely exaggerated and the real numbers leave much to be desired.[122] The implementation of such faulty FRT systems would lead to high rates of false positives and false negatives in this recognition process."
Under the Supreme Court of India's decision in Justice K.S. Puttaswamy vs Union of India (22017 10 SCC 1), any justifiable intrusion by the State into people's right to privacy, which is protected as a fundamental right under Article 21 of the Constitution, must confirm to certain thresholds, namely: legality, necessity, proportionality and procedural safeguards.[123] As per the Internet Freedom Foundation, the National Automated Facial Recognition System (AFRS) proposal fails to meet any of these thresholds, citing "absence of legality," "manifest arbitrariness," and "absence of safeguards and accountability."[124]
While the national level AFRS project is still in the works, police departments in various states in India are already deploying facial recognition technology systems, such as: TSCOP + CCTNS in Telangana,[125] Punjab Artificial Intelligence System (PAIS) in Punjab,[126] Trinetra in Uttar Pradesh,[127] Police Artificial Intelligence System in Uttarakhand,[128] AFRS in Delhi, Automated Multimodal Biometric Identification System (AMBIS) in Maharashtra, FaceTagr in Tamil Nadu. The Crime and Criminal Tracking Network and Systems (CCTNS), which is a Mission Mode Project under the National e-Governance Plan (NeGP),[129] is viewed as a system which would connect police stations across India, and help them "talk"[130] to each other. The project's objective is to digitize all FIR-related information, including FIRs registered, as well as cases investigated, charge sheets filed, and suspects and wanted persons in all police stations. This shall constitute a national database of crime and criminals in India. CCTNS is being implemented without a data protection law in place. CCTNS is proposed to be integrated with the AFRS, a repository of all crime and criminal related facial data which can be deployed to purportedly identify or verify a person from a variety of inputs ranging from images to videos.[131] This has raised privacy concerns from civil society organizations and privacy experts. Both the projects have been censured as instruments of "mass surveillance" at the hands of the state.[132] In Rajasthan, 'RajCop,' a police app has been recently integrated with a facial recognition module which can match the face of a suspect against a database of known persons in real-time. Rajasthan police is in currently working to widen the ambit of this module by making it mandatory to upload photographs of all arrested persons in CCTNS database, which will "help develop a rich database of known offenders."[133]
Helmets fixed with camera have been designed and being used by Rajasthan police in law and order situations to capture police action and activities of "the miscreants, which can later serve as evidence during the investigation of such cases."[133] PAIS (Punjab Artificial Intelligence System), App employs deep learning, machine learning, and face recognition for the identification of criminals to assist police personnel.[133] The state of Telangana has installed 8 lakh CCTV cameras,[133] with its capital city Hyderabad slowly turning into a surveillance capital.[134]
A false positive happens when facial recognition technology misidentifies a person to be someone they are not, that is, it yields an incorrect positive result. They often results in discrimination and strengthening of existing biases. For example, in 2018, Delhi Police reported that its FRT system had an accuracy rate of 2%, which sank to 1% in 2019. The FRT system even failed to distinguish accurately between different sexes.[135]
The government of Delhi in collaboration with Indian Space Research Organisation (ISRO) is developing a new technology called Crime Mapping Analytics and Predictive System (CMAPS). The project aims to deploy space technology for "controlling crime and maintaining law and order."[133] The system will be connected to a database containing data of criminals.[133] The technology is envisaged to be deployed to collect real-time data at the crime scene.[133]
In a reply dated November 25, 2020 to a Right to Information request filed by the Internet Freedom Foundation seeking information about the facial recognition system being used by the Delhi Police (with reference number DEPOL/R/E/20/07128),[136] the Office of the Deputy Commissioner of Police cum Public Information Officer: Crime stated that they cannot provide the information under section 8(d) of the Right to Information Act, 2005.[137] A Right to Information (RTI) request dated July 30, 2020 was filed with the Office of the Commissioner, Kolkata Police, seeking information about the facial recognition technology that the department was using.[138] The information sought was denied[139] stating that the department was exempted from disclosure under section 24(4) of the RTI Act.
Latin America
[edit]In the 2000 Mexican presidential election, the Mexican government employed face recognition software to prevent voter fraud. Some individuals had been registering to vote under several different names, in an attempt to place multiple votes. By comparing new face images to those already in the voter database, authorities were able to reduce duplicate registrations.[140]
In Colombia public transport busses are fitted with a facial recognition system by FaceFirst Inc to identify passengers that are sought by the National Police of Colombia. FaceFirst Inc also built the facial recognition system for Tocumen International Airport in Panama. The face recognition system is deployed to identify individuals among the travellers that are sought by the Panamanian National Police or Interpol.[141] Tocumen International Airport operates an airport-wide surveillance system using hundreds of live face recognition cameras to identify wanted individuals passing through the airport. The face recognition system was initially installed as part of a US$11 million contract and included a computer cluster of sixty computers, a fiber-optic cable network for the airport buildings, as well as the installation of 150 surveillance cameras in the airport terminal and at about 30 airport gates.[142]
At the 2014 FIFA World Cup in Brazil the Federal Police of Brazil used face recognition goggles. Face recognition systems "made in China" were also deployed at the 2016 Summer Olympics in Rio de Janeiro.[141] Nuctech Company provided 145 inspection terminals for Maracanã Stadium and 55 terminals for the Deodoro Olympic Park.[143]
European Union
[edit]Police forces in at least 21 countries of the European Union use, or plan to use, facial recognition systems, either for administrative or criminal purposes.[144]
Greece
[edit]Greek police passed a contract with Intracom-Telecom for the provision of at least 1,000 devices equipped with live facial recognition system. The delivery is expected before the summer 2021. The total value of the contract is over 4 million euros, paid for in large part by the Internal Security Fund of the European Commission.[145]
Italy
[edit]Italian police acquired a face recognition system in 2017, Sistema Automatico Riconoscimento Immagini (SARI). In November 2020, the Interior ministry announced plans to use it in real-time to identify people suspected of seeking asylum.[146]
The Netherlands
[edit]The Netherlands has deployed facial recognition and artificial intelligence technology since 2016.[147] The database of the Dutch police currently contains over 2.2 million pictures of 1.3 million Dutch citizens. This accounts for about 8% of the population. In The Netherlands, face recognition is not used by the police on municipal CCTV.[148]
South Africa
[edit]In South Africa, in 2016, the city of Johannesburg announced it was rolling out smart CCTV cameras complete with automatic number plate recognition and facial recognition.[149]
Deployment in retail stores
[edit]The US firm 3VR, now Identiv, is an example of a vendor which began offering facial recognition systems and services to retailers as early as 2007.[150] In 2012, the company advertised benefits such as "dwell and queue line analytics to decrease customer wait times", "facial surveillance analytic[s] to facilitate personalized customer greetings by employees" and the ability to "[c]reate loyalty programs by combining Point of sale (POS) data with facial recognition".[151]
United States
[edit]In 2018, the National Retail Federation Loss Prevention Research Council called facial recognition technology "a promising new tool" worth evaluating.[152]
In July 2020, the Reuters news agency reported that during the 2010s the pharmacy chain Rite Aid had deployed facial recognition video surveillance systems and components from FaceFirst, DeepCam LLC, and other vendors at some retail locations in the United States.[152] Cathy Langley, Rite Aid's vice president of asset protection, used the phrase "feature matching" to refer to the systems and said that usage of the systems resulted in less violence and organized crime in the company's stores, while former vice president of asset protection Bob Oberosler emphasized improved safety for staff and a reduced need for the involvement of law enforcement organizations.[152] In a 2020 statement to Reuters in response to the reporting, Rite Aid said that it had ceased using the facial recognition software and switched off the cameras.[152]
According to director Read Hayes of the National Retail Federation Loss Prevention Research Council, Rite Aid's surveillance program was either the largest or one of the largest programs in retail.[152] The Home Depot, Menards, Walmart, and 7-Eleven are among other US retailers also engaged in large-scale pilot programs or deployments of facial recognition technology.[152]
Of the Rite Aid stores examined by Reuters in 2020, those in communities where people of color made up the largest racial or ethnic group were three times as likely to have the technology installed,[152] raising concerns related to the substantial history of racial segregation and racial profiling in the United States. Rite Aid said that the selection of locations was "data-driven", based on the theft histories of individual stores, local and national crime data, and site infrastructure.[152]
Australia
[edit]In 2019, facial recognition to prevent theft was in use at Sydney's Star Casino and was also deployed at gaming venues in New Zealand.[153]
In June 2022, consumer group CHOICE reported facial recognition was in use in Australia at Kmart, Bunnings, and The Good Guys. The Good Guys subsequently suspended the technology pending a legal challenge by CHOICE to the Office of the Australian Information Commissioner, while Bunnings kept the technology in use and Kmart maintained its trial of the technology.[154]
Additional uses
[edit]
At the American football championship game Super Bowl XXXV in January 2001, police in Tampa Bay, Florida used Viisage face recognition software to search for potential criminals and terrorists in attendance at the event. 19 people with minor criminal records were potentially identified.[155][156]
Face recognition systems have also been used by photo management software to identify the subjects of photographs, enabling features such as searching images by person, as well as suggesting photos to be shared with a specific contact if their presence were detected in a photo.[157][158] By 2008 facial recognition systems were typically used as access control in security systems.[159]
The United States' popular music and country music celebrity Taylor Swift surreptitiously employed facial recognition technology at a concert in 2018. The camera was embedded in a kiosk near a ticket booth and scanned concert-goers as they entered the facility for known stalkers.[160]
On August 18, 2019, The Times reported that the UAE-owned Manchester City hired a Texas-based firm, Blink Identity, to deploy facial recognition systems in a driver program. The club has planned a single super-fast lane for the supporters at the Etihad stadium.[161] However, civil rights groups cautioned the club against the introduction of this technology, saying that it would risk "normalising a mass surveillance tool". The policy and campaigns officer at Liberty, Hannah Couchman said that Man City's move is alarming, since the fans will be obliged to share deeply sensitive personal information with a private company, where they could be tracked and monitored in their everyday lives.[162]
In 2019, casinos in Australia and New Zealand rolled out facial recognition to prevent theft, and a representative of Sydney's Star Casino said they would also provide 'customer service' like welcoming a patron back to a bar.[153]
In August 2020, amid the COVID-19 pandemic in the United States, American football stadiums of New York and Los Angeles announced the installation of facial recognition for upcoming matches. The purpose is to make the entry process as touchless as possible.[163] Disney's Magic Kingdom, near Orlando, Florida, likewise announced a test of facial recognition technology to create a touchless experience during the pandemic; the test was originally slated to take place between March 23 and April 23, 2021, but the limited timeframe had been removed as of late April 2021.[update][164]
Media companies have begun using face recognition technology to streamline their tracking, organizing, and archiving pictures and videos.[165]
Advantages and disadvantages
[edit]Compared to other biometric systems
[edit]In 2006, the performance of the latest face recognition algorithms was evaluated in the Face Recognition Grand Challenge (FRGC). High-resolution face images, 3-D face scans, and iris images were used in the tests. The results indicated that the new algorithms are 10 times more accurate than the face recognition algorithms of 2002 and 100 times more accurate than those of 1995. Some of the algorithms were able to outperform human participants in recognizing faces and could uniquely identify identical twins.[45][166]
One key advantage of a facial recognition system that it is able to perform mass identification as it does not require the cooperation of the test subject to work. Properly designed systems installed in airports, multiplexes, and other public places can identify individuals among the crowd, without passers-by even being aware of the system.[167] However, as compared to other biometric techniques, face recognition may not be most reliable and efficient. Quality measures are very important in facial recognition systems as large degrees of variations are possible in face images. Factors such as illumination, expression, pose and noise during face capture can affect the performance of facial recognition systems.[167] Among all biometric systems, facial recognition has the highest false acceptance and rejection rates,[167] thus questions have been raised on the effectiveness of or bias of face recognition software in cases of railway and airport security, law enforcement and housing and employment decisions.[168][5]
Weaknesses
[edit]Ralph Gross, a researcher at the Carnegie Mellon Robotics Institute in 2008, describes one obstacle related to the viewing angle of the face: "Face recognition has been getting pretty good at full frontal faces and 20 degrees off, but as soon as you go towards profile, there've been problems."[45] Besides the pose variations, low-resolution face images are also very hard to recognize. This is one of the main obstacles of face recognition in surveillance systems.[169] It has also been suggested that camera settings can favour sharper imagery of white skin than of other skin tones.[5]
Face recognition is less effective if facial expressions vary. A big smile can render the system less effective. For instance: Canada, in 2009, allowed only neutral facial expressions in passport photos.[170]
There is also inconstancy in the datasets used by researchers. Researchers may use anywhere from several subjects to scores of subjects and a few hundred images to thousands of images. Data sets may be diverse and inclusive or mainly contain images of white males. It is important for researchers to make available the datasets they used to each other, or have at least a standard or representative dataset.[171]
Although high degrees of accuracy have been claimed for some facial recognition systems, these outcomes are not universal. The consistently worst accuracy rate is for those who are 18 to 30 years old, Black and female.[5]
Racial bias and skin tone
[edit]Studies have shown that facial recognition algorithms tend to perform better on individuals with lighter skin tones compared to those with darker skin tones. This disparity arises primarily because training datasets often overrepresent lighter-skinned individuals, leading to higher error rates for darker-skinned people. For example, a 2018 study found that leading commercial gender classification models, which are facial recognition models, have an error rate up to 7 times higher for those with darker skin tones compared to those with lighter skin tones.[172]
Common image compression methods, such as JPEG chroma subsampling, have been found to disproportionately degrade performance for darker-skinned individuals. These methods inadequately represent color information, which adversely affects the ability of algorithms to recognize darker-skinned individuals accurately.[173]
Cross-race effect bias
[edit]Facial recognition systems often demonstrate lower accuracy when identifying individuals with non-Eurocentric facial features. Known as the Cross-race effect, this bias occurs when systems perform better on racial or ethnic groups that are overrepresented in their training data, resulting in reduced accuracy for underrepresented groups.[174] The overrepresented group is generally the more populous group in the location that the model is being developed. For example, models developed in Asian cultures generally perform better on Asian facial features than Eurocentric facial features due to overrepresentation in the developers training dataset. The opposite is observed in models developed in Eurocentric cultures.[175]
The systems used for facial recognition often lack the sufficient training needed to fully recognize those features not of Eurocentric descent. When the training and databases for these Machine Learning (ML) models do not contain a diverse representation, the models fail to identify the missed population, adding to their racial biases.[7]
The cross-race effect is not exclusive to machines; humans also experience difficulty recognizing faces from racial or ethnic groups different from their own. This is an example of inherent human biases being perpetuated in training datasets.[176]
Challenges for individuals with disabilities
[edit]Facial recognition technologies encounter significant challenges when identifying individuals with disabilities. For instance, systems have been shown to perform worse when recognizing individuals with Down syndrome, often leading to increased false match rates. This is due to distinct facial structures associated with the condition that are not adequately represented in training datasets.[177]
More broadly, facial recognition systems tend to overlook diverse physical characteristics related to disabilities. The lack of representative data for individuals with varying disabilities further emphasizes the need for inclusive algorithmic designs to mitigate bias and improve accuracy.[178]
Additionally, facial expression recognition technologies often fail to accurately interpret the emotional states of individuals with intellectual disabilities. This shortcoming can hinder effective communication and interaction, underscoring the necessity for systems trained on diverse datasets that include individuals with intellectual disabilities.[179]
Furthermore, biases in facial recognition algorithms can lead to discriminatory outcomes for people with disabilities. For example, certain facial features or asymmetries may result in misidentification or exclusion, highlighting the importance of developing accessible and fair biometric systems.[180]
Advancements in fairness and mitigation strategies
[edit]Efforts to address these biases include designing algorithms specifically for fairness. A notable study introduced a method to learn fair face representations by using a progressive cross-transformer model.[181] This approach highlights the importance of balancing accuracy across demographic groups while avoiding performance drops in specific populations.
Additionally, targeted dataset collection has been shown to improve racial equity in facial recognition systems. By prioritizing diverse data inputs, researchers demonstrated measurable reductions in performance disparities between racial groups.[177]
Ineffectiveness
[edit]Critics of the technology complain that the London Borough of Newham scheme has, as of 2004,[update] never recognized a single criminal, despite several criminals in the system's database living in the Borough and the system has been running for several years. "Not once, as far as the police know, has Newham's automatic face recognition system spotted a live target."[156][182] This information seems to conflict with claims that the system was credited with a 34% reduction in crime (hence why it was rolled out to Birmingham also).[183]
An experiment in 2002 by the local police department in Tampa, Florida, had similarly disappointing results.[156] A system at Boston's Logan Airport was shut down in 2003 after failing to make any matches during a two-year test period.[184]
In 2014, Facebook stated that in a standardized two-option facial recognition test, its online system scored 97.25% accuracy, compared to the human benchmark of 97.5%.[185]
Systems are often advertised as having accuracy near 100%; this is misleading as the outcomes are not universal.[5] The studies often use samples that are smaller and less diverse than would be necessary for large scale applications. Because facial recognition is not completely accurate, it creates a list of potential matches. A human operator must then look through these potential matches and studies show the operators pick the correct match out of the list only about half the time. This causes the issue of targeting the wrong suspect.[93][186]
Controversies
[edit]Privacy violations
[edit]Civil rights organizations and privacy campaigners such as the Electronic Frontier Foundation, Big Brother Watch and the ACLU express concern that privacy is being compromised by the use of surveillance technologies.[187][86][188] Face recognition can be used not just to identify an individual, but also to unearth other personal data associated with an individual – such as other photos featuring the individual, blog posts, social media profiles, Internet behavior, and travel patterns.[189] Concerns have been raised over who would have access to the knowledge of one's whereabouts and people with them at any given time.[190] Moreover, individuals have limited ability to avoid or thwart face recognition tracking unless they hide their faces. This fundamentally changes the dynamic of day-to-day privacy by enabling any marketer, government agency, or random stranger to secretly collect the identities and associated personal information of any individual captured by the face recognition system.[189] Consumers may not understand or be aware of what their data is being used for, which denies them the ability to consent to how their personal information gets shared.[190]
In July 2015, the United States Government Accountability Office conducted a Report to the Ranking Member, Subcommittee on Privacy, Technology and the Law, Committee on the Judiciary, U.S. Senate. The report discussed facial recognition technology's commercial uses, privacy issues, and the applicable federal law. It states that previously, issues concerning facial recognition technology were discussed and represent the need for updating the privacy laws of the United States so that federal law continually matches the impact of advanced technologies. The report noted that some industry, government, and private organizations were in the process of developing, or have developed, "voluntary privacy guidelines". These guidelines varied between the stakeholders, but their overall aim was to gain consent and inform citizens of the intended use of facial recognition technology. According to the report the voluntary privacy guidelines helped to counteract the privacy concerns that arise when citizens are unaware of how their personal data gets put to use.[190]
In 2016, Russian company NtechLab caused a privacy scandal in the international media when it launched the FindFace face recognition system with the promise that Russian users could take photos of strangers in the street and link them to a social media profile on the social media platform Vkontakte (VK).[191] In December 2017, Facebook rolled out a new feature that notifies a user when someone uploads a photo that includes what Facebook thinks is their face, even if they are not tagged. Facebook has attempted to frame the new functionality in a positive light, amidst prior backlashes.[192] Facebook's head of privacy, Rob Sherman, addressed this new feature as one that gives people more control over their photos online. "We've thought about this as a really empowering feature," he says. "There may be photos that exist that you don't know about."[193] Facebook's DeepFace has become the subject of several class action lawsuits under the Biometric Information Privacy Act, with claims alleging that Facebook is collecting and storing face recognition data of its users without obtaining informed consent, in direct violation of the 2008 Biometric Information Privacy Act (BIPA).[194] The most recent case was dismissed in January 2016 because the court lacked jurisdiction.[195] In the US, surveillance companies such as Clearview AI are relying on the First Amendment to the United States Constitution to data scrape user accounts on social media platforms for data that can be used in the development of facial recognition systems.[196]
In 2019, the Financial Times first reported that facial recognition software was in use in the King's Cross area of London.[197] The development around London's King's Cross mainline station includes shops, offices, Google's UK HQ and part of St Martin's College. According to the UK Information Commissioner's Office: "Scanning people's faces as they lawfully go about their daily lives, in order to identify them, is a potential threat to privacy that should concern us all."[198][199] The UK Information Commissioner Elizabeth Denham launched an investigation into the use of the King's Cross facial recognition system, operated by the company Argent. In September 2019 it was announced by Argent that facial recognition software would no longer be used at King's Cross. Argent claimed that the software had been deployed between May 2016 and March 2018 on two cameras covering a pedestrian street running through the centre of the development.[200] In October 2019, a report by the deputy London mayor Sophie Linden revealed that in a secret deal the Metropolitan Police had passed photos of seven people to Argent for use in their King's cross facial recognition system.[201]
Automated Facial Recognition was trialled by the South Wales Police on multiple occasions between 2017 and 2019. The use of the technology was challenged in court by a private individual, Edward Bridges, with support from the charity Liberty (case known as R (Bridges) v Chief Constable South Wales Police). The case was heard in the Court of Appeal and a judgement was given in August 2020.[202] The case argued that the use of Facial Recognition was a privacy violation on the basis that there was insufficient legal framework or proportionality in the use of Facial Recognition and that its use was in violation of the Data Protection Acts 1998 and 2018. The case was decided in favour of Bridges and did not award damages. The case was settled via a declaration of wrongdoing.[202] In response to the case, the British Government has repeatedly attempted to pass a Bill regulating the use of Facial Recognition in public spaces. The proposed Bills have attempted to appoint a Commissioner with the ability to regulate Facial Recognition use by Government Services in a similar manner to the Commissioner for CCTV. Such a Bill has yet to come into force [correct as of September 2021[update]].[126]
In January 2023, New York Attorney General Letitia James asked for more information on the use of facial recognition technology from Madison Square Garden Entertainment following reports that the firm used it to block lawyers involved in litigation against the company from entering Madison Square Garden. She noted such a move would could go against federal, state, and local human rights laws.[203]
Imperfect technology in law enforcement
[edit]As of 2018,[update] it is still contested as to whether or not facial recognition technology works less accurately on people of color.[204] One study by Joy Buolamwini (MIT Media Lab) and Timnit Gebru (Microsoft Research) found that the error rate for gender recognition for women of color within three commercial facial recognition systems ranged from 23.8% to 36%, whereas for lighter-skinned men it was between 0.0 and 1.6%. Overall accuracy rates for identifying men (91.9%) were higher than for women (79.4%), and none of the systems accommodated a non-binary understanding of gender.[205] It also showed that the datasets used to train commercial facial recognition models were unrepresentative of the broader population and skewed toward lighter-skinned males. However, another study showed that several commercial facial recognition software sold to law enforcement offices around the country had a lower false non-match rate for black people than for white people.[206]
Experts fear that face recognition systems may actually be hurting citizens the police claims they are trying to protect.[207] It is considered an imperfect biometric, and in a study conducted by Georgetown University researcher Clare Garvie, she concluded that "there's no consensus in the scientific community that it provides a positive identification of somebody."[208] It is believed that with such large margins of error in this technology, both legal advocates and facial recognition software companies say that the technology should only supply a portion of the case – no evidence that can lead to an arrest of an individual.[208] The lack of regulations holding facial recognition technology companies to requirements of racially biased testing can be a significant flaw in the adoption of use in law enforcement. CyberExtruder, a company that markets itself to law enforcement said that they had not performed testing or research on bias in their software. CyberExtruder did note that some skin colors are more difficult for the software to recognize with current limitations of the technology. "Just as individuals with very dark skin are hard to identify with high significance via facial recognition, individuals with very pale skin are the same," said Blake Senftner, a senior software engineer at CyberExtruder.[208]
The United States' National Institute of Standards and Technology (NIST) carried out extensive testing of FRT system 1:1 verification[209] and 1:many identification.[209] It also tested for the differing accuracy of FRT across different demographic groups. The independent study concluded at present, no FRT system has 100% accuracy.[210]
Data protection
[edit]In 2010, Peru passed the Law for Personal Data Protection, which defines biometric information that can be used to identify an individual as sensitive data. In 2012, Colombia passed a comprehensive Data Protection Law which defines biometric data as senstivite information.[141] According to Article 9(1) of the EU's 2016 General Data Protection Regulation (GDPR) the processing of biometric data for the purpose of "uniquely identifying a natural person" is sensitive and the facial recognition data processed in this way becomes sensitive personal data. In response to the GDPR passing into the law of EU member states, EU based researchers voiced concern that if they were required under the GDPR to obtain individual's consent for the processing of their facial recognition data, a face database on the scale of MegaFace could never be established again.[211] In September 2019 the Swedish Data Protection Authority (DPA) issued its first ever financial penalty for a violation of the EU's General Data Protection Regulation (GDPR) against a school that was using the technology to replace time-consuming roll calls during class. The DPA found that the school illegally obtained the biometric data of its students without completing an impact assessment. In addition the school did not make the DPA aware of the pilot scheme. A 200,000 SEK fine (€19,000/$21,000) was issued.[citation needed]
In the United States of America several U.S. states have passed laws to protect the privacy of biometric data. Examples include the Illinois Biometric Information Privacy Act (BIPA) and the California Consumer Privacy Act (CCPA).[212] In March 2020 California residents filed a class action against Clearview AI, alleging that the company had illegally collected biometric data online and with the help of face recognition technology built up a database of biometric data which was sold to companies and police forces. At the time Clearview AI already faced two lawsuits under BIPA[213] and an investigation by the Privacy Commissioner of Canada for compliance with the Personal Information Protection and Electronic Documents Act (PIPEDA).[214]
Bans on the use of facial recognition technology
[edit]United States of America
[edit]In May 2019, San Francisco, California became the first major United States city to ban the use of facial recognition software for police and other local government agencies' usage.[215] San Francisco Supervisor, Aaron Peskin, introduced regulations that will require agencies to gain approval from the San Francisco Board of Supervisors to purchase surveillance technology.[216] The regulations also require that agencies publicly disclose the intended use for new surveillance technology.[216] In June 2019, Somerville, Massachusetts became the first city on the East Coast to ban face surveillance software for government use,[217] specifically in police investigations and municipal surveillance.[218] In July 2019, Oakland, California banned the usage of facial recognition technology by city departments.[219]
The American Civil Liberties Union ("ACLU") has campaigned across the United States for transparency in surveillance technology[218] and has supported both San Francisco and Somerville's ban on facial recognition software. The ACLU works to challenge the secrecy and surveillance with this technology.[citation needed][220]
During the George Floyd protests, use of facial recognition by city government was banned in Boston, Massachusetts.[221] As of June 10, 2020,[update] municipal use has been banned in:[11]
- Berkeley, California
- Oakland, California
- Boston, Massachusetts – June 30, 2020[222]
- Brookline, Massachusetts
- Cambridge, Massachusetts
- Northampton, Massachusetts
- Springfield, Massachusetts
- Somerville, Massachusetts
- Portland, Oregon – September 2020[223]
The West Lafayette, Indiana City Council passed an ordinance banning facial recognition surveillance technology.[224]
On October 27, 2020, 22 human rights groups called upon the University of Miami to ban facial recognition technology. This came after the students accused the school of using the software to identify student protesters. The allegations were, however, denied by the university.[225]
A state police reform law in Massachusetts will take effect in July 2021; a ban passed by the legislature was rejected by governor Charlie Baker.[226] Instead, the law requires a judicial warrant, limit the personnel who can perform the search, record data about how the technology is used, and create a commission to make recommendations about future regulations.[227]
Reports in 2024 revealed that some police departments, including San Francisco Police Department, had skirted bans on facial recognition technology that had been enacted in their respective cities.[228]
European Union
[edit]In January 2020, the European Union suggested, but then quickly scrapped, a proposed moratorium on facial recognition in public spaces.[229][230]
The European "Reclaim Your Face" coalition launched in October 2020. The coalition calls for a ban on facial recognition and launched a European Citizens' Initiative in February 2021. More than 60 organizations call on the European Commission to strictly regulate the use of biometric surveillance technologies.[231]
Emotion recognition
[edit]In the 18th and 19th century, the belief that facial expressions revealed the moral worth or true inner state of a human was widespread and physiognomy was a respected science in the Western world. From the early 19th century onwards photography was used in the physiognomic analysis of facial features and facial expression to detect insanity and dementia.[232] In the 1960s and 1970s the study of human emotions and its expressions was reinvented by psychologists, who tried to define a normal range of emotional responses to events.[233] The research on automated emotion recognition has since the 1970s focused on facial expressions and speech, which are regarded as the two most important ways in which humans communicate emotions to other humans. In the 1970s the Facial Action Coding System (FACS) categorization for the physical expression of emotions was established.[234] Its developer Paul Ekman maintains that there are six emotions that are universal to all human beings and that these can be coded in facial expressions.[235] Research into automatic emotion specific expression recognition has in the past decades focused on frontal view images of human faces.[236] Facial thermography can be considered as a promising tool of emotion recognition.[237][238]
In 2016, facial feature emotion recognition algorithms were among the new technologies, alongside high-definition CCTV, high resolution 3D face recognition and iris recognition, that found their way out of university research labs.[citation needed] In 2016, Facebook acquired FacioMetrics, a facial feature emotion recognition corporate spin-off by Carnegie Mellon University. In the same year Apple Inc. acquired the facial feature emotion recognition start-up Emotient.[239] By the end of 2016, commercial vendors of facial recognition systems offered to integrate and deploy emotion recognition algorithms for facial features.[citation needed] The MIT's Media Lab spin-off Affectiva[240] by late 2019 offered a facial expression emotion detection product that can recognize emotions in humans while driving.[239]
Anti-facial recognition systems
[edit]The development of anti-facial recognition technology is effectively an arms race between privacy researchers and big data companies. Big data companies increasingly use convolutional AI technology to create ever more advanced facial recognition models. Solutions to block facial recognition may not work on newer software, or on different types of facial recognition models. One popular cited example of facial-recognition blocking is the CVDazzle makeup and haircut system, but the creators note on their website that it has been outdated for quite some time as it was designed to combat a particular facial recognition algorithm and may not work.[241] Another example is the emergence of facial recognition that can identify people wearing facemasks and sunglasses, especially after the COVID-19 pandemic.[242]
Given that big data companies have much more funding than privacy researchers, it is very difficult for anti-facial recognition systems to keep up. There is also no guarantee that obfuscation techniques that were used for images taken in the past and stored, such as masks or software obfuscation, would protect users from facial-recognition analysis of those images by future technology.[243]
In January 2013, Japanese researchers from the National Institute of Informatics created 'privacy visor' glasses that use nearly infrared light to make the face underneath it unrecognizable to face recognition software that use infrared.[244] The latest version uses a titanium frame, light-reflective material and a mask which uses angles and patterns to disrupt facial recognition technology through both absorbing and bouncing back light sources.[245][246][247][248] However, these methods are used to prevent infrared facial recognition and would not work on AI facial recognition of plain images. Some projects use adversarial machine learning to come up with new printed patterns that confuse existing face recognition software.[249]
One method that may work to protect from facial recognition systems are specific haircuts and make-up patterns that prevent the used algorithms to detect a face, known as computer vision dazzle.[241] Incidentally, the makeup styles popular with Juggalos may also protect against facial recognition.[250]
Facial masks that are worn to protect from contagious viruses can reduce the accuracy of facial recognition systems. A 2020 NIST study, tested popular one-to-one matching systems and found a failure rate between five and fifty percent on masked individuals. The Verge speculated that the accuracy rate of mass surveillance systems, which were not included in the study, would be even less accurate than the accuracy of one-to-one matching systems.[251] The facial recognition of Apple Pay can work through many barriers, including heavy makeup, thick beards and even sunglasses, but fails with masks.[252] However, facial recognition of masked faces is increasingly getting more reliable.
Another solution is the application of obfuscation to images that may fool facial recognition systems while still appearing normal to a human user. These could be used for when images are posted online or on social media. However, as it is hard to remove images once they are on the internet, the obfuscation on these images may be defeated and the face of the user identified by future advances in technology. Two examples of this technique, developed in 2020, are the ANU's 'Camera Adversaria' camera app, and the University of Chicago's Fawkes image cloaking software algorithm which applies obfuscation to already taken photos.[243] However, by 2021 the Fawkes obfuscation algorithm had already been specifically targeted by Microsoft Azure which changed its algorithm to lower Fawkes' effectiveness.[253]
See also
[edit]- AI effect
- Amazon Rekognition
- Applications of artificial intelligence
- Artificial intelligence for video surveillance
- Automatic number plate recognition
- Biometric technology in access control
- Coke Zero Facial Profiler
- Computer processing of body language
- Computer vision
- DeepFace
- FaceNet
- Face perception
- Face Recognition Grand Challenge
- FindFace
- Glasgow Face Matching Test
- ISO/IEC 19794-5
- MALINTENT
- National biometric id card
- Multimedia information retrieval
- Multilinear subspace learning
- Pattern recognition, analogy and case-based reasoning
- Retinal scan
- SenseTime
- Super recognisers
- Template matching
- Three-dimensional face recognition
- Vein matching
- Gait analysis
- Fawkes (image cloaking software)
- Lists
References
[edit]- ^ "Face Recognition based Smart Attendance System Using IoT" (PDF). International Research Journal of Engineering and Technology. 9 (3): 5. March 2022.
- ^ Thorat, S. B.; Nayak, S. K.; Jyoti P Dandale (2010). "Facial Recognition Technology: An analysis with scope in India". arXiv:1005.4263 [cs.MA].
- ^ Chen, S.K; Chang, Y.H (2014). 2014 International Conference on Artificial Intelligence and Software Engineering (AISE2014). DEStech Publications, Inc. p. 21. ISBN 978-1-60595-150-8.
- ^ Bramer, Max (2006). Artificial Intelligence in Theory and Practice: IFIP 19th World Computer Congress, TC 12: IFIP AI 2006 Stream, August 21–24, 2006, Santiago, Chile. Berlin: Springer Science+Business Media. p. 395. ISBN 978-0-387-34654-0.
- ^ a b c d e SITNFlash (October 24, 2020). "Racial Discrimination in Face Recognition Technology". Science in the News. Retrieved July 1, 2023.
- ^ "Facial Recognition Technology: Federal Law Enforcement Agencies Should Have Better Awareness of Systems Used By Employees". www.gao.gov. Retrieved September 5, 2021.
- ^ a b Security, Help Net (August 27, 2020). "Facing gender bias in facial recognition technology". Help Net Security. Retrieved July 1, 2023.
- ^ Team, Lumen Database (May 5, 2021). "Sexism in Facial Recognition Technology". Berkman Klein Center Collection. Retrieved July 1, 2023.
- ^ Understanding bias in facial recognition technologies
- ^ Wiggers, Kyle (March 5, 2022). "Study warns deepfakes can fool facial recognition". VentureBeat. Retrieved June 4, 2022.
- ^ a b "IBM bows out of facial recognition market -". GCN. June 10, 2020. Archived from the original on November 30, 2021. Retrieved October 7, 2021.
- ^ Rachel Metz (November 2, 2021). "Facebook is shutting down its facial recognition software". CNN. Retrieved November 5, 2021.
- ^ Hill, Kashmir; Mac, Ryan (November 2, 2021). "Facebook, Citing Societal Concerns, Plans to Shut Down Facial Recognition System". The New York Times. ISSN 0362-4331. Retrieved November 5, 2021.
- ^ "IBM will no longer offer, develop, or research facial recognition technology". June 9, 2020.
- ^ a b Nilsson, Nils J. (October 30, 2009). The Quest for Artificial Intelligence. Cambridge University Press. ISBN 978-1-139-64282-8.
- ^ de Leeuw, Karl; Bergstra, Jan (2007). The History of Information Security: A Comprehensive Handbook. Elsevier. p. 266. ISBN 978-0-444-51608-4.
- ^ Gates, Kelly (2011). Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance. NYU Press. pp. 48–49. ISBN 978-0-8147-3209-0.
- ^ Gates, Kelly (2011). Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance. NYU Press. pp. 49–50. ISBN 978-0-8147-3209-0.
- ^ Gates, Kelly (2011). Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance. NYU Press. p. 52. ISBN 978-0-8147-3209-0.
- ^ Gates, Kelly (2011). Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance. NYU Press. p. 53. ISBN 978-0-8147-3209-0.
- ^ Gates, Kelly (2011). Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance. NYU Press. p. 54. ISBN 978-0-8147-3209-0.
- ^ a b Malay K. Kundu; Sushmita Mitra; Debasis Mazumdar; Sankar K. Pal, eds. (2012). Perception and Machine Intelligence: First Indo-Japan Conference, PerMIn 2012, Kolkata, India, January 12–13, 2011, Proceedings. Springer Science & Business Media. p. 29. ISBN 978-3-642-27386-5.
- ^ Wechsler, Harry (2009). Malay K. Kundu; Sushmita Mitra (eds.). Reliable Face Recognition Methods: System Design, Implementation and Evaluation. Springer Science & Business Media. pp. 11–12. ISBN 978-0-387-38464-1.
- ^ Jun Wang; Laiwan Chan; DeLiang Wang, eds. (2012). Neural Information Processing: 13th International Conference, ICONIP 2006, Hong Kong, China, October 3–6, 2006, Proceedings, Part II. Springer Science & Business Media. p. 198. ISBN 978-3-540-46482-2.
- ^ Wechsler, Harry (2009). Reliable Face Recognition Methods: System Design, Implementation and Evaluation. Springer Science & Business Media. p. 12. ISBN 978-0-387-38464-1.
- ^ Wechsler, Harry (2009). Malay K. Kundu; Sushmita Mitra (eds.). Reliable Face Recognition Methods: System Design, Implementation and Evaluation. Springer Science & Business Media. p. 12. ISBN 978-0-387-38464-1.
- ^ "Mugspot Can Find A Face In The Crowd – Face-Recognition Software Prepares To Go To Work In The Streets". ScienceDaily. November 12, 1997. Retrieved November 6, 2007.
- ^ Malay K. Kundu; Sushmita Mitra; Debasis Mazumdar; Sankar K. Pal, eds. (2012). Perception and Machine Intelligence: First Indo-Japan Conference, PerMIn 2012, Kolkata, India, January 12–13, 2011, Proceedings. Springer Science & Business Media. p. 29. ISBN 978-3-642-27386-5.
- ^ Li, Stan Z.; Jain, Anil K. (2005). Handbook of Face Recognition. Springer Science & Business Media. pp. 14–15. ISBN 978-0-387-40595-7.
- ^ Kumar Datta, Asit; Datta, Madhura; Kumar Banerjee, Pradipta (2015). Face Detection and Recognition: Theory and Practice. CRC. p. 123. ISBN 978-1-4822-2657-7.
- ^ Severi, Misty (April 15, 2022). "Ukraine uses facial recognition software to identify dead Russian soldiers".
- ^ "Facial recognition technology is a valuable tool". Los Angeles Daily News. May 15, 2022.
- ^ Italiano, Laura (April 15, 2022). "Ukraine is using facial recognition to ID dead Russian soldiers and send photos of corpses home to their moms: report". Business Insider.
- ^ Li, Stan Z.; Jain, Anil K. (2005). Handbook of Face Recognition. Springer Science & Business Media. p. 1. ISBN 978-0-387-40595-7.
- ^ Li, Stan Z.; Jain, Anil K. (2005). Handbook of Face Recognition. Springer Science & Business Media. p. 2. ISBN 978-0-387-40595-7.
- ^ "Airport Facial Recognition Passenger Flow Management". hrsid.com.
- ^ a b c Bonsor, K. (September 4, 2001). "How Facial Recognition Systems Work". Retrieved June 2, 2008.
- ^ Smith, Kelly. "Face Recognition" (PDF). Retrieved June 4, 2008.
- ^ R. Brunelli and T. Poggio, "Face Recognition: Features versus Templates", IEEE Trans. on PAMI, 1993, (15)10:1042–1052
- ^ R. Brunelli, Template Matching Techniques in Computer Vision: Theory and Practice, Wiley, ISBN 978-0-470-51706-2, 2009 ([1] TM book)
- ^ Zhang, David; Jain, Anil (2006). Advances in Biometrics: International Conference, ICB 2006, Hong Kong, China, January 5–7, 2006, Proceedings. Berlin: Springer Science & Business Media. p. 183. ISBN 978-3-540-31111-9.
- ^ "A Study on the Design and Implementation of Facial Recognition Application System". International Journal of Bio-Science and Bio-Technology.
- ^ H. Ugail, Deep face recognition using full and partial face images, Elesevier, ISBN 978-0-12-822109-9, 2022 ([2] Advanced Methods and Deep Learning in Computer Vision)
- ^ Harry Wechsler (2009). Reliable Face Recognition Methods: System Design, Implementation and Evaluation. Springer Science & Business Media. p. 196. ISBN 978-0-387-38464-1.
- ^ a b c d Williams, Mark. "Better Face-Recognition Software". Technology Review. Archived from the original on June 8, 2011. Retrieved June 2, 2008.
- ^ Crawford, Mark. "Facial recognition progress report". SPIE Newsroom. Retrieved October 6, 2011.
- ^ Kimmel, Ron. "Three-dimensional face recognition" (PDF). Retrieved January 1, 2005.
- ^ Duhn, S. von; Ko, M. J.; Yin, L.; Hung, T.; Wei, X. (September 1, 2007). "Three-View Surveillance Video Based Face Modeling for Recogniton [sic]". 2007 Biometrics Symposium. pp. 1–6. doi:10.1109/BCC.2007.4430529. ISBN 978-1-4244-1548-9. S2CID 25633949.
- ^ a b Socolinsky, Diego A.; Selinger, Andrea (January 1, 2004). "Thermal Face Recognition in an Operational Scenario". CVPR'04. IEEE Computer Society. pp. 1012–1019 – via ACM Digital Library.
- ^ a b "Army Builds Face Recognition Technology that Works in Low-Light Conditions". AZoRobotics. April 18, 2018. Retrieved August 17, 2018.
- ^ Thirimachos Bourlai (2016). Face Recognition Across the Imaging Spectrum. Springer. p. 142. ISBN 978-3-319-28501-6.
- ^ Thirimachos Bourlai (2016). Face Recognition Across the Imaging Spectrum. Springer. p. 140. ISBN 978-3-319-28501-6.
- ^ "Army develops face recognition technology that works in the dark". Army Research Laboratory. April 16, 2018. Archived from the original on April 22, 2018. Retrieved August 17, 2018.
- ^ a b Riggan, Benjamin; Short, Nathaniel; Hu, Shuowen (March 2018). Thermal to Visible Synthesis of Face Images Using Multiple Regions. 2018 IEEE Winter Conference on Applications of Computer Vision (WACV). pp. 30–38. arXiv:1803.07599. Bibcode:2018arXiv180307599R. doi:10.1109/WACV.2018.00010.
- ^ Cole, Sally (June 2018). "U.S. Army's AI facial recognition works in the dark". Military Embedded Systems. p. 8.
- ^ Shontell, Alyson (September 15, 2015). "Snapchat buys Looksery, a 2-year-old startup that lets you Photoshop your face while you video chat". Business Insider Singapore. Retrieved April 9, 2018.
- ^ Kumar Mandal, Jyotsna; Bhattacharya, Debika (2019). Emerging Technology in Modelling and Graphics: Proceedings of IEM Graph 2018. Springer. p. 672. ISBN 978-981-13-7403-6.
- ^ Bryson, Kevin (May 20, 2023). "Evaluating Anti-Facial Recognition Tools". physicalsciences.uchicago.edu. Retrieved January 27, 2024.
- ^ Simonite, Tom. "Facebook Creates Software That Matches Faces Almost as Well as You Do". MIT Technology Review. Retrieved April 9, 2018.
- ^ "Facebook's DeepFace shows serious facial recognition skills". Retrieved April 9, 2018.
- ^ "Why Facebook is beating the FBI at facial recognition". The Verge. Retrieved April 9, 2018.
- ^ "How TikTok's 'For You' Algorithm Actually Works". Wired. ISSN 1059-1028. Retrieved April 17, 2021.
- ^ "How TikTok recommends videos #ForYou". TikTok. June 18, 2020. Archived from the original on June 18, 2020. Retrieved April 22, 2021.
- ^ "TikTok agrees legal payout over facial recognition". BBC News. February 26, 2021. Archived from the original on February 26, 2021. Retrieved April 22, 2021.
- ^ "A glimpse at bank branches of the future: video walls, booth-sized locations and 24/7 access". USA Today. Retrieved August 13, 2018.
- ^ Heater, Brian. "Don't rely on Face Unlock to keep your phone secure". TechCrunch. Retrieved November 2, 2017.
- ^ "Galaxy S8 face recognition already defeated with a simple picture". Ars Technica. Retrieved November 2, 2017.
- ^ "How Facial Recognition Works in Xbox Kinect". Wired. Retrieved November 2, 2017.
- ^ "Windows 10 says "Hello" to logging in with your face and the end of passwords". Ars Technica. March 17, 2015. Retrieved March 17, 2015.
- ^ Kubota, Yoko (September 27, 2017). "Apple iPhone X Production Woe Sparked by Juliet and Her Romeo". The Wall Street Journal. Archived from the original on September 28, 2017. Retrieved September 27, 2017.
- ^ Kubota, Yoko (September 27, 2017). "Apple iPhone X Production Woe Sparked by Juliet and Her Romeo". The Wall Street Journal. ISSN 0099-9660. Retrieved April 10, 2018.
- ^ a b "The five biggest questions about Apple's new facial recognition system". The Verge. Retrieved April 10, 2018.
- ^ "Apple's Face ID Feature Works With Most Sunglasses, Can Be Quickly Disabled to Thwart Thieves". Retrieved April 10, 2018.
- ^ Heisler, Yoni (November 3, 2017). "Infrared video shows off the iPhone X's new Face ID feature in action". BGR. Retrieved April 10, 2018.
- ^ Okeke, Nnamdi (October 13, 2022). "Facial Recognition: How it works, Applications, Business ideas & More". TargetTrend. Retrieved October 21, 2022.
- ^ Libby, Christopher; Ehrenfeld, Jesse (2021). "Facial Recognition Technology in 2021: Masks, Bias, and the Future of Healthcare". Journal of Medical Systems. 45 (4): 39. doi:10.1007/s10916-021-01723-w. ISSN 0148-5598. PMC 7891114. PMID 33604732.
- ^ Kesari, Ganes. "How AI Is Using Facial Detection To Spot Rare Diseases In Children". Forbes. Retrieved October 21, 2022.
- ^ "Modi govt now plans a 'touchless' vaccination process, with Aadhaar-based facial recognition". ThePrint. April 6, 2021. Retrieved February 12, 2022.
- ^ "Despite Privacy Fears, Aadhaar-Linked Facial Recognition Used For Covid-19 Vaccines". Inc42. April 7, 2021. Retrieved February 12, 2022.
- ^ "Joint Statement: Say no to Aadhaar based Facial Recognition for Vaccination!". Internet Freedom Foundation. April 14, 2021. Retrieved February 12, 2022.
- ^ "Panoptic Tracker, Finance (Pension Cell) Department, Government of Meghalaya". Panoptic Project. Retrieved February 12, 2022.
- ^ "Meghalaya clarifies on controversial app: 'Facial Recognition Technology doesn't require any anchoring legislation'". Indian Express. November 18, 2021. Retrieved February 12, 2022.
- ^ "Smartgates Face editing for the mins of the can we have". Australian Border Force. Retrieved March 11, 2019.
- ^ "Our history". New Zealand Customs Service. Retrieved March 11, 2019.
- ^ "Facial recognition technology is coming to Canadian airports this spring". CBC News. Retrieved March 3, 2017.
- ^ a b c "Face Off: The lawless growth of facial recognition in UK policing" (PDF). Big Brother Watch.
- ^ Anthony, Sebastian (June 6, 2017). "UK police arrest man via automatic face-recognition tech". Ars Technica.
- ^ a b Rees, Jenny (September 4, 2019). "Police use of facial recognition ruled lawful". Retrieved November 8, 2019.
- ^ Burgess, Matt (January 24, 2020). "The Met Police will start using live facial recognition across London". Wired UK. ISSN 1357-0978. Retrieved January 24, 2020.
- ^ Danica Kirka (August 11, 2020). "UK court says face recognition violates human rights". TechPlore. Retrieved October 4, 2020.
- ^ Sylvester, Rachel (October 5, 2024). "'No human could do this': how facial recognition is transforming policing". The Times. Retrieved October 5, 2024.
- ^ "Here's How Many Adult Faces Are Scanned From Facial Recognition Databases". Fortune. October 18, 2016.
- ^ a b Kramer, Robin; Ritchie, Kay (December 14, 2016). "The trouble with facial recognition technology (in the real world)". phys.org.
- ^ "Real-Time Facial Recognition Is Available, But Will U.S. Police Buy It?". NPR.org. NPR. May 10, 2018.
- ^ "Police Facial Recognition Databases Log About Half Of Americans". NPR.org. NPR. October 23, 2016.
- ^ Rector, Kevin; Knezevich, Alison (October 17, 2016). "Maryland's use of facial recognition software questioned by researchers, civil liberties advocates". The Baltimore Sun.
- ^ "Next Generation Identification". FBI. Retrieved April 5, 2016.
- ^ a b "ICE Uses Facial Recognition To Sift State Driver's License Records, Researchers Say". NPR.org. July 8, 2019. Retrieved December 9, 2022.
- ^ "Facial recognition at airports: Everything you need to know". USA Today. August 16, 2019.
- ^ "TSA is adding face recognition at big airports. Here's how to opt out". Washington Post. December 2, 2022. ISSN 0190-8286. Retrieved December 9, 2022.
- ^ Belanger, Ashley (May 19, 2025). "New Orleans called out for sketchiest use of facial recognition yet in the US". Ars Technica. Retrieved May 29, 2025.
- ^ Shen, Xinmei (October 4, 2018). ""Skynet", China's massive video surveillance network". South China Morning Post. Retrieved December 13, 2020.
- ^ Chan, Tara Francis (March 27, 2018). "16 parts of China are now using Skynet". Business Insider. Retrieved December 13, 2020.
- ^ "From ale to jail: facial recognition catches criminals at China beer festival". The Guardian. September 1, 2017. Retrieved March 8, 2018.
- ^ "Police use facial recognition technology to detect wanted criminals during beer festival in Chinese city of Qingdao". opengovasia.com. OpenGovAsia. Archived from the original on November 16, 2017. Retrieved March 8, 2018.
- ^ "Chinese police are using smart glasses to identify potential suspects". TechCrunch. February 8, 2018. Retrieved December 3, 2020.
- ^ "Beijing police are using facial-recognition glasses to identify car passengers and number plates". Business Insider. March 12, 2018. Retrieved December 3, 2020.
- ^ "China's massive investment in artificial intelligence has an insidious downside". Science AAAS. February 7, 2018. Retrieved February 23, 2018.
- ^ "China bets on facial recognition in big drive for total surveillance". The Washington Post. 2018. Retrieved February 23, 2018.
- ^ Liao, Rita (May 8, 2019). "Alibaba-backed facial recognition startup Megvii raises $750 million". TechCrunch. Retrieved August 28, 2019.
- ^ Dai, Sarah (June 5, 2019). "AI unicorn Megvii not behind app used for surveillance in Xinjiang, says human rights group". South China Morning Post. Retrieved August 28, 2019.
- ^ Cheng Leng; Yingzhi Yang; Ryan Woo (February 20, 2020). "Exclusive: Hundreds of Chinese businesses seek billions in loans to contend with coronavirus". Reuters. Retrieved October 5, 2020.
- ^ "A lawsuit against face-scans in China could have big consequences". The Economist. November 9, 2019.
- ^ Xiaoshan, Huang; Wen, Cheng (August 14, 2020). "New evidence showing Tencent monitors overseas users". Radio Free Asia. Archived from the original on August 16, 2020. Retrieved August 15, 2020.
- ^ Zak Doffman (August 26, 2019). "Hong Kong Exposes Both Sides Of China's Relentless Facial Recognition Machine". Forbes. Retrieved December 3, 2020.
- ^ "Facial recognition forced on 800 million Chinese internet users". Radio France Internationale. October 15, 2019. Retrieved April 21, 2024.
- ^ Dustin, Tim (April 10, 2023). "AI aims to persecute Chinese Christians". Global Christian Relief. Retrieved February 28, 2024.
- ^ "Pandemic, Persecution and Pushback - Surveillance State". Falun Gong Report. Retrieved February 27, 2024.
- ^ "Country policy and information note: Falun Gong, China, November 2023 (accessible)". The United Kingdom Government. April 4, 2024. Retrieved April 21, 2024.
- ^ Lohr, Steve (February 9, 2018). "Facial Recognition Is Accurate, if You're a White Guy". The New York Times. ISSN 0362-4331. Retrieved February 14, 2022.
- ^ "NCRB's National Automated Facial Recognition System". panoptic.in. Retrieved February 14, 2022.
- ^ a b "Watch the Watchmen Series Part 4: The National Automated Facial Recognition System". Internet Freedom Foundation. October 7, 2020. Retrieved February 14, 2022.
- ^ "Justice K.S.Puttaswamy(Retd) vs Union Of India on 26 September, 2018". Indian Kanoon.
- ^ "We might be in the market for a new kind of face mask". Internet Freedom Foundation. July 18, 2019. Retrieved February 14, 2022.
- ^ Barik, Soumyarendra (October 22, 2019). "'Fingerprint is not a big issue': Hyderabad police on collecting biometrics of 'suspects'". MediaNama. Retrieved February 14, 2022.
- ^ a b "Automated Facial Recognition Technology (Moratorium and Review) Bill [HL] - Parliamentary Bills - UK Parliament". bills.parliament.uk. Retrieved September 10, 2021.
- ^ "UP Police launch 'Trinetra', its AI-powered face recognition app to catch criminals". The Financial Express. December 27, 2018. Retrieved February 14, 2022.
- ^ Das, Kalyan (August 27, 2018). "Uttarakhand Police acquire face recognition software to help nab criminals". Hindustan Times. Retrieved February 14, 2022.
- ^ "Crime and Criminal Tracking Network & Systems (CCTNS)". National Crime Records Bureau. Archived from the original on February 18, 2022. Retrieved February 18, 2022.
- ^ Chowdhury, Sagnik (November 20, 2015). "CCTNS Project to let police stations 'talk': where it stands, and how it can help fight crime". The Indian Express. Retrieved February 18, 2022.
- ^ Sudhi Ranjan Sen (July 8, 2019). "Home Ministry moves to get automated facial recognition system for police". Hindustan Times. Retrieved February 18, 2022.
- ^ Jain, Anushka (July 15, 2020). "IFF's Legal Notice to the NCRB on the Revised RFP for the National Automated Facial Recognition System". Internet Freedom Foundation. Retrieved February 18, 2022.
- ^ a b c d e f g Parliament of India. Rajya Sabha. Two Hundred Thirty Seventh Report on Police - Training, Modernisation and Reforms (PDF). Department-related Parliamentary Standing Committee on Home Affairs, India. 2022. p. 34. Archived from the original (PDF) on August 6, 2022.
- ^ U. Sudhakar Reddy (November 10, 2021). "8.3 lakh cameras in Telangana, Hyderabad turning into surveillance city: Amnesty". The Times of India. Retrieved February 18, 2022.
- ^ "Indian govt's approach to facial recognition is flawed & driven by faulty assumptions". ThePrint. November 27, 2019. Retrieved February 15, 2022.
- ^ "Right to Information Updates from Delhi Police, Kolkata Police and Telangana State Technology Services". panoptic.in. Retrieved February 15, 2022.
- ^ "Section 8(1)(d) in The Right To Information Act, 2005". Indian Kanoon.
- ^ "Project Panoptic: RTI Updates from Delhi Police, Kolkata Police and Telangana State Technology Services". Internet Freedom Foundation. December 1, 2020. Retrieved February 15, 2022.
- ^ Chunduru, Aditya (December 2, 2020). "RTI: Kolkata, Delhi police refuse to give information on facial recognition systems". MediaNama. Retrieved February 15, 2022.
- ^ "Mexican Government Adopts FaceIt Face Recognition Technology to Eliminate Duplicate Voter Registrations in Upcoming Presidential Election". Business Wire. May 11, 2000. Archived from the original on March 5, 2016. Retrieved June 2, 2008.
- ^ a b c Selinger, Evan; Polonetsky, Jules; Tene, Omer (2018). The Cambridge Handbook of Consumer Privacy. Cambridge University Press. p. 112. ISBN 978-1-316-85927-8.
- ^ Vogel, Ben. "Panama puts names to more faces". IHS Jane's Airport Review. Archived from the original on October 12, 2014. Retrieved October 7, 2014.
- ^ "'Made-in-China' products shine at Rio Olympics". The State Council, The people's Republic of China. August 15, 2016. Retrieved November 14, 2020.
- ^ Kayser-Bril, Nicolas (December 11, 2019). "At least 11 police forces use face recognition in the EU, AlgorithmWatch reveals". AlgorithmWatch.
- ^ Pedriti, Corina (January 28, 2021). "Flush with EU funds, Greek police to introduce live face recognition before the summer". AlgorithmWatch.
- ^ Coluccini, Riccardo (January 13, 2021). "Lo scontro Viminale-Garante della privacy sul riconoscimento facciale in tempo reale". IrpiMedia.
- ^ Techredacteur, Joost Schellevis (December 16, 2016). "Politie gaat verdachten opsporen met gezichtsherkenning". nos.nl (in Dutch). Retrieved September 22, 2019.
- ^ Boon, Lex (August 25, 2018). "Meekijken met de 226 gemeentecamera's". Het Parool (in Dutch). Retrieved September 22, 2019.
- ^ Duncan, Jane (June 4, 2018). "How CCTV surveillance poses a threat to privacy in South Africa". The Conversation.
- ^ Ross, Tim (2007). "3VR Featured on Fox Business News". Money for Breakfast (Interview). Fox Business.
Interviewer: Now, can I buy something like this? Is this... do you really restrict the customers for this? Tim Ross: It's primarily being purchased by banks, retailers, and the government today and is sold through a variety of security channels.
- ^ "Improve Customer Service". 3VR. Archived from the original on August 14, 2012.
3VR's Video Intelligence Platform (VIP) transforms customer service by allowing businesses to: • Optimize staffing decisions, increase sales conversion rates and decrease customer wait times by bringing extraordinary clarity to the analysis of traffic patterns • Align staffing decisions with actual customer activity, using dwell and queue line analytics to decrease customer wait times • Increase competitiveness by using 3VR's facial surveillance analytic to facilitate personalized customer greetings by employees • Create loyalty programs by combining point of sale (POS) data with facial recognition
- ^ a b c d e f g h Dastin, Jeffrey L. (July 28, 2020). "Special Report: Rite Aid deployed facial recognition systems in hundreds of U.S. stores". U.S. Legal News. Reuters. Archived from the original on December 19, 2020.
- ^ a b Mayhew, Stephen (March 17, 2019). "Casinos down under deploy facial recognition tech to spot offenders, problem gamblers | Biometric Update". www.biometricupdate.com. Retrieved June 30, 2022.
- ^ Scanlan, Rebekah (June 29, 2022). "The Good Guys scrap 'creepy' camera feature after backlash". News.com.au.
- ^ Greene, Lisa (February 15, 2001). "Face scans match few suspects" (SHTML). St. Petersburg Times. Archived from the original on November 30, 2014. Retrieved June 30, 2011.
By using Viisage software, police matched 19 people's faces to photos of people arrested in the past for minor pickpocketing, fraud and other charges. They weren't charged with any game-day misdeeds. THIS IS A FARCE
- ^ a b c Krause, Mike (January 14, 2002). "Is face recognition just high-tech snake oil?". Enter Stage Right. ISSN 1488-1756. Archived from the original on January 24, 2002. Retrieved June 30, 2011.
- ^ "Windows 10's Photos app is getting smarter image search just like Google Photos". The Verge. Retrieved November 2, 2017.
- ^ Perez, Sarah. "Google Photos upgraded with new sharing features, photo books, and Google Lens". TechCrunch. Retrieved November 2, 2017.
- ^ "Face Recognition Applications". Animetrics. Archived from the original on July 13, 2008. Retrieved June 4, 2008.
- ^ Giaritelli, Anna (December 13, 2018). "Taylor Swift used airport-style facial recognition on concertgoers". washingtonexaminer.com. Retrieved December 13, 2018.
- ^ "Manchester City tries facial recognition to beat football queues". The Times. Retrieved August 18, 2019.
- ^ "Manchester City warned against using facial recognition on fans". The Guardian. Retrieved August 18, 2019.
- ^ Olson, Parmy (August 1, 2020). "Facial Recognition's Next Big Play: the Sports Stadium". The Wall Street Journal. ISSN 0099-9660. Retrieved August 3, 2020.
- ^ "Facial Recognition Technology Test". Walt Disney World Park Entry Technology Test. Disney. Archived from the original on April 22, 2021. Retrieved April 22, 2021.
- ^ "Face recognition - Everything you need to know | Vidispine". www.vidispine.com.
- ^ R. Kimmel and G. Sapiro (April 30, 2003). "The Mathematics of Face Recognition". SIAM News. Archived from the original on July 15, 2007. Retrieved April 30, 2003.
- ^ a b c "Top Five Biometrics: Face, Fingerprint, Iris, Palm and Voice". Bayometric. January 23, 2017. Retrieved April 10, 2018.
- ^ "Privacy Principles for Facial Recognition Technology in Commercial Applications" (PDF). fpf.org.
- ^ Haghighat, Mohammad; Abdel-Mottaleb, Mohamed (2017). "Low Resolution Face Recognition in Surveillance Systems Using Discriminant Correlation Analysis". 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017). pp. 912–917. doi:10.1109/FG.2017.130. ISBN 978-1-5090-4023-0. S2CID 36639614.
- ^ "Passport Canada – Photos". passportcanada.gc.ca. Archived from the original on March 1, 2009.
- ^ Albiol, A., Albiol, A., Oliver, J., Mossi, J.M.(2012). Who is who at different cameras: people re-identification using depth cameras. Computer Vision, IET. Vol 6(5), 378–387.
- ^ Buolamwini, Joy; Gebru, Timnit (2018). "Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification". Proceedings of Machine Learning Research. 81: 1–15.
- ^ Mallick, Sumit (2022). "The Influence of the Other-Race Effect on Susceptibility to Face Morphing Attacks". arXiv:2204.12591 [cs.CV].
- ^ Jeckeln, Gabriel (2023). "Human-Machine Comparison for Cross-Race Face Verification: Race Bias at the Upper Limits of Performance". arXiv:2305.16443 [cs.CV].
- ^ Phillips, P. Jonathon; Jiang, Fang; Narvekar, Abhijit; Ayyad, Julianne; O'Toole, Alice J. (February 2, 2011). "An other-race effect for face recognition algorithms". ACM Trans. Appl. Percept. 8 (2): 14:1–14:11. doi:10.1145/1870076.1870082. ISSN 1544-3558.
- ^ Sangrigoli, Sophie; de Schonen, Scania (2004). "Recognition of Own-Race and Other-Race Faces by Three-Month-Old Infants". Journal of Child Psychology and Psychiatry. 45 (7): 1219–1227. doi:10.1111/j.1469-7610.2004.00319.x. PMID 15335342.
- ^ a b Kohno, Yoshi (2023). Evaluation of Targeted Dataset Collection on Racial Equity in Face Recognition (PDF). Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society. pp. 1–11.
- ^ Kolla, Manasa (2023). "The Impact of Racial Distribution in Training Data on Face Recognition Bias: A Closer Look". 2023 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW). pp. 313–322. arXiv:2211.14498. doi:10.1109/WACVW58289.2023.00035. ISBN 979-8-3503-2056-5.
- ^ Ramis, Silvia (2024). "Explainable Facial Expression Recognition for People with Intellectual Disabilities". XXIII International Conference on Human Computer Interaction. pp. 1–7. arXiv:2405.11482. doi:10.1145/3612783.3612789. ISBN 979-8-4007-0790-2.
- ^ Mankoff, Jennifer (2022). "Areas of Strategic Visibility: Disability Bias in Biometrics". arXiv:2208.04712 [cs.CY].
- ^ Li, Yingjie (2021). "Learning Fair Face Representation With Progressive Cross Transformer". arXiv:2108.04983 [cs.CV].
- ^ Meek, James (June 13, 2002). "Robo cop". London: UK Guardian newspaper.
- ^ "Birmingham City Centre CCTV Installs Visionics' FaceIt". Business Wire. June 2, 2008.
- ^ Willing, Richard (September 2, 2003). "Airport anti-terror systems flub tests; Face-recognition technology fails to flag 'suspects'". USA Today. Archived from the original (Abstract) on October 1, 2007. Retrieved September 17, 2007.
- ^ Meyer, Robinson (2015). "How Worried Should We Be About Facial Recognition?". The Atlantic. Retrieved March 2, 2018.
- ^ White, David; Dunn, James D.; Schmid, Alexandra C.; Kemp, Richard I. (October 14, 2015). "Error Rates in Users of Automatic Face Recognition Software". PLOS ONE. 10 (10) e0139827. Bibcode:2015PLoSO..1039827W. doi:10.1371/journal.pone.0139827. PMC 4605725. PMID 26465631.
- ^ "EFF Sues FBI For Access to Facial-Recognition Records". Electronic Frontier Foundation. June 26, 2013.
- ^ "Q&A On Face-Recognition". American Civil Liberties Union. Archived from the original on March 24, 2015. Retrieved July 23, 2014.
- ^ a b Harley Geiger (December 6, 2011). "Facial Recognition and Privacy". Center for Democracy & Technology. Retrieved January 10, 2012.
- ^ a b c Cackley, Alicia Puente (July 2015). "FACIAL RECOGNITION TECHNOLOGY Commercial Uses, Privacy Issues, and Applicable Federal Law" (PDF).
- ^ Thomas Brewster (September 22, 2020). "This Russian Facial Recognition Startup Plans To Take Its 'Aggression Detection' Tech Global With $15 Million Backing From Sovereign Wealth Funds". Forbes. Retrieved October 4, 2020.
- ^ "Singel-Minded: Anatomy of a Backlash, or How Facebook Got an 'F' for Facial Recognition". WIRED. Retrieved April 10, 2018.
- ^ "Facebook Can Now Find Your Face, Even When It's Not Tagged". WIRED. Retrieved April 10, 2018.
- ^ "Facebook Keeps Getting Sued Over Face-Recognition Software, And Privacy Groups Say We Should Be Paying More Attention". International Business Times. September 3, 2015. Retrieved April 5, 2016.
- ^ Herra, Dana. "Judge tosses Illinois privacy law class action vs Facebook over photo tagging; California cases still pending". cookcountyrecord.com. Retrieved April 5, 2016.
- ^ Skinner-Thompson, Scott (2020). Privacy at the Margins. Cambridge University Press. p. 110. ISBN 978-1-107-18137-3.
- ^ Murgia, Madhumita (August 12, 2019). "London's King's Cross uses facial recognition in security cameras". Financial Times (subscription site). Archived from the original on December 10, 2022. Retrieved August 17, 2019.
- ^ "King's Cross facial recognition investigated". BBC News. August 15, 2019. Retrieved August 17, 2019.
- ^ Cellan-Jones, Rory (August 16, 2019). "Tech Tent: Is your face on a watch list?". BBC News. Retrieved August 17, 2019.
- ^ Sabbagh, Dan (September 2, 2019). "Facial recognition technology scrapped at King's Cross site". The Guardian. ISSN 0261-3077. Retrieved September 2, 2019.
- ^ Sabbagh, Dan (October 4, 2019). "Facial recognition row: police gave King's Cross owner images of seven people". The Guardian. Retrieved October 4, 2020.
- ^ a b Judiciary UK (August 11, 2020). "Judgement: Bridges v South Wales Police - Courts and Tribunals Judiciary" (PDF). Judiciary UK. Retrieved September 10, 2021.
- ^ Sheidlower, Noah (January 25, 2023). "NY AG Letitia James presses MSG over use of facial recognition technology". CNBC. Retrieved January 25, 2023.
- ^ "Photo Algorithms ID White Men Fine—Black Women, Not So Much". WIRED. Retrieved April 10, 2018.
- ^ Joy Buolamwini; Timnit Gebru (2018). "Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification". Proceedings of Machine Learning Research. Vol. 81. pp. 77–91. ISSN 1533-7928. Retrieved March 8, 2018.
- ^ Grother, Patrick; Quinn, George; Phillips, P. Jonathon (August 24, 2011). "Report on the Evaluation of 2D Still-Image Face Recognition Algorithms" (PDF). National Institute of Standards and Technology.
- ^ Buranyi, Stephen (August 8, 2017). "Rise of the racist robots – how AI is learning all our worst impulses". The Guardian. Retrieved April 10, 2018.
- ^ a b c Brel, Ali (December 4, 2017). "How white engineers built racist code – and why it's dangerous for black people". The Guardian. Retrieved April 10, 2018.
- ^ a b "Face Recognition Vendor Test (FRVT) Ongoing". NIST. December 14, 2016. Retrieved February 15, 2022.
- ^ Grother, Patrick J.; Ngan, Mei L.; Hanaoka, Kayee K. (December 19, 2019). "Face Recognition Vendor Test Part 3: Demographic Effects". nist.gov.
- ^ Ronald Leenes; Rosamunde van Brakel; Serge Gutwirth; Paul de Hert, eds. (2018). Data Protection and Privacy: The Internet of Bodies. Bloomsbury Publishing. p. 176. ISBN 978-1-5099-2621-3.
- ^ Bock, Lisa (2020). Identity Management with Biometrics: Explore the latest innovative solutions to provide secure identification and authentication. Packt Publishing. p. 320. ISBN 978-1-83921-321-2.
- ^ Pascu, Luana (March 16, 2020). "California residents file class action against Clearview AI biometric data collection citing CCPA". BiometricUpdate.com. Retrieved October 25, 2020.
- ^ Burt, Chris (February 24, 2020). "Canadian Privacy Commissioners investigate Clearview AI, develop guidance for police use of biometrics". BiometricUpdate.com. Retrieved October 25, 2020.
- ^ Conger, Kate; Fausset, Richard; Kovaleski, Serge F. (May 14, 2019). "San Francisco Bans Facial Recognition Technology". The New York Times. ISSN 0362-4331. Retrieved March 26, 2020.
- ^ a b "San Francisco Bans Agency Use of Facial Recognition Tech". Wired. ISSN 1059-1028. Retrieved March 26, 2020.
- ^ "Somerville Bans Government Use Of Facial Recognition Tech". wbur.org. June 28, 2019. Retrieved March 26, 2020.
- ^ a b "Somerville City Council passes facial recognition ban – The Boston Globe". The Boston Globe. Retrieved March 26, 2020.
- ^ Haskins, Caroline (July 17, 2019). "Oakland Becomes Third U.S. City to Ban Facial Recognition". Vice. Retrieved April 11, 2020.
- ^ Nkonde, Mutale (2019). "Automated Anti-Blackness: Facial Recognition in Brooklyn, New York". Kennedy School Review. 20: 30–26. ProQuest 2404400349.
- ^ Boston mayor OKs ban on facial recognition tech
- ^ Boston mayor OKs ban on facial recognition tech
- ^ Rachel Metz (September 10, 2020). "Portland passes broadest facial recognition ban in the US". CNN. Retrieved September 13, 2020.
- ^ "West Lafayette City Council approves ban on facial recognition technology" [3]
- ^ "Human Rights Groups Call On The University of Miami To Ban Facial Recognition". Forbes. Retrieved October 27, 2020.
- ^ "Governor signs police overhaul into law - The Boston Globe". BostonGlobe.com.
- ^ "Massachusetts is one of the first states to create rules around facial recognition in criminal investigations". The New York Times. March 1, 2021 – via NYTimes.com.
- ^ MacMillan, Douglas. "MSN". www.msn.com.
- ^ "EU drops idea of facial recognition ban in public areas: paper". Reuters. January 29, 2020. Retrieved April 12, 2020.
- ^ "Facial recognition: EU considers ban". BBC News. January 17, 2020. Retrieved April 12, 2020.
- ^ "Reclaim Your Face: Ban Biometric Mass Surveillance!". Reclaim Your Face. Retrieved June 12, 2021.
- ^ Gates, Kelly (2011). Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance. NYU Press. p. 156. ISBN 978-0-8147-3209-0.
- ^ Gates, Kelly (2011). Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance. NYU Press. p. 161. ISBN 978-0-8147-3209-0.
- ^ Konar, Amit; Chakraborty, Aruna (2015). Emotion Recognition: A Pattern Analysis Approach. John Wiley & Sons. p. 185. ISBN 978-1-118-13066-7.
- ^ Konar, Amit; Chakraborty, Aruna (2015). Emotion Recognition: A Pattern Analysis Approach. John Wiley & Sons. p. 186. ISBN 978-1-118-13066-7.
- ^ Konar, Amit; Chakraborty, Aruna (2015). Emotion Recognition: A Pattern Analysis Approach. John Wiley & Sons. p. 187. ISBN 978-1-118-13066-7.
- ^ Ioannou, Stephanos; Gallese, Vittorio; Merla, Arcangelo (2014). "Thermal infrared imaging in psychophysiology: Potentialities and limits". Psychophysiology. 51 (10): 951–963. doi:10.1111/psyp.12243. ISSN 1469-8986. PMC 4286005. PMID 24961292.
- ^ Kosonogov, Vladimir; Zorzi, Lucas De; Honoré, Jacques; Martínez-Velázquez, Eduardo S.; Nandrino, Jean-Louis; Martinez-Selva, José M.; Sequeira, Henrique (September 18, 2017). "Facial thermal variations: A new marker of emotional arousal". PLOS ONE. 12 (9) e0183592. Bibcode:2017PLoSO..1283592K. doi:10.1371/journal.pone.0183592. ISSN 1932-6203. PMC 5603162. PMID 28922392.
- ^ a b Fowler, Gary (October 14, 2019). "How Emotional AI Is Creating Personalized Customer Experiences And Making A Social Impact". Frobes. Retrieved October 17, 2020.
- ^ "Eureka Park Returns" (Press release). National Science Foundation. January 7, 2013. Retrieved February 3, 2013.
- ^ a b Harvey, Adam. "CV Dazzle: Camouflage from Face Detection". cvdazzle.com. Retrieved September 15, 2017.
- ^ Heilweil, Rebecca (July 28, 2020). "Masks can fool facial recognition systems, but the algorithms are learning fast". www.vox.com. Retrieved June 30, 2022.
- ^ a b Marks, Paul (2020). "Blocking Facial Recognition". cacm.acm.org. Retrieved June 30, 2022.
- ^ "These Goofy-Looking Glasses Could Make You Invisible to Facial Recognition Technology". Slate. January 18, 2013. Retrieved January 22, 2013.
- ^ Hongo, Jun. "Eyeglasses with Face Un-Recognition Function to Debut in Japan". The Wall Street Journal. Retrieved February 9, 2017.
- ^ Osborne, Charlie. "Privacy visor which blocks facial recognition software set for public release". ZDNet. Retrieved February 9, 2017.
- ^ Stone, Maddie (August 8, 2015). "These Glasses Block Facial Recognition Technology". Gizmodo. Retrieved February 9, 2017.
- ^ "How Japan's Privacy Visor fools face-recognition cameras". PC World. Retrieved February 9, 2017.
- ^ Cox, Kate (April 10, 2020). "Some shirts hide you from cameras—but will anyone wear them?". Ars Technica. Retrieved April 12, 2020.
- ^ Schreiber, Hope (July 2, 2018). "Worried about facial recognition technology? Juggalo makeup prevents involuntary surveillance". Retrieved July 18, 2019.
- ^ Vincent, James (July 28, 2020). "Face masks are breaking facial recognition algorithms, says new government study". The Verge. Retrieved August 27, 2020.
- ^ Hern, Alex (August 21, 2020). "Face masks give facial recognition software an identity crisis". The Guardian. ISSN 0261-3077. Retrieved August 24, 2020.
- ^ "Fawkes AI - University of Chicago". Retrieved June 30, 2022.
Further reading
[edit]- Farokhi, Sajad; Shamsuddin, Siti Mariyam; Flusser, Jan; Sheikh, U.U; Khansari, Mohammad; Jafari-Khouzani, Kourosh (2014). "Near infrared face recognition by combining Zernike moments and undecimated discrete wavelet transform". Digital Signal Processing. 31 (1): 13–27. Bibcode:2014DSP....31...13F. doi:10.1016/j.dsp.2014.04.008.
- "The Face Detection Algorithm Set to Revolutionize Image Search" (Feb. 2015), MIT Technology Review
- Garvie, Clare; Bedoya, Alvaro; Frankle, Jonathan (October 18, 2016). Perpetual Line Up: Unregulated Police Face Recognition in America. Center on Privacy & Technology at Georgetown Law. Retrieved October 22, 2016.
- "Facial Recognition Software 'Sounds Like Science Fiction,' but May Affect Half of Americans". As It Happens. Canadian Broadcasting Corporation. October 20, 2016. Retrieved October 22, 2016. Interview with Alvaro Bedoya, executive director of the Center on Privacy & Technology at Georgetown Law and co-author of Perpetual Line Up: Unregulated Police Face Recognition in America.
- Press, Eyal, "In Front of Their Faces: Does facial-recognition technology lead police to ignore contradictory evidence?", The New Yorker, 20 November 2023, pp. 20–26.
External links
[edit]
Media related to Facial recognition system at Wikimedia Commons- A Photometric Stereo Approach to Face Recognition (master's thesis). The University of the West of England, Bristol.
Facial recognition system
View on GrokipediaHistory
Early Concepts and Pioneering Research (1960s-1990s)
In the mid-1960s, mathematician Woodrow Bledsoe pioneered semi-automated facial recognition through a "man-machine" system, where human operators used a graphical tablet to manually mark landmarks such as the eyes, nose tip, mouth, and chin on photographic images, enabling a computer to compute Euclidean distances and angles between these points for matching against stored templates.[12] This approach, tested on small datasets of up to 256 faces, achieved high accuracy in controlled comparisons but relied heavily on human input for feature localization, limiting scalability.[13] Bledsoe's work, partially funded by the CIA for counterintelligence applications, demonstrated the feasibility of quantitative facial measurement despite the era's computational constraints.[14] By 1969, researchers Arthur Goldstein, Leon Harmon, and Alan Lesk advanced feature-based methods by developing a system for numerical coding of facial attributes, including 21 quantitative measures like forehead width, eyebrow thickness, and nose shape, combined with qualitative descriptors such as hair color and lip fullness.[15] These efforts emphasized geometric ratios and holistic profiles but still required manual preprocessing, reflecting the period's emphasis on hybrid human-computer processes over full automation.[12] In 1973, Takeo Kanade introduced the first fully automated computer program for human face recognition, using edge detection and curvature analysis on photographs to extract feature points like the eyes, nostrils, mouth contours, and chin outline without manual intervention.[16] Kanade's system processed low-resolution images (approximately 100x100 pixels) and achieved recognition rates of around 70% on limited test sets of frontal faces under uniform lighting, highlighting challenges with pose variations and occlusions that persisted into later decades.[12] Throughout the 1970s and 1980s, subsequent research refined these geometric feature extraction techniques amid slow progress, constrained by processing power that could handle only dozens of comparisons per minute and error rates exceeding 20% in uncontrolled conditions.[17] The late 1980s and early 1990s marked a shift toward appearance-based holistic methods, culminating in 1991 with Matthew Turk and Alex Pentland's eigenfaces approach, which applied principal component analysis (PCA) to a training set of centered, normalized grayscale face images to generate eigenfaces—orthogonal basis vectors capturing the principal axes of facial variance, such as overall lighting and expression differences.[18] This technique projected query images into the eigenspace, comparing coefficients against stored prototypes for identification, yielding recognition accuracies of 90-96% on datasets of 10-20 subjects under consistent conditions, though performance degraded with viewpoint changes or novel illuminations. Eigenfaces represented a computational efficiency gain, reducing dimensionality from thousands of pixels to tens of eigenvectors, and influenced subsequent subspace learning paradigms despite sensitivities to outliers and non-Gaussian data distributions.[19]Commercialization and Algorithmic Breakthroughs (2000s-2010s)
The 2000s marked the initial commercialization of facial recognition systems, driven by heightened security demands post-9/11. In January 2001, the technology was trialed at Super Bowl XXXV in Tampa, Florida, where surveillance cameras scanned approximately 100,000 attendees against a watchlist, though it yielded no arrests and faced privacy backlash.[20] Early deployments expanded to airports, casinos, and government facilities, with companies like NEC introducing NeoFace in 2002 as one of the first mass-market products capable of processing large-scale biometric data.[21] In 2002, the U.S. National Institute of Justice (NIJ) funded a pilot at Prince George's County Correctional Center for staff access control, demonstrating practical utility in controlled environments.[22] Algorithmic advances underpinned this shift, with the Viola–Jones object detection framework published in 2001 revolutionizing real-time face detection via Haar-like features, integral images for rapid computation, and AdaBoost for classifier training, enabling efficient processing at 15 frames per second on modest hardware.[23] Subsequent methods like Local Binary Patterns (LBP) in 2004 improved texture-based feature extraction for recognition under varying illumination.[24] The 2005 Face Recognition Grand Challenge (FRGC) by NIST spurred algorithm enhancements using over 50,000 images, while 3D recognition progressed, with NIJ-supported systems in 2006 achieving viability at distances of 3-9 meters.[22][24] Into the 2010s, deep learning catalyzed breakthroughs, with Facebook's DeepFace system in 2014 attaining 97.35% accuracy on the Labeled Faces in the Wild (LFW) benchmark using a 9-layer convolutional neural network (CNN) trained on millions of images, approaching human-level verification.[24] The FBI's Next Generation Identification (NGI) system, incrementally deployed from February 2011 with full operational capability by September 2014, integrated facial recognition into its biometric database, supporting searches against over 15 million mugshots by law enforcement.[25] NIST vendor tests reflected rapid gains, with top algorithms reducing false non-match rates to 0.2% by 2018 from 4% in 2014, driven by deep neural networks handling pose, lighting, and occlusion variations.[22] These developments enabled scalable commercial adoption in social media tagging, border control, and retail, though accuracy disparities persisted across demographics.[24]Integration with AI and Recent Deployments (2020s)
The integration of advanced artificial intelligence techniques, particularly deep learning models like convolutional neural networks and ArcFace embeddings, has dramatically enhanced facial recognition accuracy in the 2020s, enabling robust performance under varied conditions such as occlusions and low light.[26] NIST's ongoing Face Recognition Vendor Test (FRVT) evaluations demonstrate this progress, with top algorithms achieving false non-match rates under 0.1% on datasets exceeding 12 million images by 2022, reflecting iterative improvements from submissions starting in 2020.[27] These advancements stem from larger training datasets and architectural refinements, allowing systems to extract high-dimensional facial features more effectively than earlier methods.[28] The COVID-19 pandemic prompted specific AI adaptations for masked faces, with NIST testing revealing that 65 algorithms submitted by December 2020 reduced identification errors by incorporating partial feature analysis and synthetic data generation via GANs.[29] By 2025, vendors like NEC achieved the highest NIST FRVT rankings, with accuracy scores surpassing 99.8% in controlled benchmarks, underscoring the causal link between computational scaling and error rate reductions.[30] Such integrations have also extended to multimodal systems combining facial data with iris or gait analysis for higher reliability in real-world scenarios.[31] Deployments proliferated in the 2020s, fueled by these AI enhancements. In the United Kingdom, police live facial recognition operations escalated to 256 van-based uses in 2024, up from 63 in 2023, primarily for public safety events.[32] Commercial applications expanded similarly; Disney implemented facial recognition for hotel check-ins and park access starting in 2021, streamlining guest verification across its resorts.[20] The global market for facial recognition reached $6.94 billion in 2024, projected to hit $7.92 billion in 2025, driven by integrations in smartphones—expected to feature in 90% of devices—and border controls.[33] [34] In China, railway stations like Beijing West continued deploying AI-powered fare gates for real-time passenger authentication, processing millions daily with error rates below 1%.[35] Law enforcement agencies worldwide adopted these systems for suspect identification; Clearview AI's platform, leveraging vast web-scraped datasets, reported 99.85% accuracy on mugshot matching in NIST-evaluated subsets.[36] However, real-world deployments revealed gaps between lab benchmarks and operational performance, with NIST noting that uncontrolled variables like lighting can elevate false positives, necessitating on-site validation.[37] By mid-2025, over 76% of UK police live facial recognition trials since 2015 occurred in 2024 alone, indicating accelerated institutional adoption amid ongoing accuracy gains.[38]Technical Mechanisms
Face Detection and Preprocessing
Face detection constitutes the foundational stage in facial recognition systems, tasked with identifying and localizing human faces amid complex backgrounds in images or video streams. This process employs scanning mechanisms to propose candidate regions, followed by classification to confirm facial presence, enabling isolation of relevant areas for further processing. Early algorithms prioritized computational efficiency for real-time applications, while recent advancements integrate neural networks for superior accuracy across varied conditions.[39] The Viola-Jones algorithm, developed in 2001 by Paul Viola and Michael Jones, exemplifies a seminal cascade classifier approach using Haar-like features—simple rectangular patterns sensitive to edges and contrasts common in faces, such as the bridge of the nose. Integral images facilitate rapid feature computation at multiple scales, and AdaBoost selects an optimal subset of features for weak classifiers combined into a strong detector, achieving real-time performance on standard hardware through a multi-stage rejection cascade that discards non-facial regions early. This method excels in frontal, near-frontal views but struggles with occlusions, extreme poses, or low resolution.[40][39] Deep learning has supplanted traditional methods in contemporary systems, with architectures like MTCNN—a multi-task cascaded convolutional neural network introduced around 2016—employing three sequential networks: a proposal network for coarse detection, a refinement network for bounding box adjustment, and an output network for facial landmark prediction. MTCNN enhances detection precision and supports alignment by estimating five key points (eyes, nose, mouth corners), outperforming Viola-Jones on datasets with profile views or partial occlusions. Single-stage detectors, such as adaptations of SSD or RetinaFace, further optimize speed and accuracy by predicting faces and landmarks in one forward pass, leveraging large-scale training on annotated corpora like WIDER FACE.[41][42] Preprocessing follows detection to standardize face crops, mitigating extraneous variations that could degrade recognition accuracy. Geometric alignment warps the image to a canonical pose using detected landmarks, correcting for rotation, scale, and translation via affine transformations. Photometric adjustments address illumination disparities through techniques like histogram equalization, which redistributes intensity values for uniform contrast, or local normalization to handle shadows. Additional steps often include grayscale conversion to reduce dimensionality, Gaussian blurring for noise suppression, and cropping to a fixed size (e.g., 112x112 pixels) while discarding non-facial elements. These operations, validated in empirical studies, can boost downstream matching performance by 10-20% on benchmark datasets under uncontrolled conditions.[43][44][45]Feature Extraction and Representation
Feature extraction in facial recognition systems transforms detected face images into compact, discriminative representations by identifying key characteristics such as geometric landmarks, texture patterns, or statistical variances that distinguish individual identities.[46] These features are encoded into vectors or embeddings in a lower-dimensional space to facilitate efficient matching while minimizing irrelevant variations like lighting or pose.[47] Traditional approaches classify into holistic methods, which process the entire face globally, local methods focusing on specific regions, and hybrids combining both.[46] Holistic techniques, such as principal component analysis (PCA), represent faces as linear combinations of eigenfaces—eigenvectors derived from the covariance matrix of a training set of face images.[48] Introduced by Turk and Pentland in 1991, this method projects input faces onto the principal subspace to capture the directions of maximum variance, reducing dimensionality from thousands of pixels to tens of coefficients while retaining essential identity information.[49] Linear discriminant analysis (LDA) extends PCA by maximizing class separability, optimizing for between-class scatter relative to within-class variance in supervised settings.[50] Local feature-based extraction emphasizes robust descriptors from facial components, such as local binary patterns (LBP) for texture invariance or histograms of oriented gradients (HOG) for edge distributions.[51] These methods divide the face into patches, compute invariant features resistant to illumination changes, and aggregate them into histograms or bags-of-words representations.[51] Haar-like features, rectangular patterns measuring regional contrasts, were pivotal in early detection but also contribute to extraction by highlighting structural elements like the nose bridge.[52] Contemporary systems predominantly employ deep convolutional neural networks (CNNs) for end-to-end feature learning, where convolutional layers hierarchically extract low-level edges progressing to high-level semantic features like eye spacing or jawline contours.[53] Models like DeepID, trained on large datasets to predict identity-related attributes, yield embeddings in deep feature spaces that outperform handcrafted methods on benchmarks, achieving accuracies exceeding 99% on labeled faces in the wild under controlled conditions.[54] Representation in these paradigms involves fixed-length vectors from fully connected layers, often normalized for cosine similarity matching, enabling scalability to millions of identities.[55] Empirical evaluations indicate CNN-extracted features generalize better across poses and expressions compared to PCA, though they require substantial computational resources and data volumes.[56]Matching Algorithms and Decision Processes
Matching algorithms in facial recognition systems compare the feature representation of a detected and preprocessed probe face to templates in a gallery database, computing similarity scores to identify potential matches. These algorithms fall into categories such as holistic approaches, which treat the face as a unified pattern; feature-based methods, which analyze specific landmarks like distances between eyes or nose width; and modern deep learning embeddings, which project faces into compact vector spaces for distance-based comparison. Holistic methods, exemplified by Eigenfaces introduced in 1991, apply principal component analysis (PCA) to derive eigenfaces from training images, projecting probe and gallery faces onto this subspace and matching via coefficient similarity.[57] Feature-based algorithms extract geometric or texture descriptors from keypoints and align them for metric comparison, offering robustness to pose variations but sensitivity to occlusion.[58] Contemporary systems predominantly employ deep convolutional neural networks (CNNs) for matching, such as FaceNet developed by Google in 2015, which uses triplet loss to learn 128-dimensional embeddings where Euclidean distance or cosine similarity correlates with facial identity, enabling efficient 1:1 verification or 1:N search.[59] In 1:1 verification, the probe is compared to a single enrolled template; in 1:N identification, it searches large galleries, often ranking candidates by score. Similarity metrics include Euclidean distance for embedding spaces or specialized losses like ArcFace for angular margins to enhance discriminability.[26] Decision processes apply a threshold to the computed similarity or distance score: scores above the threshold (or below for distances) declare a match, while probabilistic models may output confidence levels for human review. Threshold selection trades off false acceptance rate (FAR), the proportion of impostor pairs incorrectly matched, against false rejection rate (FRR), the proportion of genuine pairs rejected, with the equal error rate (EER) marking their intersection.[60][61] Operational thresholds are application-specific; security contexts favor low FAR to minimize risks, potentially increasing FRR. The National Institute of Standards and Technology (NIST) evaluates via its Face Recognition Vendor Test (FRVT), measuring false non-match rate (FNMR, akin to FRR) at fixed low false match rates (FMR, akin to FAR) like 10^{-6}, where leading algorithms on high-quality datasets like visa photos achieve FNMR values under 0.1%, demonstrating high efficacy under controlled conditions but degradation in unconstrained scenarios.[62][63]Specialized Sensing Modalities
Specialized sensing modalities in facial recognition systems incorporate sensors beyond standard visible-spectrum cameras to improve performance under challenging conditions such as low illumination, varying lighting, or presentation attacks. These include infrared (IR) imaging, thermal imaging, and 3D depth sensing technologies like structured light and time-of-flight (ToF), which capture physiological or geometric features not discernible in 2D RGB images.[64][65] Infrared imaging, particularly near-IR and thermal IR, enables operation in darkness or poor visibility by detecting reflected or emitted radiation rather than relying on ambient light. Thermal IR sensors capture facial heat patterns influenced by underlying blood vessels and tissue, providing illumination-invariant representations that enhance recognition accuracy; for instance, they perform effectively without external illuminators in total darkness.[66][67] Near-IR, often paired with active illumination, supports liveness detection by revealing subsurface skin textures difficult to replicate in spoofs.[68] Commercial systems, such as HID's U.are.U, integrate multi-spectral RGB-IR with structured light for day/night sensing.[69] Three-dimensional sensing reconstructs facial geometry using depth information, mitigating vulnerabilities to 2D spoofs like photographs. Structured light projects known patterns onto the face, with cameras analyzing deformations to compute depth maps, offering high precision at short ranges suitable for access control.[70] Time-of-flight sensors, by contrast, measure the round-trip time of emitted light pulses to generate distance data, enabling real-time 3D mapping over greater distances but with potential sensitivity to ambient light interference.[71][72] Hybrid systems combine 3D depth sensors with multi-spectrum optical inputs for robust feature extraction.[64] Multispectral imaging fuses data from visible, near-IR, and other wavelengths to exploit unique spectral responses of human skin, reducing effects of illumination-induced color shifts and improving anti-spoofing via physiological signatures absent in synthetic materials.[73][68] These modalities often integrate in modern devices to achieve higher false acceptance rates below 0.1% in benchmarks, though computational demands increase with sensor fusion.[74]Applications
Law Enforcement and Public Safety
Facial recognition systems enable law enforcement agencies to match images from surveillance footage, body cameras, or witness photos against databases of known individuals, facilitating the identification of suspects in crimes such as theft, assault, and homicide. In the United States, the Federal Bureau of Investigation's Next Generation Identification (NGI) Interstate Photo System (IPS), operational since 2011, allows authorized users to conduct searches against a repository of over 12 million facial images, primarily mugshots, to generate investigative leads. Between fiscal year 2017 and April 2019, the FBI processed 152,565 facial recognition search requests from law enforcement partners, yielding thousands of potential matches annually that supported investigations.[75][76] Empirical analysis across 268 U.S. cities from 1997 to 2020 demonstrates that staggered adoption of facial recognition by police departments correlated with statistically significant reductions in violent crime rates, particularly homicides, without corresponding increases in overall arrest rates or racial disparities in arrests. Using generalized difference-in-differences regressions with multiway fixed effects, the study attributed these declines to faster and more certain identifications leading to apprehensions, which enhance deterrence effects. Cities adopting the technology earlier experienced larger homicide rate drops, suggesting causal efficacy in public safety outcomes through improved investigative efficiency rather than over-policing.[77] In the United Kingdom, the Metropolitan Police Service has deployed live facial recognition (LFR) technology since 2020, scanning crowds in real-time to match faces against watchlists of wanted persons, resulting in over 1,000 arrests by mid-2025, including 93 registered sex offenders. Of these, approximately 75% led to charges or court outcomes, demonstrating practical utility in apprehending high-harm offenders during public events and routine patrols. Similar deployments by other forces, such as South Wales Police, have identified suspects in burglary and violence cases, underscoring the technology's role in proactive policing to prevent escalation of threats to public safety.[78][79]Border Security and Immigration Control
Facial recognition systems are deployed at international borders and airports to verify traveler identities against passport biometrics, detect imposters, and screen against watchlists, thereby enhancing security while expediting processing. In controlled environments like e-gates and kiosks, these systems compare live facial scans to stored images in electronic passports or databases, achieving match rates exceeding 98% in many implementations due to standardized lighting, pose requirements, and cooperative subjects.[80][81] In the United States, U.S. Customs and Border Protection (CBP) has integrated facial recognition into its Traveler Verification Service, deployed at all 328 U.S. airports for arrivals by June 2022 and at 32 airports for departures as of July 2022. The system processes over 280,000 travelers daily, confirming identities in seconds with accuracy rates over 98%, surpassing manual inspections in efficiency and reducing fraud attempts by matching against derogatory galleries. Testing has shown identification rates above 90% for air exits, with ongoing evaluations addressing demographic variations in performance.[80][82][83] The European Union's Entry/Exit System (EES), operational since October 12, 2025, mandates facial images and fingerprints from non-EU nationals at Schengen external borders to track entries, exits, and overstays across 29 countries. This biometric database supports automated verification at e-gates, improving detection of irregular migration while minimizing manual checks. Similarly, Australia's SmartGate network, using facial recognition with ePassports, processes arrivals at major airports, with expansions including new kiosks at Sydney Airport in May 2025, enabling faster clearance for eligible travelers including U.S. Global Entry members.[84][85][86][87] These deployments demonstrate facial recognition's utility in scaling border operations amid rising travel volumes, with empirical data indicating reduced processing times—often under 10 seconds per traveler—and higher interception rates for identity fraud compared to traditional methods, though efficacy depends on database quality and algorithmic updates to mitigate environmental factors.[88][89]Commercial and Retail Operations
Facial recognition systems in commercial and retail operations primarily serve loss prevention by scanning customer faces against databases of known shoplifters, enabling real-time alerts to security personnel.[90][91] Retailers such as Kmart, Bunnings, and The Good Guys in Australia employ this technology to identify repeat offenders and mitigate theft risks.[92] In Brazil, Jockey Plaza shopping center reported a 50% reduction in theft incidents following implementation in 2023.[93] This application has gained traction amid rising retail shrinkage, with systems providing investigative efficiency and visibility into organized retail crime impacts.[94][95] Beyond security, facial recognition facilitates personalized customer experiences by estimating demographics like age and gender for targeted advertising and promotions, integrating with customer relationship management systems.[96][97] Loyalty program members can be automatically recognized at entry or checkout, triggering customized offers and streamlining interactions.[98] In theme parks, which incorporate extensive retail elements, Disney has deployed facial recognition since 2021 for guest entry and personalization, enhancing operational flow in shops and concessions by linking to prior visit data.[99][100] For payments, facial recognition enables frictionless biometric authentication at checkout, reducing transaction times and fraud. In China, platforms like Alipay and WeChat Pay introduced face recognition payments around 2019, achieving high adoption for contactless retail transactions.[101] Globally, adoption lags due to trust and regulatory hurdles, though pilots report up to 10% higher purchase volumes and 95% approval rates among users.[102][103] These systems prioritize liveness detection to counter spoofing, supporting secure verification in high-volume retail environments.[104]Government Services and Identity Verification
Facial recognition systems facilitate identity verification in various government services, including airport security screenings, national identification authentication, and access to public benefits. In the United States, the Transportation Security Administration (TSA) deploys facial comparison technology at over 80 airports as of 2025 to match travelers' live facial images against photos on passports or identification documents, enabling voluntary biometric verification prior to boarding.[105] This process aims to confirm identity without retaining biometric data post-verification, though participation remains opt-in with manual checks available as alternatives.[106] The U.S. Department of Homeland Security (DHS) further integrates face recognition in biometric exit programs at international departure gates, capturing facial images of outbound travelers to verify against entry records and prevent overstays.[107] Recent expansions include electronic gates (eGates) tested at airports like Cincinnati/Northern Kentucky International, where systems match facial scans to identity documents and boarding passes for expedited processing, particularly for TSA PreCheck members.[108] Evaluations by the National Institute of Standards and Technology (NIST) indicate high efficacy in controlled settings, with error rates supporting successful verification for approximately 99.87% of travelers in flight boarding scenarios.[109] In India, the Aadhaar program, managed by the Unique Identification Authority of India (UIDAI), incorporates facial recognition for real-time authentication in government services such as welfare disbursements and identity updates.[110] The FaceRD mobile application enables users to verify identity via facial scans matched against Aadhaar biometric records, supporting offline and remote access without physical documents.[111] By 2025, enhancements including AI-driven facial authentication in the e-Aadhaar app have streamlined processes like address corrections and reduced fraud in public service delivery.[112] These implementations prioritize one-to-one matching for verification, distinguishing from broader identification searches to enhance service efficiency while relying on enrolled biometric templates.[113]Healthcare and Biometric Authentication
Facial recognition systems in healthcare primarily serve to authenticate patient identities at admission, during treatment, and for record access, addressing misidentification errors that affect up to 12% of hospital admissions and contribute to sentinel events. A 2024 clinical trial using deep learning-based facial recognition for patient verification reported accuracy rates exceeding 99% in controlled hospital settings, with no adverse safety incidents and high clinician acceptability, outperforming wristband-based methods in reliability.[114] Similarly, an open-source facial recognition implementation achieved first-match identification success for nearly 100% of patients in a 2020 study, demonstrating robustness for unique patient matching in diverse demographics.[115] For biometric authentication, these systems enable secure login to electronic health records (EHRs) and medical devices without physical tokens, reducing unauthorized access risks in high-stakes environments. A 2020 IEEE analysis highlighted facial recognition's utility in hospitals for staff and patient tracking, processing verifications faster than barcode or RFID alternatives while maintaining low false positive rates under varying lighting conditions.[116] Implementation studies using transfer learning models, such as VGGFace2 with SENet-50, have shown error rates below 1% for patient re-identification across sessions, supporting scalable deployment for medication dispensing and procedure verification.[117][118] In authentication workflows, facial biometrics integrate with mobile apps for remote patient verification, as evidenced by a 2019 study where a facial recognition app reduced identification discrepancies by 95% compared to manual checks, enhancing safety in outpatient and telemedicine scenarios.[119] This approach also counters fraud in insurance claims by linking biometric templates to treatment records, with peer-reviewed reviews confirming facial modalities' cost-effectiveness and accessibility over iris or fingerprint alternatives due to non-contact operation.[120] Clinical evaluations, including a 2020 study on multimodal biometrics, affirm that facial systems yield verification times under 2 seconds with error rates under 0.5%, making them viable for real-time authentication in procedure matching.[121]Performance and Efficacy
Empirical Accuracy Metrics from Benchmarks
The National Institute of Standards and Technology (NIST) Face Recognition Vendor Test (FRVT), now known as Face Recognition Technology Evaluation (FRTE), serves as the primary independent benchmark for assessing facial recognition algorithm accuracy, evaluating over 1,300 algorithms from hundreds of developers as of 2025.[63] In 1:1 verification tasks, performance is measured by false non-match rate (FNMR) at fixed low false match rates (FMR), such as 10^{-4} or 10^{-6}, using datasets like mugshots, visas, and border images. Leading algorithms achieve FNMR values below 0.1% on high-quality mugshot datasets at FMR=10^{-6}, corresponding to over 99.9% verification accuracy, with recent submissions from 2025 showing FNMR as low as 0.0031 (99.69% accuracy) on visa datasets at FMR=10^{-6}.[63][122] For 1:N identification in large galleries (e.g., millions of entries simulating watchlists), metrics include false negative identification rate (FNIR) at low false positive identification rates (FPIR), such as 0.003. Top-performing algorithms in 2025 evaluations report FNIR around 0.12% at FPIR=0.3%, yielding true positive rates exceeding 99.8%, though performance degrades with gallery size and image quality; for instance, one vendor achieved 99.93% accuracy in border identification scenarios.[123][124][125] These rates reflect automated thresholding; investigative modes returning top candidates (e.g., 50 per probe) further reduce errors but increase manual review needs.[123]| Benchmark Type | Key Metric | Top 2025 Performance Example | Dataset/Context | Source |
|---|---|---|---|---|
| 1:1 Verification | FNMR at FMR=10^{-6} | <0.1% (e.g., 0.0031 on visas) | Mugshots/visas/borders | NIST FRTE[63] |
| 1:N Identification | FNIR at FPIR=0.003 | ~0.12% | Large galleries, border images | NIST FRTE/Neurotechnology[123][124] |
Real-World Success Rates and Case Studies
In border security and aviation applications, facial recognition systems have demonstrated high operational success rates. The U.S. Department of Homeland Security reported that fully operational face recognition systems for airport and port identity verification achieved success rates exceeding 99% in 2024 testing.[107] The Transportation Security Administration's Credential Authentication Technology verified identities with 100% performance across all demographic groups, including variations in skin tone, race, gender, and age.[107] U.S. Customs and Border Protection systems maintained success rates of at least 97% for all demographics in face matching, with differences between groups not exceeding 1-3%.[107] In law enforcement, real-world deployments have yielded investigative leads and identifications, though performance varies by probe quality and database size. The FBI's Next Generation Identification Interstate Photo System (NGI-IPS) validated an 85% accuracy rate in internal 2017 testing for returning candidate lists from probe images against gallery databases.[75] Vendor algorithms integrated into similar systems achieved 99.12% Rank 1 accuracy in the 2018 NIST Facial Recognition Vendor Test on controlled datasets.[75] From fiscal year 2017 to April 2019, the FBI processed 152,565 facial recognition searches without reported civil liberties violations.[75] Case studies highlight practical successes in criminal investigations:- In Scranton, Pennsylvania, police used facial recognition to identify a sexual assault suspect from surveillance footage, leading to an arrest.[127]
- Arvada Police Department in Colorado applied facial recognition in 73 investigations in 2024, generating 39 positive matches that advanced cases.[128]
- In Fairfax County, Virginia, officers identified a child sex trafficking suspect by querying a social media photo against databases.[129]
- California authorities rescued a missing child trafficked for weeks using facial recognition tools to match images from online ads.[129]
Advantages Relative to Alternative Biometrics
Facial recognition systems provide non-intrusive authentication, requiring no physical contact with a sensor, unlike fingerprint or palm vein scanning which demand direct touch and can be affected by skin conditions, dirt, or protective gloves. This contactless nature enhances hygiene, particularly in shared or high-traffic environments, and supports rapid processing without user cooperation, enabling deployment in scenarios like border crossings or crowd surveillance where alternatives such as iris or retina scanning necessitate close proximity to specialized hardware.[130][131] The technology operates passively using standard visible-spectrum cameras, allowing identification at distances up to several meters, in contrast to voice recognition, which is susceptible to environmental noise, accents, or temporary vocal changes, and gait analysis, which requires extended observation periods and controlled conditions for reliable matching.[130][132] User acceptance is notably higher for facial recognition due to its hands-free operation and familiarity, as individuals routinely encounter cameras in daily life, reducing resistance compared to invasive methods like retina scanning that involve discomfort from infrared illumination or prolonged eye fixation. Empirical benchmarks indicate facial systems achieve verification times under 1 second in controlled settings, outperforming contact-based biometrics in throughput for large-scale applications, such as airport e-gates processing over 1,000 passengers per hour without physical token handling.[130][133] Deployment costs are lowered by leveraging ubiquitous CCTV infrastructure, avoiding the need for dedicated scanners required by iris or fingerprint systems, which can exceed $100 per unit for high-security variants.[134] In terms of scalability for identification against large databases (1:N matching), facial recognition excels in real-time processing of video feeds, identifying individuals in motion or crowds, whereas alternatives like fingerprints require pre-captured templates and manual enrollment, limiting passive surveillance efficacy. National Institute of Standards and Technology evaluations confirm top-performing facial algorithms maintain false non-match rates below 0.1% on datasets exceeding 12 million images, supporting its edge in operational efficiency over modalities constrained by enrollment logistics or spoofing vulnerabilities, such as voice mimicry via recordings.[2][135]Limitations and Technical Challenges
Environmental and Operational Constraints
Facial recognition systems are highly sensitive to lighting variations, with performance degrading significantly in low-light or uneven illumination conditions where facial landmarks become obscured or distorted. Algorithms trained on well-lit datasets often fail to generalize, leading to false negatives or mismatches as shadows alter feature extraction from eyes, nose, and mouth. For instance, studies indicate that recognition accuracy can drop by over 20-50% in dim environments without supplemental infrared illumination.[136][137] Adverse weather further exacerbates these issues; rain, fog, snow, or glare from sunlight degrade image quality through scattering, blurring, or reduced contrast, limiting effective deployment in outdoor surveillance. Experimental evaluations under non-ideal weather demonstrate substantial limitations in resolution and feature clarity at distances beyond 50 meters, where systems optimized for clear conditions exhibit error rates exceeding 30%.[138][139] Pose and viewing angle impose operational constraints, as most systems achieve peak accuracy only with frontal, near-orthogonal faces; off-angle views up to 45 degrees introduce geometric distortions that misalign trained models, reducing match rates by 10-40% depending on the algorithm. Occlusions from accessories like hats, glasses, masks, or hair further compound this, presenting a persistent challenge since partial blockage hides critical features, with surveys reporting accuracy declines of up to 70% under heavy occlusion without specialized robust training.[140][141] Distance and subject motion add dynamic operational hurdles; at ranges over 10-20 meters, pixel resolution falls below thresholds for reliable detection, while rapid movement causes blur, particularly in real-time video feeds where processing latency can exceed 100ms on standard hardware.[142][143] Environmental factors such as humidity, dust, or temperature fluctuations indirectly affect camera sensors and infrared components, potentially introducing noise or thermal distortions in systems relying on multispectral imaging. Operational scalability is constrained by computational demands; large-scale deployments require high-throughput processing, yet unconstrained environments amplify false positives when gallery sizes exceed 1 million entries, straining resources without edge computing optimizations. These constraints underscore the need for hybrid approaches, like combining visible and thermal imaging, though empirical tests confirm persistent vulnerabilities in fully uncontrolled settings.[144][145][140]Error Rates and Demographic Performance Variations
Facial recognition systems demonstrate variations in error rates across demographic categories, including sex, age, and race/ethnicity, as quantified in the U.S. National Institute of Standards and Technology (NIST) Face Recognition Vendor Test (FRVT) evaluations. These assessments measure false match rates (FMR, erroneous identifications) and false non-match rates (FNMR, missed matches) using large-scale datasets of operational images. In NIST's FRVT Part 3 report from December 2019, analysis of 189 algorithms revealed that demographic differentials arise primarily from disparities in image quality between probe and gallery photos, rather than algorithmic discrimination, with darker-skinned individuals often facing lower-quality images due to factors like lighting and camera calibration.[2] Leading algorithms exhibited minimal differentials, with FMR ratios (worst-to-best across demographics) often below 10 for top performers, contrasting with poorer systems showing up to 100-fold increases in FMR for Asian and African American faces relative to Caucasian faces at low base rates (e.g., 0.0001 FMR threshold).[2][146] Subsequent NIST evaluations, including FRVT Part 8 updated through 2022 and summarized as of March 2025, confirm that high-performing algorithms continue to narrow these gaps, with FMR ratios approaching 1:1 across five racial/ethnic groups (e.g., Caucasian, Asian, African American, Hispanic, Indigenous) when evaluated at a 0.00003 FMR threshold.[7][10] For FNMR, computed at a 0.00001 FMR threshold using medium- versus high-quality images, differentials persist but are attenuated in state-of-the-art systems, attributed to anatomical variations (e.g., facial structure differences) and training data imbalances rather than intentional bias.[7] By sex, many algorithms show elevated FMR for females (up to 2-5 times higher in mid-tier systems), linked to variables like hairstyles, accessories, and softer facial contours affecting feature extraction, though top vendors achieve near-equivalence. Peer-reviewed studies have further identified facial hair as a key contributor to these sex-based variations, with beards and facial hair significantly reducing accuracy by altering facial landmarks, resulting in higher FNMR and gaps between clean-shaven and bearded faces. The mustache area imposes a particularly strong negative effect compared to chin or side beards, with impacts varying across demographics, such as greater effects among Caucasians in some analyses. These hairstyle-induced biases can be mitigated through balanced training data incorporating more facial hair examples or adaptive thresholding techniques.[7][147][148][149] Age-related variations are pronounced at extremes: FNMR and FMR increase for individuals under 18 or over 65, with ratios up to 10-20 times higher in some evaluations, due to developmental changes in facial morphology (e.g., softer features in children) and age-related alterations like wrinkles or sagging skin that challenge landmark detection.[7][150] NIST data indicate that these effects are not uniform across algorithms; for instance, certain systems exhibit lower errors for non-Caucasian groups when trained on diverse datasets, underscoring that performance disparities stem from empirical factors like dataset composition and image acquisition protocols rather than systemic racial animus.[2] Ongoing FRVT updates as of 2025 show progressive mitigation, with the best 1:1 verification algorithms achieving FNMR below 0.1% overall and demographic ratios under 2 for most categories, reflecting advancements in deep learning architectures that prioritize robustness over demographic proxies.[7] Independent peer-reviewed analyses corroborate these trends, noting that while early commercial systems (pre-2018) displayed stark disparities—e.g., error rates up to 34.7% for darker-skinned females versus 0.8% for lighter-skinned males—modern benchmarks reveal convergence driven by balanced training corpora.[151][152]| Demographic Factor | Typical FMR Differential (Top Algorithms, Ratio Worst/Best) | Key Contributing Factors | Source |
|---|---|---|---|
| Sex (Female vs. Male) | 1-5 | Hairstyle variability, facial softness | [7] |
| Age (<18 or >65) | 5-20 | Morphological changes, reduced distinctiveness | [7] |
| Race/Ethnicity (Non-Caucasian vs. Caucasian) | 1-10 (improving to ~1) | Image quality disparities, training underrepresentation | [2][7] |