Cargando…

Error Rates in Users of Automatic Face Recognition Software

In recent years, wide deployment of automatic face recognition systems has been accompanied by substantial gains in algorithm performance. However, benchmarking tests designed to evaluate these systems do not account for the errors of human operators, who are often an integral part of face recogniti...

Descripción completa

Detalles Bibliográficos
Autores principales: White, David, Dunn, James D., Schmid, Alexandra C., Kemp, Richard I.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2015
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4605725/
https://www.ncbi.nlm.nih.gov/pubmed/26465631
http://dx.doi.org/10.1371/journal.pone.0139827
_version_ 1782395249062051840
author White, David
Dunn, James D.
Schmid, Alexandra C.
Kemp, Richard I.
author_facet White, David
Dunn, James D.
Schmid, Alexandra C.
Kemp, Richard I.
author_sort White, David
collection PubMed
description In recent years, wide deployment of automatic face recognition systems has been accompanied by substantial gains in algorithm performance. However, benchmarking tests designed to evaluate these systems do not account for the errors of human operators, who are often an integral part of face recognition solutions in forensic and security settings. This causes a mismatch between evaluation tests and operational accuracy. We address this by measuring user performance in a face recognition system used to screen passport applications for identity fraud. Experiment 1 measured target detection accuracy in algorithm-generated ‘candidate lists’ selected from a large database of passport images. Accuracy was notably poorer than in previous studies of unfamiliar face matching: participants made over 50% errors for adult target faces, and over 60% when matching images of children. Experiment 2 then compared performance of student participants to trained passport officers–who use the system in their daily work–and found equivalent performance in these groups. Encouragingly, a group of highly trained and experienced “facial examiners” outperformed these groups by 20 percentage points. We conclude that human performance curtails accuracy of face recognition systems–potentially reducing benchmark estimates by 50% in operational settings. Mere practise does not attenuate these limits, but superior performance of trained examiners suggests that recruitment and selection of human operators, in combination with effective training and mentorship, can improve the operational accuracy of face recognition systems.
format Online
Article
Text
id pubmed-4605725
institution National Center for Biotechnology Information
language English
publishDate 2015
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-46057252015-10-29 Error Rates in Users of Automatic Face Recognition Software White, David Dunn, James D. Schmid, Alexandra C. Kemp, Richard I. PLoS One Research Article In recent years, wide deployment of automatic face recognition systems has been accompanied by substantial gains in algorithm performance. However, benchmarking tests designed to evaluate these systems do not account for the errors of human operators, who are often an integral part of face recognition solutions in forensic and security settings. This causes a mismatch between evaluation tests and operational accuracy. We address this by measuring user performance in a face recognition system used to screen passport applications for identity fraud. Experiment 1 measured target detection accuracy in algorithm-generated ‘candidate lists’ selected from a large database of passport images. Accuracy was notably poorer than in previous studies of unfamiliar face matching: participants made over 50% errors for adult target faces, and over 60% when matching images of children. Experiment 2 then compared performance of student participants to trained passport officers–who use the system in their daily work–and found equivalent performance in these groups. Encouragingly, a group of highly trained and experienced “facial examiners” outperformed these groups by 20 percentage points. We conclude that human performance curtails accuracy of face recognition systems–potentially reducing benchmark estimates by 50% in operational settings. Mere practise does not attenuate these limits, but superior performance of trained examiners suggests that recruitment and selection of human operators, in combination with effective training and mentorship, can improve the operational accuracy of face recognition systems. Public Library of Science 2015-10-14 /pmc/articles/PMC4605725/ /pubmed/26465631 http://dx.doi.org/10.1371/journal.pone.0139827 Text en © 2015 White et al http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
White, David
Dunn, James D.
Schmid, Alexandra C.
Kemp, Richard I.
Error Rates in Users of Automatic Face Recognition Software
title Error Rates in Users of Automatic Face Recognition Software
title_full Error Rates in Users of Automatic Face Recognition Software
title_fullStr Error Rates in Users of Automatic Face Recognition Software
title_full_unstemmed Error Rates in Users of Automatic Face Recognition Software
title_short Error Rates in Users of Automatic Face Recognition Software
title_sort error rates in users of automatic face recognition software
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4605725/
https://www.ncbi.nlm.nih.gov/pubmed/26465631
http://dx.doi.org/10.1371/journal.pone.0139827
work_keys_str_mv AT whitedavid errorratesinusersofautomaticfacerecognitionsoftware
AT dunnjamesd errorratesinusersofautomaticfacerecognitionsoftware
AT schmidalexandrac errorratesinusersofautomaticfacerecognitionsoftware
AT kemprichardi errorratesinusersofautomaticfacerecognitionsoftware