Cargando…
XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks
Face Recognition aims at identifying or confirming an individual’s identity in a still image or video. Towards this end, machine learning and deep learning techniques have been successfully employed for face recognition. However, the response of the face recognition system often remains mysterious t...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer US
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9745692/ https://www.ncbi.nlm.nih.gov/pubmed/36531522 http://dx.doi.org/10.1007/s11277-022-10127-z |
_version_ | 1784849203258720256 |
---|---|
author | Rajpal, Ankit Sehra, Khushwant Bagri, Rashika Sikka, Pooja |
author_facet | Rajpal, Ankit Sehra, Khushwant Bagri, Rashika Sikka, Pooja |
author_sort | Rajpal, Ankit |
collection | PubMed |
description | Face Recognition aims at identifying or confirming an individual’s identity in a still image or video. Towards this end, machine learning and deep learning techniques have been successfully employed for face recognition. However, the response of the face recognition system often remains mysterious to the end-user. This paper aims to fill this gap by letting an end user know which features of the face has the model relied upon in recognizing a subject’s face. In this context, we evaluate the interpretability of several face recognizers employing deep neural networks namely, LeNet-5, AlexNet, Inception-V3, and VGG16. For this purpose, a recently proposed explainable AI tool–Local Interpretable Model-Agnostic Explanations (LIME) is used. Benchmark datasets such as Yale, AT &T dataset, and Labeled Faces in the Wild (LFW) are utilized for this purpose. We are able to demonstrate that LIME indeed marks the features that are visually significant features for face recognition. |
format | Online Article Text |
id | pubmed-9745692 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Springer US |
record_format | MEDLINE/PubMed |
spelling | pubmed-97456922022-12-13 XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks Rajpal, Ankit Sehra, Khushwant Bagri, Rashika Sikka, Pooja Wirel Pers Commun Article Face Recognition aims at identifying or confirming an individual’s identity in a still image or video. Towards this end, machine learning and deep learning techniques have been successfully employed for face recognition. However, the response of the face recognition system often remains mysterious to the end-user. This paper aims to fill this gap by letting an end user know which features of the face has the model relied upon in recognizing a subject’s face. In this context, we evaluate the interpretability of several face recognizers employing deep neural networks namely, LeNet-5, AlexNet, Inception-V3, and VGG16. For this purpose, a recently proposed explainable AI tool–Local Interpretable Model-Agnostic Explanations (LIME) is used. Benchmark datasets such as Yale, AT &T dataset, and Labeled Faces in the Wild (LFW) are utilized for this purpose. We are able to demonstrate that LIME indeed marks the features that are visually significant features for face recognition. Springer US 2022-12-13 2023 /pmc/articles/PMC9745692/ /pubmed/36531522 http://dx.doi.org/10.1007/s11277-022-10127-z Text en © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022, Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Rajpal, Ankit Sehra, Khushwant Bagri, Rashika Sikka, Pooja XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks |
title | XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks |
title_full | XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks |
title_fullStr | XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks |
title_full_unstemmed | XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks |
title_short | XAI-FR: Explainable AI-Based Face Recognition Using Deep Neural Networks |
title_sort | xai-fr: explainable ai-based face recognition using deep neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9745692/ https://www.ncbi.nlm.nih.gov/pubmed/36531522 http://dx.doi.org/10.1007/s11277-022-10127-z |
work_keys_str_mv | AT rajpalankit xaifrexplainableaibasedfacerecognitionusingdeepneuralnetworks AT sehrakhushwant xaifrexplainableaibasedfacerecognitionusingdeepneuralnetworks AT bagrirashika xaifrexplainableaibasedfacerecognitionusingdeepneuralnetworks AT sikkapooja xaifrexplainableaibasedfacerecognitionusingdeepneuralnetworks |