Cargando…

Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models

This work addresses the challenge of building an accurate and generalizable periocular recognition model with a small number of learnable parameters. Deeper (larger) models are typically more capable of learning complex information. For this reason, knowledge distillation (kd) was previously propose...

Descripción completa

Detalles Bibliográficos
Autores principales: Boutros, Fadi, Damer, Naser, Raja, Kiran, Kirchbuchner, Florian, Kuijper, Arjan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8914924/
https://www.ncbi.nlm.nih.gov/pubmed/35271074
http://dx.doi.org/10.3390/s22051921
_version_ 1784667875659743232
author Boutros, Fadi
Damer, Naser
Raja, Kiran
Kirchbuchner, Florian
Kuijper, Arjan
author_facet Boutros, Fadi
Damer, Naser
Raja, Kiran
Kirchbuchner, Florian
Kuijper, Arjan
author_sort Boutros, Fadi
collection PubMed
description This work addresses the challenge of building an accurate and generalizable periocular recognition model with a small number of learnable parameters. Deeper (larger) models are typically more capable of learning complex information. For this reason, knowledge distillation (kd) was previously proposed to carry this knowledge from a large model (teacher) into a small model (student). Conventional KD optimizes the student output to be similar to the teacher output (commonly classification output). In biometrics, comparison (verification) and storage operations are conducted on biometric templates, extracted from pre-classification layers. In this work, we propose a novel template-driven KD approach that optimizes the distillation process so that the student model learns to produce templates similar to those produced by the teacher model. We demonstrate our approach on intra- and cross-device periocular verification. Our results demonstrate the superiority of our proposed approach over a network trained without KD and networks trained with conventional (vanilla) KD. For example, the targeted small model achieved an equal error rate (EER) value of 22.2% on cross-device verification without KD. The same model achieved an EER of 21.9% with the conventional KD, and only 14.7% EER when using our proposed template-driven KD.
format Online
Article
Text
id pubmed-8914924
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89149242022-03-12 Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models Boutros, Fadi Damer, Naser Raja, Kiran Kirchbuchner, Florian Kuijper, Arjan Sensors (Basel) Article This work addresses the challenge of building an accurate and generalizable periocular recognition model with a small number of learnable parameters. Deeper (larger) models are typically more capable of learning complex information. For this reason, knowledge distillation (kd) was previously proposed to carry this knowledge from a large model (teacher) into a small model (student). Conventional KD optimizes the student output to be similar to the teacher output (commonly classification output). In biometrics, comparison (verification) and storage operations are conducted on biometric templates, extracted from pre-classification layers. In this work, we propose a novel template-driven KD approach that optimizes the distillation process so that the student model learns to produce templates similar to those produced by the teacher model. We demonstrate our approach on intra- and cross-device periocular verification. Our results demonstrate the superiority of our proposed approach over a network trained without KD and networks trained with conventional (vanilla) KD. For example, the targeted small model achieved an equal error rate (EER) value of 22.2% on cross-device verification without KD. The same model achieved an EER of 21.9% with the conventional KD, and only 14.7% EER when using our proposed template-driven KD. MDPI 2022-03-01 /pmc/articles/PMC8914924/ /pubmed/35271074 http://dx.doi.org/10.3390/s22051921 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Boutros, Fadi
Damer, Naser
Raja, Kiran
Kirchbuchner, Florian
Kuijper, Arjan
Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models
title Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models
title_full Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models
title_fullStr Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models
title_full_unstemmed Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models
title_short Template-Driven Knowledge Distillation for Compact and Accurate Periocular Biometrics Deep-Learning Models
title_sort template-driven knowledge distillation for compact and accurate periocular biometrics deep-learning models
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8914924/
https://www.ncbi.nlm.nih.gov/pubmed/35271074
http://dx.doi.org/10.3390/s22051921
work_keys_str_mv AT boutrosfadi templatedrivenknowledgedistillationforcompactandaccurateperiocularbiometricsdeeplearningmodels
AT damernaser templatedrivenknowledgedistillationforcompactandaccurateperiocularbiometricsdeeplearningmodels
AT rajakiran templatedrivenknowledgedistillationforcompactandaccurateperiocularbiometricsdeeplearningmodels
AT kirchbuchnerflorian templatedrivenknowledgedistillationforcompactandaccurateperiocularbiometricsdeeplearningmodels
AT kuijperarjan templatedrivenknowledgedistillationforcompactandaccurateperiocularbiometricsdeeplearningmodels