Cargando…

Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition

The recognition performance of visual recognition systems is highly dependent on extracting and representing the discriminative characteristics of image data. Convolutional neural networks (CNNs) have shown unprecedented success in a variety of visual recognition tasks due to their capability to pro...

Descripción completa

Detalles Bibliográficos
Autores principales: Alshazly, Hammam, Linse, Christoph, Barth, Erhardt, Martinetz, Thomas
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6806105/
https://www.ncbi.nlm.nih.gov/pubmed/31554303
http://dx.doi.org/10.3390/s19194139
_version_ 1783461552310976512
author Alshazly, Hammam
Linse, Christoph
Barth, Erhardt
Martinetz, Thomas
author_facet Alshazly, Hammam
Linse, Christoph
Barth, Erhardt
Martinetz, Thomas
author_sort Alshazly, Hammam
collection PubMed
description The recognition performance of visual recognition systems is highly dependent on extracting and representing the discriminative characteristics of image data. Convolutional neural networks (CNNs) have shown unprecedented success in a variety of visual recognition tasks due to their capability to provide in-depth representations exploiting visual image features of appearance, color, and texture. This paper presents a novel system for ear recognition based on ensembles of deep CNN-based models and more specifically the Visual Geometry Group (VGG)-like network architectures for extracting discriminative deep features from ear images. We began by training different networks of increasing depth on ear images with random weight initialization. Then, we examined pretrained models as feature extractors as well as fine-tuning them on ear images. After that, we built ensembles of the best models to further improve the recognition performance. We evaluated the proposed ensembles through identification experiments using ear images acquired under controlled and uncontrolled conditions from mathematical analysis of images (AMI), AMI cropped (AMIC) (introduced here), and West Pomeranian University of Technology (WPUT) ear datasets. The experimental results indicate that our ensembles of models yield the best performance with significant improvements over the recently published results. Moreover, we provide visual explanations of the learned features by highlighting the relevant image regions utilized by the models for making decisions or predictions.
format Online
Article
Text
id pubmed-6806105
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-68061052019-11-07 Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition Alshazly, Hammam Linse, Christoph Barth, Erhardt Martinetz, Thomas Sensors (Basel) Article The recognition performance of visual recognition systems is highly dependent on extracting and representing the discriminative characteristics of image data. Convolutional neural networks (CNNs) have shown unprecedented success in a variety of visual recognition tasks due to their capability to provide in-depth representations exploiting visual image features of appearance, color, and texture. This paper presents a novel system for ear recognition based on ensembles of deep CNN-based models and more specifically the Visual Geometry Group (VGG)-like network architectures for extracting discriminative deep features from ear images. We began by training different networks of increasing depth on ear images with random weight initialization. Then, we examined pretrained models as feature extractors as well as fine-tuning them on ear images. After that, we built ensembles of the best models to further improve the recognition performance. We evaluated the proposed ensembles through identification experiments using ear images acquired under controlled and uncontrolled conditions from mathematical analysis of images (AMI), AMI cropped (AMIC) (introduced here), and West Pomeranian University of Technology (WPUT) ear datasets. The experimental results indicate that our ensembles of models yield the best performance with significant improvements over the recently published results. Moreover, we provide visual explanations of the learned features by highlighting the relevant image regions utilized by the models for making decisions or predictions. MDPI 2019-09-24 /pmc/articles/PMC6806105/ /pubmed/31554303 http://dx.doi.org/10.3390/s19194139 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Alshazly, Hammam
Linse, Christoph
Barth, Erhardt
Martinetz, Thomas
Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition
title Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition
title_full Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition
title_fullStr Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition
title_full_unstemmed Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition
title_short Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition
title_sort ensembles of deep learning models and transfer learning for ear recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6806105/
https://www.ncbi.nlm.nih.gov/pubmed/31554303
http://dx.doi.org/10.3390/s19194139
work_keys_str_mv AT alshazlyhammam ensemblesofdeeplearningmodelsandtransferlearningforearrecognition
AT linsechristoph ensemblesofdeeplearningmodelsandtransferlearningforearrecognition
AT bartherhardt ensemblesofdeeplearningmodelsandtransferlearningforearrecognition
AT martinetzthomas ensemblesofdeeplearningmodelsandtransferlearningforearrecognition