Cargando…

Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes

Deep artificial neural networks are feed-forward architectures capable of very impressive performances in diverse domains. Indeed stacking multiple layers allows a hierarchical composition of local functions, providing efficient compact mappings. Compared to the brain, however, such architectures ar...

Descripción completa

Detalles Bibliográficos
Autores principales: Drumond, Thalita F., Viéville, Thierry, Alexandre, Frédéric
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6333865/
https://www.ncbi.nlm.nih.gov/pubmed/30687053
http://dx.doi.org/10.3389/fncom.2018.00100
_version_ 1783387635897597952
author Drumond, Thalita F.
Viéville, Thierry
Alexandre, Frédéric
author_facet Drumond, Thalita F.
Viéville, Thierry
Alexandre, Frédéric
author_sort Drumond, Thalita F.
collection PubMed
description Deep artificial neural networks are feed-forward architectures capable of very impressive performances in diverse domains. Indeed stacking multiple layers allows a hierarchical composition of local functions, providing efficient compact mappings. Compared to the brain, however, such architectures are closer to a single pipeline and require huge amounts of data, while concrete cases for either human or machine learning systems are often restricted to not-so-big data sets. Furthermore, interpretability of the obtained results is a key issue: since deep learning applications are increasingly present in society, it is important that the underlying processes be accessible and understandable to every one. In order to target these challenges, in this contribution we analyze how considering prototypes in a rather generalized sense (with respect to the state of the art) allows to reasonably work with small data sets while providing an interpretable view of the obtained results. Some mathematical interpretation of this proposal is discussed. Sensitivity to hyperparameters is a key issue for reproducible deep learning results, and is carefully considered in our methodology. Performances and limitations of the proposed setup are explored in details, under different hyperparameter sets, in an analogous way as biological experiments are conducted. We obtain a rather simple architecture, easy to explain, and which allows, combined with a standard method, to target both performances and interpretability.
format Online
Article
Text
id pubmed-6333865
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-63338652019-01-25 Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes Drumond, Thalita F. Viéville, Thierry Alexandre, Frédéric Front Comput Neurosci Neuroscience Deep artificial neural networks are feed-forward architectures capable of very impressive performances in diverse domains. Indeed stacking multiple layers allows a hierarchical composition of local functions, providing efficient compact mappings. Compared to the brain, however, such architectures are closer to a single pipeline and require huge amounts of data, while concrete cases for either human or machine learning systems are often restricted to not-so-big data sets. Furthermore, interpretability of the obtained results is a key issue: since deep learning applications are increasingly present in society, it is important that the underlying processes be accessible and understandable to every one. In order to target these challenges, in this contribution we analyze how considering prototypes in a rather generalized sense (with respect to the state of the art) allows to reasonably work with small data sets while providing an interpretable view of the obtained results. Some mathematical interpretation of this proposal is discussed. Sensitivity to hyperparameters is a key issue for reproducible deep learning results, and is carefully considered in our methodology. Performances and limitations of the proposed setup are explored in details, under different hyperparameter sets, in an analogous way as biological experiments are conducted. We obtain a rather simple architecture, easy to explain, and which allows, combined with a standard method, to target both performances and interpretability. Frontiers Media S.A. 2019-01-09 /pmc/articles/PMC6333865/ /pubmed/30687053 http://dx.doi.org/10.3389/fncom.2018.00100 Text en Copyright © 2019 Drumond, Viéville and Alexandre. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Drumond, Thalita F.
Viéville, Thierry
Alexandre, Frédéric
Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes
title Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes
title_full Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes
title_fullStr Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes
title_full_unstemmed Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes
title_short Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes
title_sort bio-inspired analysis of deep learning on not-so-big data using data-prototypes
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6333865/
https://www.ncbi.nlm.nih.gov/pubmed/30687053
http://dx.doi.org/10.3389/fncom.2018.00100
work_keys_str_mv AT drumondthalitaf bioinspiredanalysisofdeeplearningonnotsobigdatausingdataprototypes
AT vievillethierry bioinspiredanalysisofdeeplearningonnotsobigdatausingdataprototypes
AT alexandrefrederic bioinspiredanalysisofdeeplearningonnotsobigdatausingdataprototypes