Cargando…
Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes
Deep artificial neural networks are feed-forward architectures capable of very impressive performances in diverse domains. Indeed stacking multiple layers allows a hierarchical composition of local functions, providing efficient compact mappings. Compared to the brain, however, such architectures ar...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6333865/ https://www.ncbi.nlm.nih.gov/pubmed/30687053 http://dx.doi.org/10.3389/fncom.2018.00100 |
_version_ | 1783387635897597952 |
---|---|
author | Drumond, Thalita F. Viéville, Thierry Alexandre, Frédéric |
author_facet | Drumond, Thalita F. Viéville, Thierry Alexandre, Frédéric |
author_sort | Drumond, Thalita F. |
collection | PubMed |
description | Deep artificial neural networks are feed-forward architectures capable of very impressive performances in diverse domains. Indeed stacking multiple layers allows a hierarchical composition of local functions, providing efficient compact mappings. Compared to the brain, however, such architectures are closer to a single pipeline and require huge amounts of data, while concrete cases for either human or machine learning systems are often restricted to not-so-big data sets. Furthermore, interpretability of the obtained results is a key issue: since deep learning applications are increasingly present in society, it is important that the underlying processes be accessible and understandable to every one. In order to target these challenges, in this contribution we analyze how considering prototypes in a rather generalized sense (with respect to the state of the art) allows to reasonably work with small data sets while providing an interpretable view of the obtained results. Some mathematical interpretation of this proposal is discussed. Sensitivity to hyperparameters is a key issue for reproducible deep learning results, and is carefully considered in our methodology. Performances and limitations of the proposed setup are explored in details, under different hyperparameter sets, in an analogous way as biological experiments are conducted. We obtain a rather simple architecture, easy to explain, and which allows, combined with a standard method, to target both performances and interpretability. |
format | Online Article Text |
id | pubmed-6333865 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-63338652019-01-25 Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes Drumond, Thalita F. Viéville, Thierry Alexandre, Frédéric Front Comput Neurosci Neuroscience Deep artificial neural networks are feed-forward architectures capable of very impressive performances in diverse domains. Indeed stacking multiple layers allows a hierarchical composition of local functions, providing efficient compact mappings. Compared to the brain, however, such architectures are closer to a single pipeline and require huge amounts of data, while concrete cases for either human or machine learning systems are often restricted to not-so-big data sets. Furthermore, interpretability of the obtained results is a key issue: since deep learning applications are increasingly present in society, it is important that the underlying processes be accessible and understandable to every one. In order to target these challenges, in this contribution we analyze how considering prototypes in a rather generalized sense (with respect to the state of the art) allows to reasonably work with small data sets while providing an interpretable view of the obtained results. Some mathematical interpretation of this proposal is discussed. Sensitivity to hyperparameters is a key issue for reproducible deep learning results, and is carefully considered in our methodology. Performances and limitations of the proposed setup are explored in details, under different hyperparameter sets, in an analogous way as biological experiments are conducted. We obtain a rather simple architecture, easy to explain, and which allows, combined with a standard method, to target both performances and interpretability. Frontiers Media S.A. 2019-01-09 /pmc/articles/PMC6333865/ /pubmed/30687053 http://dx.doi.org/10.3389/fncom.2018.00100 Text en Copyright © 2019 Drumond, Viéville and Alexandre. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Drumond, Thalita F. Viéville, Thierry Alexandre, Frédéric Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes |
title | Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes |
title_full | Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes |
title_fullStr | Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes |
title_full_unstemmed | Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes |
title_short | Bio-inspired Analysis of Deep Learning on Not-So-Big Data Using Data-Prototypes |
title_sort | bio-inspired analysis of deep learning on not-so-big data using data-prototypes |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6333865/ https://www.ncbi.nlm.nih.gov/pubmed/30687053 http://dx.doi.org/10.3389/fncom.2018.00100 |
work_keys_str_mv | AT drumondthalitaf bioinspiredanalysisofdeeplearningonnotsobigdatausingdataprototypes AT vievillethierry bioinspiredanalysisofdeeplearningonnotsobigdatausingdataprototypes AT alexandrefrederic bioinspiredanalysisofdeeplearningonnotsobigdatausingdataprototypes |