Cargando…

Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere

Among the various architectures of Recurrent Neural Networks, Echo State Networks (ESNs) emerged due to their simplified and inexpensive training procedure. These networks are known to be sensitive to the setting of hyper-parameters, which critically affect their behavior. Results show that their pe...

Descripción completa

Detalles Bibliográficos
Autores principales: Verzelli, Pietro, Alippi, Cesare, Livi, Lorenzo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6761167/
https://www.ncbi.nlm.nih.gov/pubmed/31554855
http://dx.doi.org/10.1038/s41598-019-50158-4
_version_ 1783453969332305920
author Verzelli, Pietro
Alippi, Cesare
Livi, Lorenzo
author_facet Verzelli, Pietro
Alippi, Cesare
Livi, Lorenzo
author_sort Verzelli, Pietro
collection PubMed
description Among the various architectures of Recurrent Neural Networks, Echo State Networks (ESNs) emerged due to their simplified and inexpensive training procedure. These networks are known to be sensitive to the setting of hyper-parameters, which critically affect their behavior. Results show that their performance is usually maximized in a narrow region of hyper-parameter space called edge of criticality. Finding such a region requires searching in hyper-parameter space in a sensible way: hyper-parameter configurations marginally outside such a region might yield networks exhibiting fully developed chaos, hence producing unreliable computations. The performance gain due to optimizing hyper-parameters can be studied by considering the memory–nonlinearity trade-off, i.e., the fact that increasing the nonlinear behavior of the network degrades its ability to remember past inputs, and vice-versa. In this paper, we propose a model of ESNs that eliminates critical dependence on hyper-parameters, resulting in networks that provably cannot enter a chaotic regime and, at the same time, denotes nonlinear behavior in phase space characterized by a large memory of past inputs, comparable to the one of linear networks. Our contribution is supported by experiments corroborating our theoretical findings, showing that the proposed model displays dynamics that are rich-enough to approximate many common nonlinear systems used for benchmarking.
format Online
Article
Text
id pubmed-6761167
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-67611672019-11-12 Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere Verzelli, Pietro Alippi, Cesare Livi, Lorenzo Sci Rep Article Among the various architectures of Recurrent Neural Networks, Echo State Networks (ESNs) emerged due to their simplified and inexpensive training procedure. These networks are known to be sensitive to the setting of hyper-parameters, which critically affect their behavior. Results show that their performance is usually maximized in a narrow region of hyper-parameter space called edge of criticality. Finding such a region requires searching in hyper-parameter space in a sensible way: hyper-parameter configurations marginally outside such a region might yield networks exhibiting fully developed chaos, hence producing unreliable computations. The performance gain due to optimizing hyper-parameters can be studied by considering the memory–nonlinearity trade-off, i.e., the fact that increasing the nonlinear behavior of the network degrades its ability to remember past inputs, and vice-versa. In this paper, we propose a model of ESNs that eliminates critical dependence on hyper-parameters, resulting in networks that provably cannot enter a chaotic regime and, at the same time, denotes nonlinear behavior in phase space characterized by a large memory of past inputs, comparable to the one of linear networks. Our contribution is supported by experiments corroborating our theoretical findings, showing that the proposed model displays dynamics that are rich-enough to approximate many common nonlinear systems used for benchmarking. Nature Publishing Group UK 2019-09-25 /pmc/articles/PMC6761167/ /pubmed/31554855 http://dx.doi.org/10.1038/s41598-019-50158-4 Text en © The Author(s) 2019 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
spellingShingle Article
Verzelli, Pietro
Alippi, Cesare
Livi, Lorenzo
Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere
title Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere
title_full Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere
title_fullStr Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere
title_full_unstemmed Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere
title_short Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere
title_sort echo state networks with self-normalizing activations on the hyper-sphere
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6761167/
https://www.ncbi.nlm.nih.gov/pubmed/31554855
http://dx.doi.org/10.1038/s41598-019-50158-4
work_keys_str_mv AT verzellipietro echostatenetworkswithselfnormalizingactivationsonthehypersphere
AT alippicesare echostatenetworkswithselfnormalizingactivationsonthehypersphere
AT livilorenzo echostatenetworkswithselfnormalizingactivationsonthehypersphere