Cargando…

Model-size reduction for reservoir computing by concatenating internal states through time

Reservoir computing (RC) is a machine learning algorithm that can learn complex time series from data very rapidly based on the use of high-dimensional dynamical systems, such as random networks of neurons, called “reservoirs.” To implement RC in edge computing, it is highly important to reduce the...

Descripción completa

Detalles Bibliográficos
Autores principales: Sakemi, Yusuke, Morino, Kai, Leleu, Timothée, Aihara, Kazuyuki
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7733507/
https://www.ncbi.nlm.nih.gov/pubmed/33311595
http://dx.doi.org/10.1038/s41598-020-78725-0
_version_ 1783622287344271360
author Sakemi, Yusuke
Morino, Kai
Leleu, Timothée
Aihara, Kazuyuki
author_facet Sakemi, Yusuke
Morino, Kai
Leleu, Timothée
Aihara, Kazuyuki
author_sort Sakemi, Yusuke
collection PubMed
description Reservoir computing (RC) is a machine learning algorithm that can learn complex time series from data very rapidly based on the use of high-dimensional dynamical systems, such as random networks of neurons, called “reservoirs.” To implement RC in edge computing, it is highly important to reduce the amount of computational resources that RC requires. In this study, we propose methods that reduce the size of the reservoir by inputting the past or drifting states of the reservoir to the output layer at the current time step. To elucidate the mechanism of model-size reduction, the proposed methods are analyzed based on information processing capacity proposed by Dambre et al. (Sci Rep 2:514, 2012). In addition, we evaluate the effectiveness of the proposed methods on time-series prediction tasks: the generalized Hénon-map and NARMA. On these tasks, we found that the proposed methods were able to reduce the size of the reservoir up to one tenth without a substantial increase in regression error.
format Online
Article
Text
id pubmed-7733507
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-77335072020-12-15 Model-size reduction for reservoir computing by concatenating internal states through time Sakemi, Yusuke Morino, Kai Leleu, Timothée Aihara, Kazuyuki Sci Rep Article Reservoir computing (RC) is a machine learning algorithm that can learn complex time series from data very rapidly based on the use of high-dimensional dynamical systems, such as random networks of neurons, called “reservoirs.” To implement RC in edge computing, it is highly important to reduce the amount of computational resources that RC requires. In this study, we propose methods that reduce the size of the reservoir by inputting the past or drifting states of the reservoir to the output layer at the current time step. To elucidate the mechanism of model-size reduction, the proposed methods are analyzed based on information processing capacity proposed by Dambre et al. (Sci Rep 2:514, 2012). In addition, we evaluate the effectiveness of the proposed methods on time-series prediction tasks: the generalized Hénon-map and NARMA. On these tasks, we found that the proposed methods were able to reduce the size of the reservoir up to one tenth without a substantial increase in regression error. Nature Publishing Group UK 2020-12-11 /pmc/articles/PMC7733507/ /pubmed/33311595 http://dx.doi.org/10.1038/s41598-020-78725-0 Text en © The Author(s) 2020 Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
spellingShingle Article
Sakemi, Yusuke
Morino, Kai
Leleu, Timothée
Aihara, Kazuyuki
Model-size reduction for reservoir computing by concatenating internal states through time
title Model-size reduction for reservoir computing by concatenating internal states through time
title_full Model-size reduction for reservoir computing by concatenating internal states through time
title_fullStr Model-size reduction for reservoir computing by concatenating internal states through time
title_full_unstemmed Model-size reduction for reservoir computing by concatenating internal states through time
title_short Model-size reduction for reservoir computing by concatenating internal states through time
title_sort model-size reduction for reservoir computing by concatenating internal states through time
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7733507/
https://www.ncbi.nlm.nih.gov/pubmed/33311595
http://dx.doi.org/10.1038/s41598-020-78725-0
work_keys_str_mv AT sakemiyusuke modelsizereductionforreservoircomputingbyconcatenatinginternalstatesthroughtime
AT morinokai modelsizereductionforreservoircomputingbyconcatenatinginternalstatesthroughtime
AT leleutimothee modelsizereductionforreservoircomputingbyconcatenatinginternalstatesthroughtime
AT aiharakazuyuki modelsizereductionforreservoircomputingbyconcatenatinginternalstatesthroughtime