Cargando…

Stabilizing patterns in time: Neural network approach

Recurrent and feedback networks are capable of holding dynamic memories. Nonetheless, training a network for that task is challenging. In order to do so, one should face non-linear propagation of errors in the system. Small deviations from the desired dynamics due to error or inherent noise might ha...

Descripción completa

Detalles Bibliográficos
Autores principales: Ben-Shushan, Nadav, Tsodyks, Misha
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5741269/
https://www.ncbi.nlm.nih.gov/pubmed/29232710
http://dx.doi.org/10.1371/journal.pcbi.1005861
_version_ 1783288172705218560
author Ben-Shushan, Nadav
Tsodyks, Misha
author_facet Ben-Shushan, Nadav
Tsodyks, Misha
author_sort Ben-Shushan, Nadav
collection PubMed
description Recurrent and feedback networks are capable of holding dynamic memories. Nonetheless, training a network for that task is challenging. In order to do so, one should face non-linear propagation of errors in the system. Small deviations from the desired dynamics due to error or inherent noise might have a dramatic effect in the future. A method to cope with these difficulties is thus needed. In this work we focus on recurrent networks with linear activation functions and binary output unit. We characterize its ability to reproduce a temporal sequence of actions over its output unit. We suggest casting the temporal learning problem to a perceptron problem. In the discrete case a finite margin appears, providing the network, to some extent, robustness to noise, for which it performs perfectly (i.e. producing a desired sequence for an arbitrary number of cycles flawlessly). In the continuous case the margin approaches zero when the output unit changes its state, hence the network is only able to reproduce the sequence with slight jitters. Numerical simulation suggest that in the discrete time case, the longest sequence that can be learned scales, at best, as square root of the network size. A dramatic effect occurs when learning several short sequences in parallel, that is, their total length substantially exceeds the length of the longest single sequence the network can learn. This model easily generalizes to an arbitrary number of output units, which boost its performance. This effect is demonstrated by considering two practical examples for sequence learning. This work suggests a way to overcome stability problems for training recurrent networks and further quantifies the performance of a network under the specific learning scheme.
format Online
Article
Text
id pubmed-5741269
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-57412692018-01-10 Stabilizing patterns in time: Neural network approach Ben-Shushan, Nadav Tsodyks, Misha PLoS Comput Biol Research Article Recurrent and feedback networks are capable of holding dynamic memories. Nonetheless, training a network for that task is challenging. In order to do so, one should face non-linear propagation of errors in the system. Small deviations from the desired dynamics due to error or inherent noise might have a dramatic effect in the future. A method to cope with these difficulties is thus needed. In this work we focus on recurrent networks with linear activation functions and binary output unit. We characterize its ability to reproduce a temporal sequence of actions over its output unit. We suggest casting the temporal learning problem to a perceptron problem. In the discrete case a finite margin appears, providing the network, to some extent, robustness to noise, for which it performs perfectly (i.e. producing a desired sequence for an arbitrary number of cycles flawlessly). In the continuous case the margin approaches zero when the output unit changes its state, hence the network is only able to reproduce the sequence with slight jitters. Numerical simulation suggest that in the discrete time case, the longest sequence that can be learned scales, at best, as square root of the network size. A dramatic effect occurs when learning several short sequences in parallel, that is, their total length substantially exceeds the length of the longest single sequence the network can learn. This model easily generalizes to an arbitrary number of output units, which boost its performance. This effect is demonstrated by considering two practical examples for sequence learning. This work suggests a way to overcome stability problems for training recurrent networks and further quantifies the performance of a network under the specific learning scheme. Public Library of Science 2017-12-12 /pmc/articles/PMC5741269/ /pubmed/29232710 http://dx.doi.org/10.1371/journal.pcbi.1005861 Text en © 2017 Ben-Shushan, Tsodyks http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Ben-Shushan, Nadav
Tsodyks, Misha
Stabilizing patterns in time: Neural network approach
title Stabilizing patterns in time: Neural network approach
title_full Stabilizing patterns in time: Neural network approach
title_fullStr Stabilizing patterns in time: Neural network approach
title_full_unstemmed Stabilizing patterns in time: Neural network approach
title_short Stabilizing patterns in time: Neural network approach
title_sort stabilizing patterns in time: neural network approach
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5741269/
https://www.ncbi.nlm.nih.gov/pubmed/29232710
http://dx.doi.org/10.1371/journal.pcbi.1005861
work_keys_str_mv AT benshushannadav stabilizingpatternsintimeneuralnetworkapproach
AT tsodyksmisha stabilizingpatternsintimeneuralnetworkapproach