Cargando…
Learning, Memory, and the Role of Neural Network Architecture
The performance of information processing systems, from artificial neural networks to natural neuronal ensembles, depends heavily on the underlying system architecture. In this study, we compare the performance of parallel and layered network architectures during sequential tasks that require both a...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2011
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3127797/ https://www.ncbi.nlm.nih.gov/pubmed/21738455 http://dx.doi.org/10.1371/journal.pcbi.1002063 |
_version_ | 1782207374127267840 |
---|---|
author | Hermundstad, Ann M. Brown, Kevin S. Bassett, Danielle S. Carlson, Jean M. |
author_facet | Hermundstad, Ann M. Brown, Kevin S. Bassett, Danielle S. Carlson, Jean M. |
author_sort | Hermundstad, Ann M. |
collection | PubMed |
description | The performance of information processing systems, from artificial neural networks to natural neuronal ensembles, depends heavily on the underlying system architecture. In this study, we compare the performance of parallel and layered network architectures during sequential tasks that require both acquisition and retention of information, thereby identifying tradeoffs between learning and memory processes. During the task of supervised, sequential function approximation, networks produce and adapt representations of external information. Performance is evaluated by statistically analyzing the error in these representations while varying the initial network state, the structure of the external information, and the time given to learn the information. We link performance to complexity in network architecture by characterizing local error landscape curvature. We find that variations in error landscape structure give rise to tradeoffs in performance; these include the ability of the network to maximize accuracy versus minimize inaccuracy and produce specific versus generalizable representations of information. Parallel networks generate smooth error landscapes with deep, narrow minima, enabling them to find highly specific representations given sufficient time. While accurate, however, these representations are difficult to generalize. In contrast, layered networks generate rough error landscapes with a variety of local minima, allowing them to quickly find coarse representations. Although less accurate, these representations are easily adaptable. The presence of measurable performance tradeoffs in both layered and parallel networks has implications for understanding the behavior of a wide variety of natural and artificial learning systems. |
format | Online Article Text |
id | pubmed-3127797 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2011 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-31277972011-07-07 Learning, Memory, and the Role of Neural Network Architecture Hermundstad, Ann M. Brown, Kevin S. Bassett, Danielle S. Carlson, Jean M. PLoS Comput Biol Research Article The performance of information processing systems, from artificial neural networks to natural neuronal ensembles, depends heavily on the underlying system architecture. In this study, we compare the performance of parallel and layered network architectures during sequential tasks that require both acquisition and retention of information, thereby identifying tradeoffs between learning and memory processes. During the task of supervised, sequential function approximation, networks produce and adapt representations of external information. Performance is evaluated by statistically analyzing the error in these representations while varying the initial network state, the structure of the external information, and the time given to learn the information. We link performance to complexity in network architecture by characterizing local error landscape curvature. We find that variations in error landscape structure give rise to tradeoffs in performance; these include the ability of the network to maximize accuracy versus minimize inaccuracy and produce specific versus generalizable representations of information. Parallel networks generate smooth error landscapes with deep, narrow minima, enabling them to find highly specific representations given sufficient time. While accurate, however, these representations are difficult to generalize. In contrast, layered networks generate rough error landscapes with a variety of local minima, allowing them to quickly find coarse representations. Although less accurate, these representations are easily adaptable. The presence of measurable performance tradeoffs in both layered and parallel networks has implications for understanding the behavior of a wide variety of natural and artificial learning systems. Public Library of Science 2011-06-30 /pmc/articles/PMC3127797/ /pubmed/21738455 http://dx.doi.org/10.1371/journal.pcbi.1002063 Text en Hermundstad et al. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited. |
spellingShingle | Research Article Hermundstad, Ann M. Brown, Kevin S. Bassett, Danielle S. Carlson, Jean M. Learning, Memory, and the Role of Neural Network Architecture |
title | Learning, Memory, and the Role of Neural Network Architecture |
title_full | Learning, Memory, and the Role of Neural Network Architecture |
title_fullStr | Learning, Memory, and the Role of Neural Network Architecture |
title_full_unstemmed | Learning, Memory, and the Role of Neural Network Architecture |
title_short | Learning, Memory, and the Role of Neural Network Architecture |
title_sort | learning, memory, and the role of neural network architecture |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3127797/ https://www.ncbi.nlm.nih.gov/pubmed/21738455 http://dx.doi.org/10.1371/journal.pcbi.1002063 |
work_keys_str_mv | AT hermundstadannm learningmemoryandtheroleofneuralnetworkarchitecture AT brownkevins learningmemoryandtheroleofneuralnetworkarchitecture AT bassettdanielles learningmemoryandtheroleofneuralnetworkarchitecture AT carlsonjeanm learningmemoryandtheroleofneuralnetworkarchitecture |