Cargando…

A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning

Synaptic connections between neurons in the brain are dynamic because of continuously ongoing spine dynamics, axonal sprouting, and other processes. In fact, it was recently shown that the spontaneous synapse-autonomous component of spine dynamics is at least as large as the component that depends o...

Descripción completa

Detalles Bibliográficos
Autores principales: Kappel, David, Legenstein, Robert, Habenschuss, Stefan, Hsieh, Michael, Maass, Wolfgang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Society for Neuroscience 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5913731/
https://www.ncbi.nlm.nih.gov/pubmed/29696150
http://dx.doi.org/10.1523/ENEURO.0301-17.2018
_version_ 1783316595653738496
author Kappel, David
Legenstein, Robert
Habenschuss, Stefan
Hsieh, Michael
Maass, Wolfgang
author_facet Kappel, David
Legenstein, Robert
Habenschuss, Stefan
Hsieh, Michael
Maass, Wolfgang
author_sort Kappel, David
collection PubMed
description Synaptic connections between neurons in the brain are dynamic because of continuously ongoing spine dynamics, axonal sprouting, and other processes. In fact, it was recently shown that the spontaneous synapse-autonomous component of spine dynamics is at least as large as the component that depends on the history of pre- and postsynaptic neural activity. These data are inconsistent with common models for network plasticity and raise the following questions: how can neural circuits maintain a stable computational function in spite of these continuously ongoing processes, and what could be functional uses of these ongoing processes? Here, we present a rigorous theoretical framework for these seemingly stochastic spine dynamics and rewiring processes in the context of reward-based learning tasks. We show that spontaneous synapse-autonomous processes, in combination with reward signals such as dopamine, can explain the capability of networks of neurons in the brain to configure themselves for specific computational tasks, and to compensate automatically for later changes in the network or task. Furthermore, we show theoretically and through computer simulations that stable computational performance is compatible with continuously ongoing synapse-autonomous changes. After reaching good computational performance it causes primarily a slow drift of network architecture and dynamics in task-irrelevant dimensions, as observed for neural activity in motor cortex and other areas. On the more abstract level of reinforcement learning the resulting model gives rise to an understanding of reward-driven network plasticity as continuous sampling of network configurations.
format Online
Article
Text
id pubmed-5913731
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher Society for Neuroscience
record_format MEDLINE/PubMed
spelling pubmed-59137312018-04-25 A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning Kappel, David Legenstein, Robert Habenschuss, Stefan Hsieh, Michael Maass, Wolfgang eNeuro Theory/New Concepts Synaptic connections between neurons in the brain are dynamic because of continuously ongoing spine dynamics, axonal sprouting, and other processes. In fact, it was recently shown that the spontaneous synapse-autonomous component of spine dynamics is at least as large as the component that depends on the history of pre- and postsynaptic neural activity. These data are inconsistent with common models for network plasticity and raise the following questions: how can neural circuits maintain a stable computational function in spite of these continuously ongoing processes, and what could be functional uses of these ongoing processes? Here, we present a rigorous theoretical framework for these seemingly stochastic spine dynamics and rewiring processes in the context of reward-based learning tasks. We show that spontaneous synapse-autonomous processes, in combination with reward signals such as dopamine, can explain the capability of networks of neurons in the brain to configure themselves for specific computational tasks, and to compensate automatically for later changes in the network or task. Furthermore, we show theoretically and through computer simulations that stable computational performance is compatible with continuously ongoing synapse-autonomous changes. After reaching good computational performance it causes primarily a slow drift of network architecture and dynamics in task-irrelevant dimensions, as observed for neural activity in motor cortex and other areas. On the more abstract level of reinforcement learning the resulting model gives rise to an understanding of reward-driven network plasticity as continuous sampling of network configurations. Society for Neuroscience 2018-04-24 /pmc/articles/PMC5913731/ /pubmed/29696150 http://dx.doi.org/10.1523/ENEURO.0301-17.2018 Text en Copyright © 2018 Kappel et al. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.
spellingShingle Theory/New Concepts
Kappel, David
Legenstein, Robert
Habenschuss, Stefan
Hsieh, Michael
Maass, Wolfgang
A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning
title A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning
title_full A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning
title_fullStr A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning
title_full_unstemmed A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning
title_short A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning
title_sort dynamic connectome supports the emergence of stable computational function of neural circuits through reward-based learning
topic Theory/New Concepts
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5913731/
https://www.ncbi.nlm.nih.gov/pubmed/29696150
http://dx.doi.org/10.1523/ENEURO.0301-17.2018
work_keys_str_mv AT kappeldavid adynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT legensteinrobert adynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT habenschussstefan adynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT hsiehmichael adynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT maasswolfgang adynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT kappeldavid dynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT legensteinrobert dynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT habenschussstefan dynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT hsiehmichael dynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning
AT maasswolfgang dynamicconnectomesupportstheemergenceofstablecomputationalfunctionofneuralcircuitsthroughrewardbasedlearning