Cargando…

Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning

Biologically plausible modeling of behavioral reinforcement learning tasks has seen great improvements over the past decades. Less work has been dedicated to tasks involving contingency reversals, i.e., tasks in which the original behavioral goal is reversed one or multiple times. The ability to adj...

Descripción completa

Detalles Bibliográficos
Autores principales: Jarvers, Christian, Brosch, Tobias, Brechmann, André, Woldeit, Marie L., Schulz, Andreas L., Ohl, Frank W., Lommerzheim, Marcel, Neumann, Heiko
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5112252/
https://www.ncbi.nlm.nih.gov/pubmed/27909395
http://dx.doi.org/10.3389/fnins.2016.00535
_version_ 1782467957772779520
author Jarvers, Christian
Brosch, Tobias
Brechmann, André
Woldeit, Marie L.
Schulz, Andreas L.
Ohl, Frank W.
Lommerzheim, Marcel
Neumann, Heiko
author_facet Jarvers, Christian
Brosch, Tobias
Brechmann, André
Woldeit, Marie L.
Schulz, Andreas L.
Ohl, Frank W.
Lommerzheim, Marcel
Neumann, Heiko
author_sort Jarvers, Christian
collection PubMed
description Biologically plausible modeling of behavioral reinforcement learning tasks has seen great improvements over the past decades. Less work has been dedicated to tasks involving contingency reversals, i.e., tasks in which the original behavioral goal is reversed one or multiple times. The ability to adjust to such reversals is a key element of behavioral flexibility. Here, we investigate the neural mechanisms underlying contingency-reversal tasks. We first conduct experiments with humans and gerbils to demonstrate memory effects, including multiple reversals in which subjects (humans and animals) show a faster learning rate when a previously learned contingency re-appears. Motivated by recurrent mechanisms of learning and memory for object categories, we propose a network architecture which involves reinforcement learning to steer an orienting system that monitors the success in reward acquisition. We suggest that a model sensory system provides feature representations which are further processed by category-related subnetworks which constitute a neural analog of expert networks. Categories are selected dynamically in a competitive field and predict the expected reward. Learning occurs in sequentialized phases to selectively focus the weight adaptation to synapses in the hierarchical network and modulate their weight changes by a global modulator signal. The orienting subsystem itself learns to bias the competition in the presence of continuous monotonic reward accumulation. In case of sudden changes in the discrepancy of predicted and acquired reward the activated motor category can be switched. We suggest that this subsystem is composed of a hierarchically organized network of dis-inhibitory mechanisms, dubbed a dynamic control network (DCN), which resembles components of the basal ganglia. The DCN selectively activates an expert network, corresponding to the current behavioral strategy. The trace of the accumulated reward is monitored such that large sudden deviations from the monotonicity of its evolution trigger a reset after which another expert subnetwork can be activated—if it has already been established before—or new categories can be recruited and associated with novel behavioral patterns.
format Online
Article
Text
id pubmed-5112252
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-51122522016-12-01 Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning Jarvers, Christian Brosch, Tobias Brechmann, André Woldeit, Marie L. Schulz, Andreas L. Ohl, Frank W. Lommerzheim, Marcel Neumann, Heiko Front Neurosci Neuroscience Biologically plausible modeling of behavioral reinforcement learning tasks has seen great improvements over the past decades. Less work has been dedicated to tasks involving contingency reversals, i.e., tasks in which the original behavioral goal is reversed one or multiple times. The ability to adjust to such reversals is a key element of behavioral flexibility. Here, we investigate the neural mechanisms underlying contingency-reversal tasks. We first conduct experiments with humans and gerbils to demonstrate memory effects, including multiple reversals in which subjects (humans and animals) show a faster learning rate when a previously learned contingency re-appears. Motivated by recurrent mechanisms of learning and memory for object categories, we propose a network architecture which involves reinforcement learning to steer an orienting system that monitors the success in reward acquisition. We suggest that a model sensory system provides feature representations which are further processed by category-related subnetworks which constitute a neural analog of expert networks. Categories are selected dynamically in a competitive field and predict the expected reward. Learning occurs in sequentialized phases to selectively focus the weight adaptation to synapses in the hierarchical network and modulate their weight changes by a global modulator signal. The orienting subsystem itself learns to bias the competition in the presence of continuous monotonic reward accumulation. In case of sudden changes in the discrepancy of predicted and acquired reward the activated motor category can be switched. We suggest that this subsystem is composed of a hierarchically organized network of dis-inhibitory mechanisms, dubbed a dynamic control network (DCN), which resembles components of the basal ganglia. The DCN selectively activates an expert network, corresponding to the current behavioral strategy. The trace of the accumulated reward is monitored such that large sudden deviations from the monotonicity of its evolution trigger a reset after which another expert subnetwork can be activated—if it has already been established before—or new categories can be recruited and associated with novel behavioral patterns. Frontiers Media S.A. 2016-11-17 /pmc/articles/PMC5112252/ /pubmed/27909395 http://dx.doi.org/10.3389/fnins.2016.00535 Text en Copyright © 2016 Jarvers, Brosch, Brechmann, Woldeit, Schulz, Ohl, Lommerzheim and Neumann. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Jarvers, Christian
Brosch, Tobias
Brechmann, André
Woldeit, Marie L.
Schulz, Andreas L.
Ohl, Frank W.
Lommerzheim, Marcel
Neumann, Heiko
Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning
title Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning
title_full Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning
title_fullStr Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning
title_full_unstemmed Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning
title_short Reversal Learning in Humans and Gerbils: Dynamic Control Network Facilitates Learning
title_sort reversal learning in humans and gerbils: dynamic control network facilitates learning
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5112252/
https://www.ncbi.nlm.nih.gov/pubmed/27909395
http://dx.doi.org/10.3389/fnins.2016.00535
work_keys_str_mv AT jarverschristian reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning
AT broschtobias reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning
AT brechmannandre reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning
AT woldeitmariel reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning
AT schulzandreasl reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning
AT ohlfrankw reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning
AT lommerzheimmarcel reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning
AT neumannheiko reversallearninginhumansandgerbilsdynamiccontrolnetworkfacilitateslearning