Cargando…
Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice
The exploration/exploitation tradeoff – pursuing a known reward vs. sampling from lesser known options in the hope of finding a better payoff – is a fundamental aspect of learning and decision making. In humans, this has been studied using multi-armed bandit tasks. The same processes have also been...
Autores principales: | , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6962304/ https://www.ncbi.nlm.nih.gov/pubmed/31998088 http://dx.doi.org/10.3389/fnbeh.2019.00270 |
_version_ | 1783488135878934528 |
---|---|
author | Metha, Jeremy A. Brian, Maddison L. Oberrauch, Sara Barnes, Samuel A. Featherby, Travis J. Bossaerts, Peter Murawski, Carsten Hoyer, Daniel Jacobson, Laura H. |
author_facet | Metha, Jeremy A. Brian, Maddison L. Oberrauch, Sara Barnes, Samuel A. Featherby, Travis J. Bossaerts, Peter Murawski, Carsten Hoyer, Daniel Jacobson, Laura H. |
author_sort | Metha, Jeremy A. |
collection | PubMed |
description | The exploration/exploitation tradeoff – pursuing a known reward vs. sampling from lesser known options in the hope of finding a better payoff – is a fundamental aspect of learning and decision making. In humans, this has been studied using multi-armed bandit tasks. The same processes have also been studied using simplified probabilistic reversal learning (PRL) tasks with binary choices. Our investigations suggest that protocols previously used to explore PRL in mice may prove beyond their cognitive capacities, with animals performing at a no-better-than-chance level. We sought a novel probabilistic learning task to improve behavioral responding in mice, whilst allowing the investigation of the exploration/exploitation tradeoff in decision making. To achieve this, we developed a two-lever operant chamber task with levers corresponding to different probabilities (high/low) of receiving a saccharin reward, reversing the reward contingencies associated with levers once animals reached a threshold of 80% responding at the high rewarding lever. We found that, unlike in existing PRL tasks, mice are able to learn and behave near optimally with 80% high/20% low reward probabilities. Altering the reward contingencies towards equality showed that some mice displayed preference for the high rewarding lever with probabilities as close as 60% high/40% low. Additionally, we show that animal choice behavior can be effectively modelled using reinforcement learning (RL) models incorporating learning rates for positive and negative prediction error, a perseveration parameter, and a noise parameter. This new decision task, coupled with RL analyses, advances access to investigate the neuroscience of the exploration/exploitation tradeoff in decision making. |
format | Online Article Text |
id | pubmed-6962304 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-69623042020-01-29 Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice Metha, Jeremy A. Brian, Maddison L. Oberrauch, Sara Barnes, Samuel A. Featherby, Travis J. Bossaerts, Peter Murawski, Carsten Hoyer, Daniel Jacobson, Laura H. Front Behav Neurosci Neuroscience The exploration/exploitation tradeoff – pursuing a known reward vs. sampling from lesser known options in the hope of finding a better payoff – is a fundamental aspect of learning and decision making. In humans, this has been studied using multi-armed bandit tasks. The same processes have also been studied using simplified probabilistic reversal learning (PRL) tasks with binary choices. Our investigations suggest that protocols previously used to explore PRL in mice may prove beyond their cognitive capacities, with animals performing at a no-better-than-chance level. We sought a novel probabilistic learning task to improve behavioral responding in mice, whilst allowing the investigation of the exploration/exploitation tradeoff in decision making. To achieve this, we developed a two-lever operant chamber task with levers corresponding to different probabilities (high/low) of receiving a saccharin reward, reversing the reward contingencies associated with levers once animals reached a threshold of 80% responding at the high rewarding lever. We found that, unlike in existing PRL tasks, mice are able to learn and behave near optimally with 80% high/20% low reward probabilities. Altering the reward contingencies towards equality showed that some mice displayed preference for the high rewarding lever with probabilities as close as 60% high/40% low. Additionally, we show that animal choice behavior can be effectively modelled using reinforcement learning (RL) models incorporating learning rates for positive and negative prediction error, a perseveration parameter, and a noise parameter. This new decision task, coupled with RL analyses, advances access to investigate the neuroscience of the exploration/exploitation tradeoff in decision making. Frontiers Media S.A. 2020-01-09 /pmc/articles/PMC6962304/ /pubmed/31998088 http://dx.doi.org/10.3389/fnbeh.2019.00270 Text en Copyright © 2020 Metha, Brian, Oberrauch, Barnes, Featherby, Bossaerts, Murawski, Hoyer and Jacobson. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Metha, Jeremy A. Brian, Maddison L. Oberrauch, Sara Barnes, Samuel A. Featherby, Travis J. Bossaerts, Peter Murawski, Carsten Hoyer, Daniel Jacobson, Laura H. Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice |
title | Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice |
title_full | Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice |
title_fullStr | Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice |
title_full_unstemmed | Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice |
title_short | Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice |
title_sort | separating probability and reversal learning in a novel probabilistic reversal learning task for mice |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6962304/ https://www.ncbi.nlm.nih.gov/pubmed/31998088 http://dx.doi.org/10.3389/fnbeh.2019.00270 |
work_keys_str_mv | AT methajeremya separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT brianmaddisonl separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT oberrauchsara separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT barnessamuela separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT featherbytravisj separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT bossaertspeter separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT murawskicarsten separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT hoyerdaniel separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice AT jacobsonlaurah separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice |