Cargando…

Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice

The exploration/exploitation tradeoff – pursuing a known reward vs. sampling from lesser known options in the hope of finding a better payoff – is a fundamental aspect of learning and decision making. In humans, this has been studied using multi-armed bandit tasks. The same processes have also been...

Descripción completa

Detalles Bibliográficos
Autores principales: Metha, Jeremy A., Brian, Maddison L., Oberrauch, Sara, Barnes, Samuel A., Featherby, Travis J., Bossaerts, Peter, Murawski, Carsten, Hoyer, Daniel, Jacobson, Laura H.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6962304/
https://www.ncbi.nlm.nih.gov/pubmed/31998088
http://dx.doi.org/10.3389/fnbeh.2019.00270
_version_ 1783488135878934528
author Metha, Jeremy A.
Brian, Maddison L.
Oberrauch, Sara
Barnes, Samuel A.
Featherby, Travis J.
Bossaerts, Peter
Murawski, Carsten
Hoyer, Daniel
Jacobson, Laura H.
author_facet Metha, Jeremy A.
Brian, Maddison L.
Oberrauch, Sara
Barnes, Samuel A.
Featherby, Travis J.
Bossaerts, Peter
Murawski, Carsten
Hoyer, Daniel
Jacobson, Laura H.
author_sort Metha, Jeremy A.
collection PubMed
description The exploration/exploitation tradeoff – pursuing a known reward vs. sampling from lesser known options in the hope of finding a better payoff – is a fundamental aspect of learning and decision making. In humans, this has been studied using multi-armed bandit tasks. The same processes have also been studied using simplified probabilistic reversal learning (PRL) tasks with binary choices. Our investigations suggest that protocols previously used to explore PRL in mice may prove beyond their cognitive capacities, with animals performing at a no-better-than-chance level. We sought a novel probabilistic learning task to improve behavioral responding in mice, whilst allowing the investigation of the exploration/exploitation tradeoff in decision making. To achieve this, we developed a two-lever operant chamber task with levers corresponding to different probabilities (high/low) of receiving a saccharin reward, reversing the reward contingencies associated with levers once animals reached a threshold of 80% responding at the high rewarding lever. We found that, unlike in existing PRL tasks, mice are able to learn and behave near optimally with 80% high/20% low reward probabilities. Altering the reward contingencies towards equality showed that some mice displayed preference for the high rewarding lever with probabilities as close as 60% high/40% low. Additionally, we show that animal choice behavior can be effectively modelled using reinforcement learning (RL) models incorporating learning rates for positive and negative prediction error, a perseveration parameter, and a noise parameter. This new decision task, coupled with RL analyses, advances access to investigate the neuroscience of the exploration/exploitation tradeoff in decision making.
format Online
Article
Text
id pubmed-6962304
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-69623042020-01-29 Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice Metha, Jeremy A. Brian, Maddison L. Oberrauch, Sara Barnes, Samuel A. Featherby, Travis J. Bossaerts, Peter Murawski, Carsten Hoyer, Daniel Jacobson, Laura H. Front Behav Neurosci Neuroscience The exploration/exploitation tradeoff – pursuing a known reward vs. sampling from lesser known options in the hope of finding a better payoff – is a fundamental aspect of learning and decision making. In humans, this has been studied using multi-armed bandit tasks. The same processes have also been studied using simplified probabilistic reversal learning (PRL) tasks with binary choices. Our investigations suggest that protocols previously used to explore PRL in mice may prove beyond their cognitive capacities, with animals performing at a no-better-than-chance level. We sought a novel probabilistic learning task to improve behavioral responding in mice, whilst allowing the investigation of the exploration/exploitation tradeoff in decision making. To achieve this, we developed a two-lever operant chamber task with levers corresponding to different probabilities (high/low) of receiving a saccharin reward, reversing the reward contingencies associated with levers once animals reached a threshold of 80% responding at the high rewarding lever. We found that, unlike in existing PRL tasks, mice are able to learn and behave near optimally with 80% high/20% low reward probabilities. Altering the reward contingencies towards equality showed that some mice displayed preference for the high rewarding lever with probabilities as close as 60% high/40% low. Additionally, we show that animal choice behavior can be effectively modelled using reinforcement learning (RL) models incorporating learning rates for positive and negative prediction error, a perseveration parameter, and a noise parameter. This new decision task, coupled with RL analyses, advances access to investigate the neuroscience of the exploration/exploitation tradeoff in decision making. Frontiers Media S.A. 2020-01-09 /pmc/articles/PMC6962304/ /pubmed/31998088 http://dx.doi.org/10.3389/fnbeh.2019.00270 Text en Copyright © 2020 Metha, Brian, Oberrauch, Barnes, Featherby, Bossaerts, Murawski, Hoyer and Jacobson. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Metha, Jeremy A.
Brian, Maddison L.
Oberrauch, Sara
Barnes, Samuel A.
Featherby, Travis J.
Bossaerts, Peter
Murawski, Carsten
Hoyer, Daniel
Jacobson, Laura H.
Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice
title Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice
title_full Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice
title_fullStr Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice
title_full_unstemmed Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice
title_short Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice
title_sort separating probability and reversal learning in a novel probabilistic reversal learning task for mice
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6962304/
https://www.ncbi.nlm.nih.gov/pubmed/31998088
http://dx.doi.org/10.3389/fnbeh.2019.00270
work_keys_str_mv AT methajeremya separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT brianmaddisonl separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT oberrauchsara separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT barnessamuela separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT featherbytravisj separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT bossaertspeter separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT murawskicarsten separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT hoyerdaniel separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice
AT jacobsonlaurah separatingprobabilityandreversallearninginanovelprobabilisticreversallearningtaskformice