Cargando…
Computational mechanisms underlying motivation to earn symbolic reinforcers
Reinforcement learning (RL) is a theoretical framework that describes how agents learn to select options that maximize rewards and minimize punishments over time. We often make choices, however, to obtain symbolic reinforcers (e.g. money, points) that can later be exchanged for primary reinforcers (...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Cold Spring Harbor Laboratory
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10592730/ https://www.ncbi.nlm.nih.gov/pubmed/37873311 http://dx.doi.org/10.1101/2023.10.11.561900 |
_version_ | 1785124334907424768 |
---|---|
author | Burk, Diana C. Taswell, Craig Tang, Hua Averbeck, Bruno B. |
author_facet | Burk, Diana C. Taswell, Craig Tang, Hua Averbeck, Bruno B. |
author_sort | Burk, Diana C. |
collection | PubMed |
description | Reinforcement learning (RL) is a theoretical framework that describes how agents learn to select options that maximize rewards and minimize punishments over time. We often make choices, however, to obtain symbolic reinforcers (e.g. money, points) that can later be exchanged for primary reinforcers (e.g. food, drink). Although symbolic reinforcers are motivating, little is understood about the neural or computational mechanisms underlying the motivation to earn them. In the present study, we examined how monkeys learn to make choices that maximize fluid rewards through reinforcement with tokens. The question addressed here is how the value of a state, which is a function of multiple task features (e.g. current number of accumulated tokens, choice options, task epoch, trials since last delivery of primary reinforcer, etc.), drives value and affects motivation. We constructed a Markov decision process model that computes the value of task states given task features to capture the motivational state of the animal. Fixation times, choice reaction times, and abort frequency were all significantly related to values of task states during the tokens task (n=5 monkeys). Furthermore, the model makes predictions for how neural responses could change on a moment-by-moment basis relative to changes in state value. Together, this task and model allow us to capture learning and behavior related to symbolic reinforcement. |
format | Online Article Text |
id | pubmed-10592730 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Cold Spring Harbor Laboratory |
record_format | MEDLINE/PubMed |
spelling | pubmed-105927302023-10-24 Computational mechanisms underlying motivation to earn symbolic reinforcers Burk, Diana C. Taswell, Craig Tang, Hua Averbeck, Bruno B. bioRxiv Article Reinforcement learning (RL) is a theoretical framework that describes how agents learn to select options that maximize rewards and minimize punishments over time. We often make choices, however, to obtain symbolic reinforcers (e.g. money, points) that can later be exchanged for primary reinforcers (e.g. food, drink). Although symbolic reinforcers are motivating, little is understood about the neural or computational mechanisms underlying the motivation to earn them. In the present study, we examined how monkeys learn to make choices that maximize fluid rewards through reinforcement with tokens. The question addressed here is how the value of a state, which is a function of multiple task features (e.g. current number of accumulated tokens, choice options, task epoch, trials since last delivery of primary reinforcer, etc.), drives value and affects motivation. We constructed a Markov decision process model that computes the value of task states given task features to capture the motivational state of the animal. Fixation times, choice reaction times, and abort frequency were all significantly related to values of task states during the tokens task (n=5 monkeys). Furthermore, the model makes predictions for how neural responses could change on a moment-by-moment basis relative to changes in state value. Together, this task and model allow us to capture learning and behavior related to symbolic reinforcement. Cold Spring Harbor Laboratory 2023-10-11 /pmc/articles/PMC10592730/ /pubmed/37873311 http://dx.doi.org/10.1101/2023.10.11.561900 Text en https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (https://creativecommons.org/licenses/by-nc-nd/4.0/) , which allows reusers to copy and distribute the material in any medium or format in unadapted form only, for noncommercial purposes only, and only so long as attribution is given to the creator. |
spellingShingle | Article Burk, Diana C. Taswell, Craig Tang, Hua Averbeck, Bruno B. Computational mechanisms underlying motivation to earn symbolic reinforcers |
title | Computational mechanisms underlying motivation to earn symbolic reinforcers |
title_full | Computational mechanisms underlying motivation to earn symbolic reinforcers |
title_fullStr | Computational mechanisms underlying motivation to earn symbolic reinforcers |
title_full_unstemmed | Computational mechanisms underlying motivation to earn symbolic reinforcers |
title_short | Computational mechanisms underlying motivation to earn symbolic reinforcers |
title_sort | computational mechanisms underlying motivation to earn symbolic reinforcers |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10592730/ https://www.ncbi.nlm.nih.gov/pubmed/37873311 http://dx.doi.org/10.1101/2023.10.11.561900 |
work_keys_str_mv | AT burkdianac computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers AT taswellcraig computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers AT tanghua computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers AT averbeckbrunob computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers |