Cargando…

Computational mechanisms underlying motivation to earn symbolic reinforcers

Reinforcement learning (RL) is a theoretical framework that describes how agents learn to select options that maximize rewards and minimize punishments over time. We often make choices, however, to obtain symbolic reinforcers (e.g. money, points) that can later be exchanged for primary reinforcers (...

Descripción completa

Detalles Bibliográficos
Autores principales: Burk, Diana C., Taswell, Craig, Tang, Hua, Averbeck, Bruno B.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Cold Spring Harbor Laboratory 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10592730/
https://www.ncbi.nlm.nih.gov/pubmed/37873311
http://dx.doi.org/10.1101/2023.10.11.561900
_version_ 1785124334907424768
author Burk, Diana C.
Taswell, Craig
Tang, Hua
Averbeck, Bruno B.
author_facet Burk, Diana C.
Taswell, Craig
Tang, Hua
Averbeck, Bruno B.
author_sort Burk, Diana C.
collection PubMed
description Reinforcement learning (RL) is a theoretical framework that describes how agents learn to select options that maximize rewards and minimize punishments over time. We often make choices, however, to obtain symbolic reinforcers (e.g. money, points) that can later be exchanged for primary reinforcers (e.g. food, drink). Although symbolic reinforcers are motivating, little is understood about the neural or computational mechanisms underlying the motivation to earn them. In the present study, we examined how monkeys learn to make choices that maximize fluid rewards through reinforcement with tokens. The question addressed here is how the value of a state, which is a function of multiple task features (e.g. current number of accumulated tokens, choice options, task epoch, trials since last delivery of primary reinforcer, etc.), drives value and affects motivation. We constructed a Markov decision process model that computes the value of task states given task features to capture the motivational state of the animal. Fixation times, choice reaction times, and abort frequency were all significantly related to values of task states during the tokens task (n=5 monkeys). Furthermore, the model makes predictions for how neural responses could change on a moment-by-moment basis relative to changes in state value. Together, this task and model allow us to capture learning and behavior related to symbolic reinforcement.
format Online
Article
Text
id pubmed-10592730
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Cold Spring Harbor Laboratory
record_format MEDLINE/PubMed
spelling pubmed-105927302023-10-24 Computational mechanisms underlying motivation to earn symbolic reinforcers Burk, Diana C. Taswell, Craig Tang, Hua Averbeck, Bruno B. bioRxiv Article Reinforcement learning (RL) is a theoretical framework that describes how agents learn to select options that maximize rewards and minimize punishments over time. We often make choices, however, to obtain symbolic reinforcers (e.g. money, points) that can later be exchanged for primary reinforcers (e.g. food, drink). Although symbolic reinforcers are motivating, little is understood about the neural or computational mechanisms underlying the motivation to earn them. In the present study, we examined how monkeys learn to make choices that maximize fluid rewards through reinforcement with tokens. The question addressed here is how the value of a state, which is a function of multiple task features (e.g. current number of accumulated tokens, choice options, task epoch, trials since last delivery of primary reinforcer, etc.), drives value and affects motivation. We constructed a Markov decision process model that computes the value of task states given task features to capture the motivational state of the animal. Fixation times, choice reaction times, and abort frequency were all significantly related to values of task states during the tokens task (n=5 monkeys). Furthermore, the model makes predictions for how neural responses could change on a moment-by-moment basis relative to changes in state value. Together, this task and model allow us to capture learning and behavior related to symbolic reinforcement. Cold Spring Harbor Laboratory 2023-10-11 /pmc/articles/PMC10592730/ /pubmed/37873311 http://dx.doi.org/10.1101/2023.10.11.561900 Text en https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (https://creativecommons.org/licenses/by-nc-nd/4.0/) , which allows reusers to copy and distribute the material in any medium or format in unadapted form only, for noncommercial purposes only, and only so long as attribution is given to the creator.
spellingShingle Article
Burk, Diana C.
Taswell, Craig
Tang, Hua
Averbeck, Bruno B.
Computational mechanisms underlying motivation to earn symbolic reinforcers
title Computational mechanisms underlying motivation to earn symbolic reinforcers
title_full Computational mechanisms underlying motivation to earn symbolic reinforcers
title_fullStr Computational mechanisms underlying motivation to earn symbolic reinforcers
title_full_unstemmed Computational mechanisms underlying motivation to earn symbolic reinforcers
title_short Computational mechanisms underlying motivation to earn symbolic reinforcers
title_sort computational mechanisms underlying motivation to earn symbolic reinforcers
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10592730/
https://www.ncbi.nlm.nih.gov/pubmed/37873311
http://dx.doi.org/10.1101/2023.10.11.561900
work_keys_str_mv AT burkdianac computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers
AT taswellcraig computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers
AT tanghua computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers
AT averbeckbrunob computationalmechanismsunderlyingmotivationtoearnsymbolicreinforcers