Cargando…

Reinforcement Learning With Parsimonious Computation and a Forgetting Process

Decision-making is assumed to be supported by model-free and model-based systems: the model-free system is based purely on experience, while the model-based system uses a cognitive map of the environment and is more accurate. The recently developed multistep decision-making task and its computationa...

Descripción completa

Detalles Bibliográficos
Autores principales: Toyama, Asako, Katahira, Kentaro, Ohira, Hideki
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6520826/
https://www.ncbi.nlm.nih.gov/pubmed/31143107
http://dx.doi.org/10.3389/fnhum.2019.00153
_version_ 1783418818353168384
author Toyama, Asako
Katahira, Kentaro
Ohira, Hideki
author_facet Toyama, Asako
Katahira, Kentaro
Ohira, Hideki
author_sort Toyama, Asako
collection PubMed
description Decision-making is assumed to be supported by model-free and model-based systems: the model-free system is based purely on experience, while the model-based system uses a cognitive map of the environment and is more accurate. The recently developed multistep decision-making task and its computational model can dissociate the contributions of the two systems and have been used widely. This study used this task and model to understand our value-based learning process and tested alternative algorithms for the model-free and model-based learning systems. The task used in this study had a deterministic transition structure, and the degree of use of this structure in learning is estimated as the relative contribution of the model-based system to choices. We obtained data from 29 participants and fitted them with various computational models that differ in the model-free and model-based assumptions. The results of model comparison and parameter estimation showed that the participants update the value of action sequences and not each action. Additionally, the model fit was improved substantially by assuming that the learning mechanism includes a forgetting process, where the values of unselected options change to a certain default value over time. We also examined the relationships between the estimated parameters and psychopathology and other traits measured by self-reported questionnaires, and the results suggested that the difference in model assumptions can change the conclusion. In particular, inclusion of the forgetting process in the computational models had a strong impact on estimation of the weighting parameter of the model-free and model-based systems.
format Online
Article
Text
id pubmed-6520826
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-65208262019-05-29 Reinforcement Learning With Parsimonious Computation and a Forgetting Process Toyama, Asako Katahira, Kentaro Ohira, Hideki Front Hum Neurosci Neuroscience Decision-making is assumed to be supported by model-free and model-based systems: the model-free system is based purely on experience, while the model-based system uses a cognitive map of the environment and is more accurate. The recently developed multistep decision-making task and its computational model can dissociate the contributions of the two systems and have been used widely. This study used this task and model to understand our value-based learning process and tested alternative algorithms for the model-free and model-based learning systems. The task used in this study had a deterministic transition structure, and the degree of use of this structure in learning is estimated as the relative contribution of the model-based system to choices. We obtained data from 29 participants and fitted them with various computational models that differ in the model-free and model-based assumptions. The results of model comparison and parameter estimation showed that the participants update the value of action sequences and not each action. Additionally, the model fit was improved substantially by assuming that the learning mechanism includes a forgetting process, where the values of unselected options change to a certain default value over time. We also examined the relationships between the estimated parameters and psychopathology and other traits measured by self-reported questionnaires, and the results suggested that the difference in model assumptions can change the conclusion. In particular, inclusion of the forgetting process in the computational models had a strong impact on estimation of the weighting parameter of the model-free and model-based systems. Frontiers Media S.A. 2019-05-09 /pmc/articles/PMC6520826/ /pubmed/31143107 http://dx.doi.org/10.3389/fnhum.2019.00153 Text en Copyright © 2019 Toyama, Katahira and Ohira. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Toyama, Asako
Katahira, Kentaro
Ohira, Hideki
Reinforcement Learning With Parsimonious Computation and a Forgetting Process
title Reinforcement Learning With Parsimonious Computation and a Forgetting Process
title_full Reinforcement Learning With Parsimonious Computation and a Forgetting Process
title_fullStr Reinforcement Learning With Parsimonious Computation and a Forgetting Process
title_full_unstemmed Reinforcement Learning With Parsimonious Computation and a Forgetting Process
title_short Reinforcement Learning With Parsimonious Computation and a Forgetting Process
title_sort reinforcement learning with parsimonious computation and a forgetting process
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6520826/
https://www.ncbi.nlm.nih.gov/pubmed/31143107
http://dx.doi.org/10.3389/fnhum.2019.00153
work_keys_str_mv AT toyamaasako reinforcementlearningwithparsimoniouscomputationandaforgettingprocess
AT katahirakentaro reinforcementlearningwithparsimoniouscomputationandaforgettingprocess
AT ohirahideki reinforcementlearningwithparsimoniouscomputationandaforgettingprocess