Cargando…

A new model of decision processing in instrumental learning tasks

Learning and decision-making are interactive processes, yet cognitive modeling of error-driven learning and decision-making have largely evolved separately. Recently, evidence accumulation models (EAMs) of decision-making and reinforcement learning (RL) models of error-driven learning have been comb...

Descripción completa

Detalles Bibliográficos
Autores principales: Miletić, Steven, Boag, Russell J, Trutti, Anne C, Stevenson, Niek, Forstmann, Birte U, Heathcote, Andrew
Formato: Online Artículo Texto
Lenguaje:English
Publicado: eLife Sciences Publications, Ltd 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7880686/
https://www.ncbi.nlm.nih.gov/pubmed/33501916
http://dx.doi.org/10.7554/eLife.63055
_version_ 1783650746966736896
author Miletić, Steven
Boag, Russell J
Trutti, Anne C
Stevenson, Niek
Forstmann, Birte U
Heathcote, Andrew
author_facet Miletić, Steven
Boag, Russell J
Trutti, Anne C
Stevenson, Niek
Forstmann, Birte U
Heathcote, Andrew
author_sort Miletić, Steven
collection PubMed
description Learning and decision-making are interactive processes, yet cognitive modeling of error-driven learning and decision-making have largely evolved separately. Recently, evidence accumulation models (EAMs) of decision-making and reinforcement learning (RL) models of error-driven learning have been combined into joint RL-EAMs that can in principle address these interactions. However, we show that the most commonly used combination, based on the diffusion decision model (DDM) for binary choice, consistently fails to capture crucial aspects of response times observed during reinforcement learning. We propose a new RL-EAM based on an advantage racing diffusion (ARD) framework for choices among two or more options that not only addresses this problem but captures stimulus difficulty, speed-accuracy trade-off, and stimulus-response-mapping reversal effects. The RL-ARD avoids fundamental limitations imposed by the DDM on addressing effects of absolute values of choices, as well as extensions beyond binary choice, and provides a computationally tractable basis for wider applications.
format Online
Article
Text
id pubmed-7880686
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher eLife Sciences Publications, Ltd
record_format MEDLINE/PubMed
spelling pubmed-78806862021-02-16 A new model of decision processing in instrumental learning tasks Miletić, Steven Boag, Russell J Trutti, Anne C Stevenson, Niek Forstmann, Birte U Heathcote, Andrew eLife Neuroscience Learning and decision-making are interactive processes, yet cognitive modeling of error-driven learning and decision-making have largely evolved separately. Recently, evidence accumulation models (EAMs) of decision-making and reinforcement learning (RL) models of error-driven learning have been combined into joint RL-EAMs that can in principle address these interactions. However, we show that the most commonly used combination, based on the diffusion decision model (DDM) for binary choice, consistently fails to capture crucial aspects of response times observed during reinforcement learning. We propose a new RL-EAM based on an advantage racing diffusion (ARD) framework for choices among two or more options that not only addresses this problem but captures stimulus difficulty, speed-accuracy trade-off, and stimulus-response-mapping reversal effects. The RL-ARD avoids fundamental limitations imposed by the DDM on addressing effects of absolute values of choices, as well as extensions beyond binary choice, and provides a computationally tractable basis for wider applications. eLife Sciences Publications, Ltd 2021-01-27 /pmc/articles/PMC7880686/ /pubmed/33501916 http://dx.doi.org/10.7554/eLife.63055 Text en © 2021, Miletić et al http://creativecommons.org/licenses/by/4.0/ http://creativecommons.org/licenses/by/4.0/This article is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited.
spellingShingle Neuroscience
Miletić, Steven
Boag, Russell J
Trutti, Anne C
Stevenson, Niek
Forstmann, Birte U
Heathcote, Andrew
A new model of decision processing in instrumental learning tasks
title A new model of decision processing in instrumental learning tasks
title_full A new model of decision processing in instrumental learning tasks
title_fullStr A new model of decision processing in instrumental learning tasks
title_full_unstemmed A new model of decision processing in instrumental learning tasks
title_short A new model of decision processing in instrumental learning tasks
title_sort new model of decision processing in instrumental learning tasks
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7880686/
https://www.ncbi.nlm.nih.gov/pubmed/33501916
http://dx.doi.org/10.7554/eLife.63055
work_keys_str_mv AT mileticsteven anewmodelofdecisionprocessingininstrumentallearningtasks
AT boagrussellj anewmodelofdecisionprocessingininstrumentallearningtasks
AT truttiannec anewmodelofdecisionprocessingininstrumentallearningtasks
AT stevensonniek anewmodelofdecisionprocessingininstrumentallearningtasks
AT forstmannbirteu anewmodelofdecisionprocessingininstrumentallearningtasks
AT heathcoteandrew anewmodelofdecisionprocessingininstrumentallearningtasks
AT mileticsteven newmodelofdecisionprocessingininstrumentallearningtasks
AT boagrussellj newmodelofdecisionprocessingininstrumentallearningtasks
AT truttiannec newmodelofdecisionprocessingininstrumentallearningtasks
AT stevensonniek newmodelofdecisionprocessingininstrumentallearningtasks
AT forstmannbirteu newmodelofdecisionprocessingininstrumentallearningtasks
AT heathcoteandrew newmodelofdecisionprocessingininstrumentallearningtasks