Cargando…
A new model of decision processing in instrumental learning tasks
Learning and decision-making are interactive processes, yet cognitive modeling of error-driven learning and decision-making have largely evolved separately. Recently, evidence accumulation models (EAMs) of decision-making and reinforcement learning (RL) models of error-driven learning have been comb...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
eLife Sciences Publications, Ltd
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7880686/ https://www.ncbi.nlm.nih.gov/pubmed/33501916 http://dx.doi.org/10.7554/eLife.63055 |
_version_ | 1783650746966736896 |
---|---|
author | Miletić, Steven Boag, Russell J Trutti, Anne C Stevenson, Niek Forstmann, Birte U Heathcote, Andrew |
author_facet | Miletić, Steven Boag, Russell J Trutti, Anne C Stevenson, Niek Forstmann, Birte U Heathcote, Andrew |
author_sort | Miletić, Steven |
collection | PubMed |
description | Learning and decision-making are interactive processes, yet cognitive modeling of error-driven learning and decision-making have largely evolved separately. Recently, evidence accumulation models (EAMs) of decision-making and reinforcement learning (RL) models of error-driven learning have been combined into joint RL-EAMs that can in principle address these interactions. However, we show that the most commonly used combination, based on the diffusion decision model (DDM) for binary choice, consistently fails to capture crucial aspects of response times observed during reinforcement learning. We propose a new RL-EAM based on an advantage racing diffusion (ARD) framework for choices among two or more options that not only addresses this problem but captures stimulus difficulty, speed-accuracy trade-off, and stimulus-response-mapping reversal effects. The RL-ARD avoids fundamental limitations imposed by the DDM on addressing effects of absolute values of choices, as well as extensions beyond binary choice, and provides a computationally tractable basis for wider applications. |
format | Online Article Text |
id | pubmed-7880686 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | eLife Sciences Publications, Ltd |
record_format | MEDLINE/PubMed |
spelling | pubmed-78806862021-02-16 A new model of decision processing in instrumental learning tasks Miletić, Steven Boag, Russell J Trutti, Anne C Stevenson, Niek Forstmann, Birte U Heathcote, Andrew eLife Neuroscience Learning and decision-making are interactive processes, yet cognitive modeling of error-driven learning and decision-making have largely evolved separately. Recently, evidence accumulation models (EAMs) of decision-making and reinforcement learning (RL) models of error-driven learning have been combined into joint RL-EAMs that can in principle address these interactions. However, we show that the most commonly used combination, based on the diffusion decision model (DDM) for binary choice, consistently fails to capture crucial aspects of response times observed during reinforcement learning. We propose a new RL-EAM based on an advantage racing diffusion (ARD) framework for choices among two or more options that not only addresses this problem but captures stimulus difficulty, speed-accuracy trade-off, and stimulus-response-mapping reversal effects. The RL-ARD avoids fundamental limitations imposed by the DDM on addressing effects of absolute values of choices, as well as extensions beyond binary choice, and provides a computationally tractable basis for wider applications. eLife Sciences Publications, Ltd 2021-01-27 /pmc/articles/PMC7880686/ /pubmed/33501916 http://dx.doi.org/10.7554/eLife.63055 Text en © 2021, Miletić et al http://creativecommons.org/licenses/by/4.0/ http://creativecommons.org/licenses/by/4.0/This article is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited. |
spellingShingle | Neuroscience Miletić, Steven Boag, Russell J Trutti, Anne C Stevenson, Niek Forstmann, Birte U Heathcote, Andrew A new model of decision processing in instrumental learning tasks |
title | A new model of decision processing in instrumental learning tasks |
title_full | A new model of decision processing in instrumental learning tasks |
title_fullStr | A new model of decision processing in instrumental learning tasks |
title_full_unstemmed | A new model of decision processing in instrumental learning tasks |
title_short | A new model of decision processing in instrumental learning tasks |
title_sort | new model of decision processing in instrumental learning tasks |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7880686/ https://www.ncbi.nlm.nih.gov/pubmed/33501916 http://dx.doi.org/10.7554/eLife.63055 |
work_keys_str_mv | AT mileticsteven anewmodelofdecisionprocessingininstrumentallearningtasks AT boagrussellj anewmodelofdecisionprocessingininstrumentallearningtasks AT truttiannec anewmodelofdecisionprocessingininstrumentallearningtasks AT stevensonniek anewmodelofdecisionprocessingininstrumentallearningtasks AT forstmannbirteu anewmodelofdecisionprocessingininstrumentallearningtasks AT heathcoteandrew anewmodelofdecisionprocessingininstrumentallearningtasks AT mileticsteven newmodelofdecisionprocessingininstrumentallearningtasks AT boagrussellj newmodelofdecisionprocessingininstrumentallearningtasks AT truttiannec newmodelofdecisionprocessingininstrumentallearningtasks AT stevensonniek newmodelofdecisionprocessingininstrumentallearningtasks AT forstmannbirteu newmodelofdecisionprocessingininstrumentallearningtasks AT heathcoteandrew newmodelofdecisionprocessingininstrumentallearningtasks |