Cargando…
A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning
Learning in a new environment is influenced by prior learning and experience. Correctly applying a rule that maps a context to stimuli, actions, and outcomes enables faster learning and better outcomes compared to relying on strategies for learning that are ignorant of task structure. However, it is...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2016
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4811957/ https://www.ncbi.nlm.nih.gov/pubmed/27064794 http://dx.doi.org/10.3389/fnins.2016.00125 |
_version_ | 1782424061194797056 |
---|---|
author | Balcarras, Matthew Womelsdorf, Thilo |
author_facet | Balcarras, Matthew Womelsdorf, Thilo |
author_sort | Balcarras, Matthew |
collection | PubMed |
description | Learning in a new environment is influenced by prior learning and experience. Correctly applying a rule that maps a context to stimuli, actions, and outcomes enables faster learning and better outcomes compared to relying on strategies for learning that are ignorant of task structure. However, it is often difficult to know when and how to apply learned rules in new contexts. In our study we explored how subjects employ different strategies for learning the relationship between stimulus features and positive outcomes in a probabilistic task context. We test the hypothesis that task naive subjects will show enhanced learning of feature specific reward associations by switching to the use of an abstract rule that associates stimuli by feature type and restricts selections to that dimension. To test this hypothesis we designed a decision making task where subjects receive probabilistic feedback following choices between pairs of stimuli. In the task, trials are grouped in two contexts by blocks, where in one type of block there is no unique relationship between a specific feature dimension (stimulus shape or color) and positive outcomes, and following an un-cued transition, alternating blocks have outcomes that are linked to either stimulus shape or color. Two-thirds of subjects (n = 22/32) exhibited behavior that was best fit by a hierarchical feature-rule model. Supporting the prediction of the model mechanism these subjects showed significantly enhanced performance in feature-reward blocks, and rapidly switched their choice strategy to using abstract feature rules when reward contingencies changed. Choice behavior of other subjects (n = 10/32) was fit by a range of alternative reinforcement learning models representing strategies that do not benefit from applying previously learned rules. In summary, these results show that untrained subjects are capable of flexibly shifting between behavioral rules by leveraging simple model-free reinforcement learning and context-specific selections to drive responses. |
format | Online Article Text |
id | pubmed-4811957 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2016 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-48119572016-04-08 A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning Balcarras, Matthew Womelsdorf, Thilo Front Neurosci Neuroscience Learning in a new environment is influenced by prior learning and experience. Correctly applying a rule that maps a context to stimuli, actions, and outcomes enables faster learning and better outcomes compared to relying on strategies for learning that are ignorant of task structure. However, it is often difficult to know when and how to apply learned rules in new contexts. In our study we explored how subjects employ different strategies for learning the relationship between stimulus features and positive outcomes in a probabilistic task context. We test the hypothesis that task naive subjects will show enhanced learning of feature specific reward associations by switching to the use of an abstract rule that associates stimuli by feature type and restricts selections to that dimension. To test this hypothesis we designed a decision making task where subjects receive probabilistic feedback following choices between pairs of stimuli. In the task, trials are grouped in two contexts by blocks, where in one type of block there is no unique relationship between a specific feature dimension (stimulus shape or color) and positive outcomes, and following an un-cued transition, alternating blocks have outcomes that are linked to either stimulus shape or color. Two-thirds of subjects (n = 22/32) exhibited behavior that was best fit by a hierarchical feature-rule model. Supporting the prediction of the model mechanism these subjects showed significantly enhanced performance in feature-reward blocks, and rapidly switched their choice strategy to using abstract feature rules when reward contingencies changed. Choice behavior of other subjects (n = 10/32) was fit by a range of alternative reinforcement learning models representing strategies that do not benefit from applying previously learned rules. In summary, these results show that untrained subjects are capable of flexibly shifting between behavioral rules by leveraging simple model-free reinforcement learning and context-specific selections to drive responses. Frontiers Media S.A. 2016-03-30 /pmc/articles/PMC4811957/ /pubmed/27064794 http://dx.doi.org/10.3389/fnins.2016.00125 Text en Copyright © 2016 Balcarras and Womelsdorf. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Balcarras, Matthew Womelsdorf, Thilo A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning |
title | A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning |
title_full | A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning |
title_fullStr | A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning |
title_full_unstemmed | A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning |
title_short | A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning |
title_sort | flexible mechanism of rule selection enables rapid feature-based reinforcement learning |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4811957/ https://www.ncbi.nlm.nih.gov/pubmed/27064794 http://dx.doi.org/10.3389/fnins.2016.00125 |
work_keys_str_mv | AT balcarrasmatthew aflexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning AT womelsdorfthilo aflexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning AT balcarrasmatthew flexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning AT womelsdorfthilo flexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning |