Cargando…

A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning

Learning in a new environment is influenced by prior learning and experience. Correctly applying a rule that maps a context to stimuli, actions, and outcomes enables faster learning and better outcomes compared to relying on strategies for learning that are ignorant of task structure. However, it is...

Descripción completa

Detalles Bibliográficos
Autores principales: Balcarras, Matthew, Womelsdorf, Thilo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4811957/
https://www.ncbi.nlm.nih.gov/pubmed/27064794
http://dx.doi.org/10.3389/fnins.2016.00125
_version_ 1782424061194797056
author Balcarras, Matthew
Womelsdorf, Thilo
author_facet Balcarras, Matthew
Womelsdorf, Thilo
author_sort Balcarras, Matthew
collection PubMed
description Learning in a new environment is influenced by prior learning and experience. Correctly applying a rule that maps a context to stimuli, actions, and outcomes enables faster learning and better outcomes compared to relying on strategies for learning that are ignorant of task structure. However, it is often difficult to know when and how to apply learned rules in new contexts. In our study we explored how subjects employ different strategies for learning the relationship between stimulus features and positive outcomes in a probabilistic task context. We test the hypothesis that task naive subjects will show enhanced learning of feature specific reward associations by switching to the use of an abstract rule that associates stimuli by feature type and restricts selections to that dimension. To test this hypothesis we designed a decision making task where subjects receive probabilistic feedback following choices between pairs of stimuli. In the task, trials are grouped in two contexts by blocks, where in one type of block there is no unique relationship between a specific feature dimension (stimulus shape or color) and positive outcomes, and following an un-cued transition, alternating blocks have outcomes that are linked to either stimulus shape or color. Two-thirds of subjects (n = 22/32) exhibited behavior that was best fit by a hierarchical feature-rule model. Supporting the prediction of the model mechanism these subjects showed significantly enhanced performance in feature-reward blocks, and rapidly switched their choice strategy to using abstract feature rules when reward contingencies changed. Choice behavior of other subjects (n = 10/32) was fit by a range of alternative reinforcement learning models representing strategies that do not benefit from applying previously learned rules. In summary, these results show that untrained subjects are capable of flexibly shifting between behavioral rules by leveraging simple model-free reinforcement learning and context-specific selections to drive responses.
format Online
Article
Text
id pubmed-4811957
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-48119572016-04-08 A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning Balcarras, Matthew Womelsdorf, Thilo Front Neurosci Neuroscience Learning in a new environment is influenced by prior learning and experience. Correctly applying a rule that maps a context to stimuli, actions, and outcomes enables faster learning and better outcomes compared to relying on strategies for learning that are ignorant of task structure. However, it is often difficult to know when and how to apply learned rules in new contexts. In our study we explored how subjects employ different strategies for learning the relationship between stimulus features and positive outcomes in a probabilistic task context. We test the hypothesis that task naive subjects will show enhanced learning of feature specific reward associations by switching to the use of an abstract rule that associates stimuli by feature type and restricts selections to that dimension. To test this hypothesis we designed a decision making task where subjects receive probabilistic feedback following choices between pairs of stimuli. In the task, trials are grouped in two contexts by blocks, where in one type of block there is no unique relationship between a specific feature dimension (stimulus shape or color) and positive outcomes, and following an un-cued transition, alternating blocks have outcomes that are linked to either stimulus shape or color. Two-thirds of subjects (n = 22/32) exhibited behavior that was best fit by a hierarchical feature-rule model. Supporting the prediction of the model mechanism these subjects showed significantly enhanced performance in feature-reward blocks, and rapidly switched their choice strategy to using abstract feature rules when reward contingencies changed. Choice behavior of other subjects (n = 10/32) was fit by a range of alternative reinforcement learning models representing strategies that do not benefit from applying previously learned rules. In summary, these results show that untrained subjects are capable of flexibly shifting between behavioral rules by leveraging simple model-free reinforcement learning and context-specific selections to drive responses. Frontiers Media S.A. 2016-03-30 /pmc/articles/PMC4811957/ /pubmed/27064794 http://dx.doi.org/10.3389/fnins.2016.00125 Text en Copyright © 2016 Balcarras and Womelsdorf. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Balcarras, Matthew
Womelsdorf, Thilo
A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning
title A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning
title_full A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning
title_fullStr A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning
title_full_unstemmed A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning
title_short A Flexible Mechanism of Rule Selection Enables Rapid Feature-Based Reinforcement Learning
title_sort flexible mechanism of rule selection enables rapid feature-based reinforcement learning
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4811957/
https://www.ncbi.nlm.nih.gov/pubmed/27064794
http://dx.doi.org/10.3389/fnins.2016.00125
work_keys_str_mv AT balcarrasmatthew aflexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning
AT womelsdorfthilo aflexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning
AT balcarrasmatthew flexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning
AT womelsdorfthilo flexiblemechanismofruleselectionenablesrapidfeaturebasedreinforcementlearning