Cargando…
Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events
Learning is a crucial basis for biological systems to adapt to environments. Environments include various states or episodes, and episode-dependent learning is essential in adaptation to such complex situations. Here, we developed a model for learning a two-target search task used in primate physiol...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9201426/ https://www.ncbi.nlm.nih.gov/pubmed/35720772 http://dx.doi.org/10.3389/fncom.2022.784604 |
_version_ | 1784728318055022592 |
---|---|
author | Sakamoto, Kazuhiro Yamada, Hinata Kawaguchi, Norihiko Furusawa, Yoshito Saito, Naohiro Mushiake, Hajime |
author_facet | Sakamoto, Kazuhiro Yamada, Hinata Kawaguchi, Norihiko Furusawa, Yoshito Saito, Naohiro Mushiake, Hajime |
author_sort | Sakamoto, Kazuhiro |
collection | PubMed |
description | Learning is a crucial basis for biological systems to adapt to environments. Environments include various states or episodes, and episode-dependent learning is essential in adaptation to such complex situations. Here, we developed a model for learning a two-target search task used in primate physiological experiments. In the task, the agent is required to gaze one of the four presented light spots. Two neighboring spots are served as the correct target alternately, and the correct target pair is switched after a certain number of consecutive successes. In order for the agent to obtain rewards with a high probability, it is necessary to make decisions based on the actions and results of the previous two trials. Our previous work achieved this by using a dynamic state space. However, to learn a task that includes events such as fixation to the initial central spot, the model framework should be extended. For this purpose, here we propose a “history-in-episode architecture.” Specifically, we divide states into episodes and histories, and actions are selected based on the histories within each episode. When we compared the proposed model including the dynamic state space with the conventional SARSA method in the two-target search task, the former performed close to the theoretical optimum, while the latter never achieved target-pair switch because it had to re-learn each correct target each time. The reinforcement learning model including the proposed history-in-episode architecture and dynamic state scape enables episode-dependent learning and provides a basis for highly adaptable learning systems to complex environments. |
format | Online Article Text |
id | pubmed-9201426 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-92014262022-06-17 Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events Sakamoto, Kazuhiro Yamada, Hinata Kawaguchi, Norihiko Furusawa, Yoshito Saito, Naohiro Mushiake, Hajime Front Comput Neurosci Neuroscience Learning is a crucial basis for biological systems to adapt to environments. Environments include various states or episodes, and episode-dependent learning is essential in adaptation to such complex situations. Here, we developed a model for learning a two-target search task used in primate physiological experiments. In the task, the agent is required to gaze one of the four presented light spots. Two neighboring spots are served as the correct target alternately, and the correct target pair is switched after a certain number of consecutive successes. In order for the agent to obtain rewards with a high probability, it is necessary to make decisions based on the actions and results of the previous two trials. Our previous work achieved this by using a dynamic state space. However, to learn a task that includes events such as fixation to the initial central spot, the model framework should be extended. For this purpose, here we propose a “history-in-episode architecture.” Specifically, we divide states into episodes and histories, and actions are selected based on the histories within each episode. When we compared the proposed model including the dynamic state space with the conventional SARSA method in the two-target search task, the former performed close to the theoretical optimum, while the latter never achieved target-pair switch because it had to re-learn each correct target each time. The reinforcement learning model including the proposed history-in-episode architecture and dynamic state scape enables episode-dependent learning and provides a basis for highly adaptable learning systems to complex environments. Frontiers Media S.A. 2022-06-02 /pmc/articles/PMC9201426/ /pubmed/35720772 http://dx.doi.org/10.3389/fncom.2022.784604 Text en Copyright © 2022 Sakamoto, Yamada, Kawaguchi, Furusawa, Saito and Mushiake. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Sakamoto, Kazuhiro Yamada, Hinata Kawaguchi, Norihiko Furusawa, Yoshito Saito, Naohiro Mushiake, Hajime Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events |
title | Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events |
title_full | Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events |
title_fullStr | Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events |
title_full_unstemmed | Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events |
title_short | Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events |
title_sort | reinforcement learning model with dynamic state space tested on target search tasks for monkeys: extension to learning task events |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9201426/ https://www.ncbi.nlm.nih.gov/pubmed/35720772 http://dx.doi.org/10.3389/fncom.2022.784604 |
work_keys_str_mv | AT sakamotokazuhiro reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents AT yamadahinata reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents AT kawaguchinorihiko reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents AT furusawayoshito reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents AT saitonaohiro reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents AT mushiakehajime reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents |