Cargando…

Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events

Learning is a crucial basis for biological systems to adapt to environments. Environments include various states or episodes, and episode-dependent learning is essential in adaptation to such complex situations. Here, we developed a model for learning a two-target search task used in primate physiol...

Descripción completa

Detalles Bibliográficos
Autores principales: Sakamoto, Kazuhiro, Yamada, Hinata, Kawaguchi, Norihiko, Furusawa, Yoshito, Saito, Naohiro, Mushiake, Hajime
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9201426/
https://www.ncbi.nlm.nih.gov/pubmed/35720772
http://dx.doi.org/10.3389/fncom.2022.784604
_version_ 1784728318055022592
author Sakamoto, Kazuhiro
Yamada, Hinata
Kawaguchi, Norihiko
Furusawa, Yoshito
Saito, Naohiro
Mushiake, Hajime
author_facet Sakamoto, Kazuhiro
Yamada, Hinata
Kawaguchi, Norihiko
Furusawa, Yoshito
Saito, Naohiro
Mushiake, Hajime
author_sort Sakamoto, Kazuhiro
collection PubMed
description Learning is a crucial basis for biological systems to adapt to environments. Environments include various states or episodes, and episode-dependent learning is essential in adaptation to such complex situations. Here, we developed a model for learning a two-target search task used in primate physiological experiments. In the task, the agent is required to gaze one of the four presented light spots. Two neighboring spots are served as the correct target alternately, and the correct target pair is switched after a certain number of consecutive successes. In order for the agent to obtain rewards with a high probability, it is necessary to make decisions based on the actions and results of the previous two trials. Our previous work achieved this by using a dynamic state space. However, to learn a task that includes events such as fixation to the initial central spot, the model framework should be extended. For this purpose, here we propose a “history-in-episode architecture.” Specifically, we divide states into episodes and histories, and actions are selected based on the histories within each episode. When we compared the proposed model including the dynamic state space with the conventional SARSA method in the two-target search task, the former performed close to the theoretical optimum, while the latter never achieved target-pair switch because it had to re-learn each correct target each time. The reinforcement learning model including the proposed history-in-episode architecture and dynamic state scape enables episode-dependent learning and provides a basis for highly adaptable learning systems to complex environments.
format Online
Article
Text
id pubmed-9201426
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-92014262022-06-17 Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events Sakamoto, Kazuhiro Yamada, Hinata Kawaguchi, Norihiko Furusawa, Yoshito Saito, Naohiro Mushiake, Hajime Front Comput Neurosci Neuroscience Learning is a crucial basis for biological systems to adapt to environments. Environments include various states or episodes, and episode-dependent learning is essential in adaptation to such complex situations. Here, we developed a model for learning a two-target search task used in primate physiological experiments. In the task, the agent is required to gaze one of the four presented light spots. Two neighboring spots are served as the correct target alternately, and the correct target pair is switched after a certain number of consecutive successes. In order for the agent to obtain rewards with a high probability, it is necessary to make decisions based on the actions and results of the previous two trials. Our previous work achieved this by using a dynamic state space. However, to learn a task that includes events such as fixation to the initial central spot, the model framework should be extended. For this purpose, here we propose a “history-in-episode architecture.” Specifically, we divide states into episodes and histories, and actions are selected based on the histories within each episode. When we compared the proposed model including the dynamic state space with the conventional SARSA method in the two-target search task, the former performed close to the theoretical optimum, while the latter never achieved target-pair switch because it had to re-learn each correct target each time. The reinforcement learning model including the proposed history-in-episode architecture and dynamic state scape enables episode-dependent learning and provides a basis for highly adaptable learning systems to complex environments. Frontiers Media S.A. 2022-06-02 /pmc/articles/PMC9201426/ /pubmed/35720772 http://dx.doi.org/10.3389/fncom.2022.784604 Text en Copyright © 2022 Sakamoto, Yamada, Kawaguchi, Furusawa, Saito and Mushiake. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Sakamoto, Kazuhiro
Yamada, Hinata
Kawaguchi, Norihiko
Furusawa, Yoshito
Saito, Naohiro
Mushiake, Hajime
Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events
title Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events
title_full Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events
title_fullStr Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events
title_full_unstemmed Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events
title_short Reinforcement Learning Model With Dynamic State Space Tested on Target Search Tasks for Monkeys: Extension to Learning Task Events
title_sort reinforcement learning model with dynamic state space tested on target search tasks for monkeys: extension to learning task events
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9201426/
https://www.ncbi.nlm.nih.gov/pubmed/35720772
http://dx.doi.org/10.3389/fncom.2022.784604
work_keys_str_mv AT sakamotokazuhiro reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents
AT yamadahinata reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents
AT kawaguchinorihiko reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents
AT furusawayoshito reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents
AT saitonaohiro reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents
AT mushiakehajime reinforcementlearningmodelwithdynamicstatespacetestedontargetsearchtasksformonkeysextensiontolearningtaskevents