Cargando…
Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game
We study a simple model for social-learning agents in a restless multiarmed bandit (rMAB). The bandit has one good arm that changes to a bad one with a certain probability. Each agent stochastically selects one of the two methods, random search (individual learning) or copying information from other...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5434024/ https://www.ncbi.nlm.nih.gov/pubmed/28512339 http://dx.doi.org/10.1038/s41598-017-01750-z |
_version_ | 1783236959138742272 |
---|---|
author | Nakayama, Kazuaki Hisakado, Masato Mori, Shintaro |
author_facet | Nakayama, Kazuaki Hisakado, Masato Mori, Shintaro |
author_sort | Nakayama, Kazuaki |
collection | PubMed |
description | We study a simple model for social-learning agents in a restless multiarmed bandit (rMAB). The bandit has one good arm that changes to a bad one with a certain probability. Each agent stochastically selects one of the two methods, random search (individual learning) or copying information from other agents (social learning), using which he/she seeks the good arm. Fitness of an agent is the probability to know the good arm in the steady state of the agent system. In this model, we explicitly construct the unique Nash equilibrium state and show that the corresponding strategy for each agent is an evolutionarily stable strategy (ESS) in the sense of Thomas. It is shown that the fitness of an agent with ESS is superior to that of an asocial learner when the success probability of social learning is greater than a threshold determined from the probability of success of individual learning, the probability of change of state of the rMAB, and the number of agents. The ESS Nash equilibrium is a solution to Rogers’ paradox. |
format | Online Article Text |
id | pubmed-5434024 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-54340242017-05-17 Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game Nakayama, Kazuaki Hisakado, Masato Mori, Shintaro Sci Rep Article We study a simple model for social-learning agents in a restless multiarmed bandit (rMAB). The bandit has one good arm that changes to a bad one with a certain probability. Each agent stochastically selects one of the two methods, random search (individual learning) or copying information from other agents (social learning), using which he/she seeks the good arm. Fitness of an agent is the probability to know the good arm in the steady state of the agent system. In this model, we explicitly construct the unique Nash equilibrium state and show that the corresponding strategy for each agent is an evolutionarily stable strategy (ESS) in the sense of Thomas. It is shown that the fitness of an agent with ESS is superior to that of an asocial learner when the success probability of social learning is greater than a threshold determined from the probability of success of individual learning, the probability of change of state of the rMAB, and the number of agents. The ESS Nash equilibrium is a solution to Rogers’ paradox. Nature Publishing Group UK 2017-05-16 /pmc/articles/PMC5434024/ /pubmed/28512339 http://dx.doi.org/10.1038/s41598-017-01750-z Text en © The Author(s) 2017 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/. |
spellingShingle | Article Nakayama, Kazuaki Hisakado, Masato Mori, Shintaro Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game |
title | Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game |
title_full | Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game |
title_fullStr | Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game |
title_full_unstemmed | Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game |
title_short | Nash Equilibrium of Social-Learning Agents in a Restless Multiarmed Bandit Game |
title_sort | nash equilibrium of social-learning agents in a restless multiarmed bandit game |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5434024/ https://www.ncbi.nlm.nih.gov/pubmed/28512339 http://dx.doi.org/10.1038/s41598-017-01750-z |
work_keys_str_mv | AT nakayamakazuaki nashequilibriumofsociallearningagentsinarestlessmultiarmedbanditgame AT hisakadomasato nashequilibriumofsociallearningagentsinarestlessmultiarmedbanditgame AT morishintaro nashequilibriumofsociallearningagentsinarestlessmultiarmedbanditgame |