Cargando…
Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria
We study the distribution of strategies in a large game that models how agents choose among different double auction markets. We classify the possible mean field Nash equilibria, which include potentially segregated states where an agent population can split into subpopulations adopting different st...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6037372/ https://www.ncbi.nlm.nih.gov/pubmed/29985923 http://dx.doi.org/10.1371/journal.pone.0196577 |
_version_ | 1783338318776238080 |
---|---|
author | Nicole, Robin Sollich, Peter |
author_facet | Nicole, Robin Sollich, Peter |
author_sort | Nicole, Robin |
collection | PubMed |
description | We study the distribution of strategies in a large game that models how agents choose among different double auction markets. We classify the possible mean field Nash equilibria, which include potentially segregated states where an agent population can split into subpopulations adopting different strategies. As the game is aggregative, the actual equilibrium strategy distributions remain undetermined, however. We therefore compare with the results of a reinforcement learning dynamics inspired by Experience-Weighted Attraction (EWA) learning, which at long times leads to Nash equilibria in the appropriate limits of large intensity of choice, low noise (long agent memory) and perfect imputation of missing scores (fictitious play). The learning dynamics breaks the indeterminacy of the Nash equilibria. Non-trivially, depending on how the relevant limits are taken, more than one type of equilibrium can be selected. These include the standard homogeneous mixed and heterogeneous pure states, but also heterogeneous mixed states where different agents play different strategies that are not all pure. The analysis of the reinforcement learning involves Fokker-Planck modeling combined with large deviation methods. The theoretical results are confirmed by multi-agent simulations. |
format | Online Article Text |
id | pubmed-6037372 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-60373722018-07-19 Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria Nicole, Robin Sollich, Peter PLoS One Research Article We study the distribution of strategies in a large game that models how agents choose among different double auction markets. We classify the possible mean field Nash equilibria, which include potentially segregated states where an agent population can split into subpopulations adopting different strategies. As the game is aggregative, the actual equilibrium strategy distributions remain undetermined, however. We therefore compare with the results of a reinforcement learning dynamics inspired by Experience-Weighted Attraction (EWA) learning, which at long times leads to Nash equilibria in the appropriate limits of large intensity of choice, low noise (long agent memory) and perfect imputation of missing scores (fictitious play). The learning dynamics breaks the indeterminacy of the Nash equilibria. Non-trivially, depending on how the relevant limits are taken, more than one type of equilibrium can be selected. These include the standard homogeneous mixed and heterogeneous pure states, but also heterogeneous mixed states where different agents play different strategies that are not all pure. The analysis of the reinforcement learning involves Fokker-Planck modeling combined with large deviation methods. The theoretical results are confirmed by multi-agent simulations. Public Library of Science 2018-07-09 /pmc/articles/PMC6037372/ /pubmed/29985923 http://dx.doi.org/10.1371/journal.pone.0196577 Text en © 2018 Nicole, Sollich http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Nicole, Robin Sollich, Peter Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria |
title | Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria |
title_full | Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria |
title_fullStr | Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria |
title_full_unstemmed | Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria |
title_short | Dynamical selection of Nash equilibria using reinforcement learning: Emergence of heterogeneous mixed equilibria |
title_sort | dynamical selection of nash equilibria using reinforcement learning: emergence of heterogeneous mixed equilibria |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6037372/ https://www.ncbi.nlm.nih.gov/pubmed/29985923 http://dx.doi.org/10.1371/journal.pone.0196577 |
work_keys_str_mv | AT nicolerobin dynamicalselectionofnashequilibriausingreinforcementlearningemergenceofheterogeneousmixedequilibria AT sollichpeter dynamicalselectionofnashequilibriausingreinforcementlearningemergenceofheterogeneousmixedequilibria |