Cargando…
Optimizing the depth and the direction of prospective planning using information values
Evaluating the future consequences of actions is achievable by simulating a mental search tree into the future. Expanding deep trees, however, is computationally taxing. Therefore, machines and humans use a plan-until-habit scheme that simulates the environment up to a limited depth and then exploit...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6440644/ https://www.ncbi.nlm.nih.gov/pubmed/30861001 http://dx.doi.org/10.1371/journal.pcbi.1006827 |
_version_ | 1783407427969875968 |
---|---|
author | Sezener, Can Eren Dezfouli, Amir Keramati, Mehdi |
author_facet | Sezener, Can Eren Dezfouli, Amir Keramati, Mehdi |
author_sort | Sezener, Can Eren |
collection | PubMed |
description | Evaluating the future consequences of actions is achievable by simulating a mental search tree into the future. Expanding deep trees, however, is computationally taxing. Therefore, machines and humans use a plan-until-habit scheme that simulates the environment up to a limited depth and then exploits habitual values as proxies for consequences that may arise in the future. Two outstanding questions in this scheme are “in which directions the search tree should be expanded?”, and “when should the expansion stop?”. Here we propose a principled solution to these questions based on a speed/accuracy tradeoff: deeper expansion in the appropriate directions leads to more accurate planning, but at the cost of slower decision-making. Our simulation results show how this algorithm expands the search tree effectively and efficiently in a grid-world environment. We further show that our algorithm can explain several behavioral patterns in animals and humans, namely the effect of time-pressure on the depth of planning, the effect of reward magnitudes on the direction of planning, and the gradual shift from goal-directed to habitual behavior over the course of training. The algorithm also provides several predictions testable in animal/human experiments. |
format | Online Article Text |
id | pubmed-6440644 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-64406442019-04-12 Optimizing the depth and the direction of prospective planning using information values Sezener, Can Eren Dezfouli, Amir Keramati, Mehdi PLoS Comput Biol Research Article Evaluating the future consequences of actions is achievable by simulating a mental search tree into the future. Expanding deep trees, however, is computationally taxing. Therefore, machines and humans use a plan-until-habit scheme that simulates the environment up to a limited depth and then exploits habitual values as proxies for consequences that may arise in the future. Two outstanding questions in this scheme are “in which directions the search tree should be expanded?”, and “when should the expansion stop?”. Here we propose a principled solution to these questions based on a speed/accuracy tradeoff: deeper expansion in the appropriate directions leads to more accurate planning, but at the cost of slower decision-making. Our simulation results show how this algorithm expands the search tree effectively and efficiently in a grid-world environment. We further show that our algorithm can explain several behavioral patterns in animals and humans, namely the effect of time-pressure on the depth of planning, the effect of reward magnitudes on the direction of planning, and the gradual shift from goal-directed to habitual behavior over the course of training. The algorithm also provides several predictions testable in animal/human experiments. Public Library of Science 2019-03-12 /pmc/articles/PMC6440644/ /pubmed/30861001 http://dx.doi.org/10.1371/journal.pcbi.1006827 Text en © 2019 Sezener et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Sezener, Can Eren Dezfouli, Amir Keramati, Mehdi Optimizing the depth and the direction of prospective planning using information values |
title | Optimizing the depth and the direction of prospective planning using information values |
title_full | Optimizing the depth and the direction of prospective planning using information values |
title_fullStr | Optimizing the depth and the direction of prospective planning using information values |
title_full_unstemmed | Optimizing the depth and the direction of prospective planning using information values |
title_short | Optimizing the depth and the direction of prospective planning using information values |
title_sort | optimizing the depth and the direction of prospective planning using information values |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6440644/ https://www.ncbi.nlm.nih.gov/pubmed/30861001 http://dx.doi.org/10.1371/journal.pcbi.1006827 |
work_keys_str_mv | AT sezenercaneren optimizingthedepthandthedirectionofprospectiveplanningusinginformationvalues AT dezfouliamir optimizingthedepthandthedirectionofprospectiveplanningusinginformationvalues AT keramatimehdi optimizingthedepthandthedirectionofprospectiveplanningusinginformationvalues |