Cargando…

Context Transfer in Reinforcement Learning Using Action-Value Functions

This paper discusses the notion of context transfer in reinforcement learning tasks. Context transfer, as defined in this paper, implies knowledge transfer between source and target tasks that share the same environment dynamics and reward function but have different states or action spaces. In othe...

Descripción completa

Detalles Bibliográficos
Autores principales: Mousavi, Amin, Nadjar Araabi, Babak, Nili Ahmadabadi, Majid
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi Publishing Corporation 2014
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4293791/
https://www.ncbi.nlm.nih.gov/pubmed/25610457
http://dx.doi.org/10.1155/2014/428567
_version_ 1782352645522980864
author Mousavi, Amin
Nadjar Araabi, Babak
Nili Ahmadabadi, Majid
author_facet Mousavi, Amin
Nadjar Araabi, Babak
Nili Ahmadabadi, Majid
author_sort Mousavi, Amin
collection PubMed
description This paper discusses the notion of context transfer in reinforcement learning tasks. Context transfer, as defined in this paper, implies knowledge transfer between source and target tasks that share the same environment dynamics and reward function but have different states or action spaces. In other words, the agents learn the same task while using different sensors and actuators. This requires the existence of an underlying common Markov decision process (MDP) to which all the agents' MDPs can be mapped. This is formulated in terms of the notion of MDP homomorphism. The learning framework is Q-learning. To transfer the knowledge between these tasks, the feature space is used as a translator and is expressed as a partial mapping between the state-action spaces of different tasks. The Q-values learned during the learning process of the source tasks are mapped to the sets of Q-values for the target task. These transferred Q-values are merged together and used to initialize the learning process of the target task. An interval-based approach is used to represent and merge the knowledge of the source tasks. Empirical results show that the transferred initialization can be beneficial to the learning process of the target task.
format Online
Article
Text
id pubmed-4293791
institution National Center for Biotechnology Information
language English
publishDate 2014
publisher Hindawi Publishing Corporation
record_format MEDLINE/PubMed
spelling pubmed-42937912015-01-21 Context Transfer in Reinforcement Learning Using Action-Value Functions Mousavi, Amin Nadjar Araabi, Babak Nili Ahmadabadi, Majid Comput Intell Neurosci Research Article This paper discusses the notion of context transfer in reinforcement learning tasks. Context transfer, as defined in this paper, implies knowledge transfer between source and target tasks that share the same environment dynamics and reward function but have different states or action spaces. In other words, the agents learn the same task while using different sensors and actuators. This requires the existence of an underlying common Markov decision process (MDP) to which all the agents' MDPs can be mapped. This is formulated in terms of the notion of MDP homomorphism. The learning framework is Q-learning. To transfer the knowledge between these tasks, the feature space is used as a translator and is expressed as a partial mapping between the state-action spaces of different tasks. The Q-values learned during the learning process of the source tasks are mapped to the sets of Q-values for the target task. These transferred Q-values are merged together and used to initialize the learning process of the target task. An interval-based approach is used to represent and merge the knowledge of the source tasks. Empirical results show that the transferred initialization can be beneficial to the learning process of the target task. Hindawi Publishing Corporation 2014 2014-12-31 /pmc/articles/PMC4293791/ /pubmed/25610457 http://dx.doi.org/10.1155/2014/428567 Text en Copyright © 2014 Amin Mousavi et al. https://creativecommons.org/licenses/by/3.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Mousavi, Amin
Nadjar Araabi, Babak
Nili Ahmadabadi, Majid
Context Transfer in Reinforcement Learning Using Action-Value Functions
title Context Transfer in Reinforcement Learning Using Action-Value Functions
title_full Context Transfer in Reinforcement Learning Using Action-Value Functions
title_fullStr Context Transfer in Reinforcement Learning Using Action-Value Functions
title_full_unstemmed Context Transfer in Reinforcement Learning Using Action-Value Functions
title_short Context Transfer in Reinforcement Learning Using Action-Value Functions
title_sort context transfer in reinforcement learning using action-value functions
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4293791/
https://www.ncbi.nlm.nih.gov/pubmed/25610457
http://dx.doi.org/10.1155/2014/428567
work_keys_str_mv AT mousaviamin contexttransferinreinforcementlearningusingactionvaluefunctions
AT nadjararaabibabak contexttransferinreinforcementlearningusingactionvaluefunctions
AT niliahmadabadimajid contexttransferinreinforcementlearningusingactionvaluefunctions