Cargando…
Compositional RL Agents That Follow Language Commands in Temporal Logic
We demonstrate how a reinforcement learning agent can use compositional recurrent neural networks to learn to carry out commands specified in linear temporal logic (LTL). Our approach takes as input an LTL formula, structures a deep network according to the parse of the formula, and determines satis...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8326833/ https://www.ncbi.nlm.nih.gov/pubmed/34350213 http://dx.doi.org/10.3389/frobt.2021.689550 |
_version_ | 1783731933529767936 |
---|---|
author | Kuo, Yen-Ling Katz, Boris Barbu, Andrei |
author_facet | Kuo, Yen-Ling Katz, Boris Barbu, Andrei |
author_sort | Kuo, Yen-Ling |
collection | PubMed |
description | We demonstrate how a reinforcement learning agent can use compositional recurrent neural networks to learn to carry out commands specified in linear temporal logic (LTL). Our approach takes as input an LTL formula, structures a deep network according to the parse of the formula, and determines satisfying actions. This compositional structure of the network enables zero-shot generalization to significantly more complex unseen formulas. We demonstrate this ability in multiple problem domains with both discrete and continuous state-action spaces. In a symbolic domain, the agent finds a sequence of letters that satisfy a specification. In a Minecraft-like environment, the agent finds a sequence of actions that conform to a formula. In the Fetch environment, the robot finds a sequence of arm configurations that move blocks on a table to fulfill the commands. While most prior work can learn to execute one formula reliably, we develop a novel form of multi-task learning for RL agents that allows them to learn from a diverse set of tasks and generalize to a new set of diverse tasks without any additional training. The compositional structures presented here are not specific to LTL, thus opening the path to RL agents that perform zero-shot generalization in other compositional domains. |
format | Online Article Text |
id | pubmed-8326833 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-83268332021-08-03 Compositional RL Agents That Follow Language Commands in Temporal Logic Kuo, Yen-Ling Katz, Boris Barbu, Andrei Front Robot AI Robotics and AI We demonstrate how a reinforcement learning agent can use compositional recurrent neural networks to learn to carry out commands specified in linear temporal logic (LTL). Our approach takes as input an LTL formula, structures a deep network according to the parse of the formula, and determines satisfying actions. This compositional structure of the network enables zero-shot generalization to significantly more complex unseen formulas. We demonstrate this ability in multiple problem domains with both discrete and continuous state-action spaces. In a symbolic domain, the agent finds a sequence of letters that satisfy a specification. In a Minecraft-like environment, the agent finds a sequence of actions that conform to a formula. In the Fetch environment, the robot finds a sequence of arm configurations that move blocks on a table to fulfill the commands. While most prior work can learn to execute one formula reliably, we develop a novel form of multi-task learning for RL agents that allows them to learn from a diverse set of tasks and generalize to a new set of diverse tasks without any additional training. The compositional structures presented here are not specific to LTL, thus opening the path to RL agents that perform zero-shot generalization in other compositional domains. Frontiers Media S.A. 2021-07-19 /pmc/articles/PMC8326833/ /pubmed/34350213 http://dx.doi.org/10.3389/frobt.2021.689550 Text en Copyright © 2021 Kuo, Katz and Barbu. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Robotics and AI Kuo, Yen-Ling Katz, Boris Barbu, Andrei Compositional RL Agents That Follow Language Commands in Temporal Logic |
title | Compositional RL Agents That Follow Language Commands in Temporal Logic |
title_full | Compositional RL Agents That Follow Language Commands in Temporal Logic |
title_fullStr | Compositional RL Agents That Follow Language Commands in Temporal Logic |
title_full_unstemmed | Compositional RL Agents That Follow Language Commands in Temporal Logic |
title_short | Compositional RL Agents That Follow Language Commands in Temporal Logic |
title_sort | compositional rl agents that follow language commands in temporal logic |
topic | Robotics and AI |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8326833/ https://www.ncbi.nlm.nih.gov/pubmed/34350213 http://dx.doi.org/10.3389/frobt.2021.689550 |
work_keys_str_mv | AT kuoyenling compositionalrlagentsthatfollowlanguagecommandsintemporallogic AT katzboris compositionalrlagentsthatfollowlanguagecommandsintemporallogic AT barbuandrei compositionalrlagentsthatfollowlanguagecommandsintemporallogic |