Cargando…

Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions

An important characteristic of human language is compositionality. We can efficiently express a wide variety of real-world situations, events, and behaviors by compositionally constructing the meaning of a complex expression from a finite number of elements. Previous studies have analyzed how machin...

Descripción completa

Detalles Bibliográficos
Autores principales: Yamada, Tatsuro, Murata, Shingo, Arie, Hiroaki, Ogata, Tetsuya
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5744442/
https://www.ncbi.nlm.nih.gov/pubmed/29311891
http://dx.doi.org/10.3389/fnbot.2017.00070
_version_ 1783288745728933888
author Yamada, Tatsuro
Murata, Shingo
Arie, Hiroaki
Ogata, Tetsuya
author_facet Yamada, Tatsuro
Murata, Shingo
Arie, Hiroaki
Ogata, Tetsuya
author_sort Yamada, Tatsuro
collection PubMed
description An important characteristic of human language is compositionality. We can efficiently express a wide variety of real-world situations, events, and behaviors by compositionally constructing the meaning of a complex expression from a finite number of elements. Previous studies have analyzed how machine-learning models, particularly neural networks, can learn from experience to represent compositional relationships between language and robot actions with the aim of understanding the symbol grounding structure and achieving intelligent communicative agents. Such studies have mainly dealt with the words (nouns, adjectives, and verbs) that directly refer to real-world matters. In addition to these words, the current study deals with logic words, such as “not,” “and,” and “or” simultaneously. These words are not directly referring to the real world, but are logical operators that contribute to the construction of meaning in sentences. In human–robot communication, these words may be used often. The current study builds a recurrent neural network model with long short-term memory units and trains it to learn to translate sentences including logic words into robot actions. We investigate what kind of compositional representations, which mediate sentences and robot actions, emerge as the network's internal states via the learning process. Analysis after learning shows that referential words are merged with visual information and the robot's own current state, and the logical words are represented by the model in accordance with their functions as logical operators. Words such as “true,” “false,” and “not” work as non-linear transformations to encode orthogonal phrases into the same area in a memory cell state space. The word “and,” which required a robot to lift up both its hands, worked as if it was a universal quantifier. The word “or,” which required action generation that looked apparently random, was represented as an unstable space of the network's dynamical system.
format Online
Article
Text
id pubmed-5744442
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-57444422018-01-08 Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions Yamada, Tatsuro Murata, Shingo Arie, Hiroaki Ogata, Tetsuya Front Neurorobot Neuroscience An important characteristic of human language is compositionality. We can efficiently express a wide variety of real-world situations, events, and behaviors by compositionally constructing the meaning of a complex expression from a finite number of elements. Previous studies have analyzed how machine-learning models, particularly neural networks, can learn from experience to represent compositional relationships between language and robot actions with the aim of understanding the symbol grounding structure and achieving intelligent communicative agents. Such studies have mainly dealt with the words (nouns, adjectives, and verbs) that directly refer to real-world matters. In addition to these words, the current study deals with logic words, such as “not,” “and,” and “or” simultaneously. These words are not directly referring to the real world, but are logical operators that contribute to the construction of meaning in sentences. In human–robot communication, these words may be used often. The current study builds a recurrent neural network model with long short-term memory units and trains it to learn to translate sentences including logic words into robot actions. We investigate what kind of compositional representations, which mediate sentences and robot actions, emerge as the network's internal states via the learning process. Analysis after learning shows that referential words are merged with visual information and the robot's own current state, and the logical words are represented by the model in accordance with their functions as logical operators. Words such as “true,” “false,” and “not” work as non-linear transformations to encode orthogonal phrases into the same area in a memory cell state space. The word “and,” which required a robot to lift up both its hands, worked as if it was a universal quantifier. The word “or,” which required action generation that looked apparently random, was represented as an unstable space of the network's dynamical system. Frontiers Media S.A. 2017-12-22 /pmc/articles/PMC5744442/ /pubmed/29311891 http://dx.doi.org/10.3389/fnbot.2017.00070 Text en Copyright © 2017 Yamada, Murata, Arie and Ogata. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Yamada, Tatsuro
Murata, Shingo
Arie, Hiroaki
Ogata, Tetsuya
Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions
title Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions
title_full Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions
title_fullStr Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions
title_full_unstemmed Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions
title_short Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions
title_sort representation learning of logic words by an rnn: from word sequences to robot actions
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5744442/
https://www.ncbi.nlm.nih.gov/pubmed/29311891
http://dx.doi.org/10.3389/fnbot.2017.00070
work_keys_str_mv AT yamadatatsuro representationlearningoflogicwordsbyanrnnfromwordsequencestorobotactions
AT muratashingo representationlearningoflogicwordsbyanrnnfromwordsequencestorobotactions
AT ariehiroaki representationlearningoflogicwordsbyanrnnfromwordsequencestorobotactions
AT ogatatetsuya representationlearningoflogicwordsbyanrnnfromwordsequencestorobotactions