Cargando…
Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction
Similar to specific natural language instructions, intention-related natural language queries also play an essential role in our daily life communication. Inspired by the psychology term “affordance” and its applications in Human-Robot interaction, we propose an object affordance-based natural langu...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7238763/ https://www.ncbi.nlm.nih.gov/pubmed/32477091 http://dx.doi.org/10.3389/fnbot.2020.00026 |
_version_ | 1783536593802362880 |
---|---|
author | Mi, Jinpeng Liang, Hongzhuo Katsakis, Nikolaos Tang, Song Li, Qingdu Zhang, Changshui Zhang, Jianwei |
author_facet | Mi, Jinpeng Liang, Hongzhuo Katsakis, Nikolaos Tang, Song Li, Qingdu Zhang, Changshui Zhang, Jianwei |
author_sort | Mi, Jinpeng |
collection | PubMed |
description | Similar to specific natural language instructions, intention-related natural language queries also play an essential role in our daily life communication. Inspired by the psychology term “affordance” and its applications in Human-Robot interaction, we propose an object affordance-based natural language visual grounding architecture to ground intention-related natural language queries. Formally, we first present an attention-based multi-visual features fusion network to detect object affordances from RGB images. While fusing deep visual features extracted from a pre-trained CNN model with deep texture features encoded by a deep texture encoding network, the presented object affordance detection network takes into account the interaction of the multi-visual features, and reserves the complementary nature of the different features by integrating attention weights learned from sparse representations of the multi-visual features. We train and validate the attention-based object affordance recognition network on a self-built dataset in which a large number of images originate from MSCOCO and ImageNet. Moreover, we introduce an intention semantic extraction module to extract intention semantics from intention-related natural language queries. Finally, we ground intention-related natural language queries by integrating the detected object affordances with the extracted intention semantics. We conduct extensive experiments to validate the performance of the object affordance detection network and the intention-related natural language queries grounding architecture. |
format | Online Article Text |
id | pubmed-7238763 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-72387632020-05-29 Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction Mi, Jinpeng Liang, Hongzhuo Katsakis, Nikolaos Tang, Song Li, Qingdu Zhang, Changshui Zhang, Jianwei Front Neurorobot Neuroscience Similar to specific natural language instructions, intention-related natural language queries also play an essential role in our daily life communication. Inspired by the psychology term “affordance” and its applications in Human-Robot interaction, we propose an object affordance-based natural language visual grounding architecture to ground intention-related natural language queries. Formally, we first present an attention-based multi-visual features fusion network to detect object affordances from RGB images. While fusing deep visual features extracted from a pre-trained CNN model with deep texture features encoded by a deep texture encoding network, the presented object affordance detection network takes into account the interaction of the multi-visual features, and reserves the complementary nature of the different features by integrating attention weights learned from sparse representations of the multi-visual features. We train and validate the attention-based object affordance recognition network on a self-built dataset in which a large number of images originate from MSCOCO and ImageNet. Moreover, we introduce an intention semantic extraction module to extract intention semantics from intention-related natural language queries. Finally, we ground intention-related natural language queries by integrating the detected object affordances with the extracted intention semantics. We conduct extensive experiments to validate the performance of the object affordance detection network and the intention-related natural language queries grounding architecture. Frontiers Media S.A. 2020-05-13 /pmc/articles/PMC7238763/ /pubmed/32477091 http://dx.doi.org/10.3389/fnbot.2020.00026 Text en Copyright © 2020 Mi, Liang, Katsakis, Tang, Li, Zhang and Zhang. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Mi, Jinpeng Liang, Hongzhuo Katsakis, Nikolaos Tang, Song Li, Qingdu Zhang, Changshui Zhang, Jianwei Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction |
title | Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction |
title_full | Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction |
title_fullStr | Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction |
title_full_unstemmed | Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction |
title_short | Intention-Related Natural Language Grounding via Object Affordance Detection and Intention Semantic Extraction |
title_sort | intention-related natural language grounding via object affordance detection and intention semantic extraction |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7238763/ https://www.ncbi.nlm.nih.gov/pubmed/32477091 http://dx.doi.org/10.3389/fnbot.2020.00026 |
work_keys_str_mv | AT mijinpeng intentionrelatednaturallanguagegroundingviaobjectaffordancedetectionandintentionsemanticextraction AT lianghongzhuo intentionrelatednaturallanguagegroundingviaobjectaffordancedetectionandintentionsemanticextraction AT katsakisnikolaos intentionrelatednaturallanguagegroundingviaobjectaffordancedetectionandintentionsemanticextraction AT tangsong intentionrelatednaturallanguagegroundingviaobjectaffordancedetectionandintentionsemanticextraction AT liqingdu intentionrelatednaturallanguagegroundingviaobjectaffordancedetectionandintentionsemanticextraction AT zhangchangshui intentionrelatednaturallanguagegroundingviaobjectaffordancedetectionandintentionsemanticextraction AT zhangjianwei intentionrelatednaturallanguagegroundingviaobjectaffordancedetectionandintentionsemanticextraction |