Cargando…

Affordance embeddings for situated language understanding

Much progress in AI over the last decade has been driven by advances in natural language processing technology, in turn facilitated by large datasets and increased computation power used to train large neural language models. These systems demonstrate apparently sophisticated linguistic understandin...

Descripción completa

Detalles Bibliográficos
Autores principales: Krishnaswamy, Nikhil, Pustejovsky, James
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9538673/
https://www.ncbi.nlm.nih.gov/pubmed/36213167
http://dx.doi.org/10.3389/frai.2022.774752
_version_ 1784803380274659328
author Krishnaswamy, Nikhil
Pustejovsky, James
author_facet Krishnaswamy, Nikhil
Pustejovsky, James
author_sort Krishnaswamy, Nikhil
collection PubMed
description Much progress in AI over the last decade has been driven by advances in natural language processing technology, in turn facilitated by large datasets and increased computation power used to train large neural language models. These systems demonstrate apparently sophisticated linguistic understanding or generation capabilities, but often fail to transfer their skills to situations they have not encountered before. We argue that computational situated grounding of linguistic information to real or simulated scenarios provide a solution to some of these learning challenges by creating situational representations that both serve as a formal model of the salient phenomena, and contain rich amounts of exploitable, task-appropriate data for training new, flexible computational models. We approach this problem from a neurosymbolic perspective, using multimodal contextual modeling of interactive situations, events, and object properties, particularly afforded behaviors, and habitats, the situations that condition them. These properties are tightly coupled to processes of situated grounding, and herein we discuss we combine neural and symbolic methods with multimodal simulations to create a platform, VoxWorld, for modeling communication in context, and we demonstrate how neural embedding vectors of symbolically-encoded object affordances facilitate transferring knowledge of objects and situations to novel entities, and learning how to recognize and generate linguistic and gestural denotations.
format Online
Article
Text
id pubmed-9538673
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-95386732022-10-08 Affordance embeddings for situated language understanding Krishnaswamy, Nikhil Pustejovsky, James Front Artif Intell Artificial Intelligence Much progress in AI over the last decade has been driven by advances in natural language processing technology, in turn facilitated by large datasets and increased computation power used to train large neural language models. These systems demonstrate apparently sophisticated linguistic understanding or generation capabilities, but often fail to transfer their skills to situations they have not encountered before. We argue that computational situated grounding of linguistic information to real or simulated scenarios provide a solution to some of these learning challenges by creating situational representations that both serve as a formal model of the salient phenomena, and contain rich amounts of exploitable, task-appropriate data for training new, flexible computational models. We approach this problem from a neurosymbolic perspective, using multimodal contextual modeling of interactive situations, events, and object properties, particularly afforded behaviors, and habitats, the situations that condition them. These properties are tightly coupled to processes of situated grounding, and herein we discuss we combine neural and symbolic methods with multimodal simulations to create a platform, VoxWorld, for modeling communication in context, and we demonstrate how neural embedding vectors of symbolically-encoded object affordances facilitate transferring knowledge of objects and situations to novel entities, and learning how to recognize and generate linguistic and gestural denotations. Frontiers Media S.A. 2022-09-23 /pmc/articles/PMC9538673/ /pubmed/36213167 http://dx.doi.org/10.3389/frai.2022.774752 Text en Copyright © 2022 Krishnaswamy and Pustejovsky. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Krishnaswamy, Nikhil
Pustejovsky, James
Affordance embeddings for situated language understanding
title Affordance embeddings for situated language understanding
title_full Affordance embeddings for situated language understanding
title_fullStr Affordance embeddings for situated language understanding
title_full_unstemmed Affordance embeddings for situated language understanding
title_short Affordance embeddings for situated language understanding
title_sort affordance embeddings for situated language understanding
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9538673/
https://www.ncbi.nlm.nih.gov/pubmed/36213167
http://dx.doi.org/10.3389/frai.2022.774752
work_keys_str_mv AT krishnaswamynikhil affordanceembeddingsforsituatedlanguageunderstanding
AT pustejovskyjames affordanceembeddingsforsituatedlanguageunderstanding