Cargando…

Multi-modal self-adaptation during object recognition in an artificial cognitive system

The cognitive connection between the senses of touch and vision is probably the best-known case of multimodality. Recent discoveries suggest that the mapping between both senses is learned rather than innate. This evidence opens the door to a dynamic multimodality that allows individuals to adaptive...

Descripción completa

Detalles Bibliográficos
Autores principales: Miralles, David, Garrofé, Guillem, Parés, Carlota, González, Alejandro, Serra, Gerard, Soto, Alberto, Sevillano, Xavier, de Beeck, Hans Op, Masson, Haemy Lee
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8904602/
https://www.ncbi.nlm.nih.gov/pubmed/35260603
http://dx.doi.org/10.1038/s41598-022-07424-9
_version_ 1784664990372855808
author Miralles, David
Garrofé, Guillem
Parés, Carlota
González, Alejandro
Serra, Gerard
Soto, Alberto
Sevillano, Xavier
de Beeck, Hans Op
Masson, Haemy Lee
author_facet Miralles, David
Garrofé, Guillem
Parés, Carlota
González, Alejandro
Serra, Gerard
Soto, Alberto
Sevillano, Xavier
de Beeck, Hans Op
Masson, Haemy Lee
author_sort Miralles, David
collection PubMed
description The cognitive connection between the senses of touch and vision is probably the best-known case of multimodality. Recent discoveries suggest that the mapping between both senses is learned rather than innate. This evidence opens the door to a dynamic multimodality that allows individuals to adaptively develop within their environment. By mimicking this aspect of human learning, we propose a new multimodal mechanism that allows artificial cognitive systems (ACS) to quickly adapt to unforeseen perceptual anomalies generated by the environment or by the system itself. In this context, visual recognition systems have advanced remarkably in recent years thanks to the creation of large-scale datasets together with the advent of deep learning algorithms. However, this has not been the case for the haptic modality, where the lack of two-handed dexterous datasets has limited the ability of learning systems to process the tactile information of human object exploration. This data imbalance hinders the creation of synchronized datasets that would enable the development of multimodality in ACS during object exploration. In this work, we use a multimodal dataset recently generated from tactile sensors placed on a collection of objects that capture haptic data from human manipulation, together with the corresponding visual counterpart. Using this data, we create a multimodal learning transfer mechanism capable of both detecting sudden and permanent anomalies in the visual channel and maintaining visual object recognition performance by retraining the visual mode for a few minutes using haptic information. Our proposal for perceptual awareness and self-adaptation is of noteworthy relevance as can be applied by any system that satisfies two very generic conditions: it can classify each mode independently and is provided with a synchronized multimodal data set.
format Online
Article
Text
id pubmed-8904602
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-89046022022-03-09 Multi-modal self-adaptation during object recognition in an artificial cognitive system Miralles, David Garrofé, Guillem Parés, Carlota González, Alejandro Serra, Gerard Soto, Alberto Sevillano, Xavier de Beeck, Hans Op Masson, Haemy Lee Sci Rep Article The cognitive connection between the senses of touch and vision is probably the best-known case of multimodality. Recent discoveries suggest that the mapping between both senses is learned rather than innate. This evidence opens the door to a dynamic multimodality that allows individuals to adaptively develop within their environment. By mimicking this aspect of human learning, we propose a new multimodal mechanism that allows artificial cognitive systems (ACS) to quickly adapt to unforeseen perceptual anomalies generated by the environment or by the system itself. In this context, visual recognition systems have advanced remarkably in recent years thanks to the creation of large-scale datasets together with the advent of deep learning algorithms. However, this has not been the case for the haptic modality, where the lack of two-handed dexterous datasets has limited the ability of learning systems to process the tactile information of human object exploration. This data imbalance hinders the creation of synchronized datasets that would enable the development of multimodality in ACS during object exploration. In this work, we use a multimodal dataset recently generated from tactile sensors placed on a collection of objects that capture haptic data from human manipulation, together with the corresponding visual counterpart. Using this data, we create a multimodal learning transfer mechanism capable of both detecting sudden and permanent anomalies in the visual channel and maintaining visual object recognition performance by retraining the visual mode for a few minutes using haptic information. Our proposal for perceptual awareness and self-adaptation is of noteworthy relevance as can be applied by any system that satisfies two very generic conditions: it can classify each mode independently and is provided with a synchronized multimodal data set. Nature Publishing Group UK 2022-03-08 /pmc/articles/PMC8904602/ /pubmed/35260603 http://dx.doi.org/10.1038/s41598-022-07424-9 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Miralles, David
Garrofé, Guillem
Parés, Carlota
González, Alejandro
Serra, Gerard
Soto, Alberto
Sevillano, Xavier
de Beeck, Hans Op
Masson, Haemy Lee
Multi-modal self-adaptation during object recognition in an artificial cognitive system
title Multi-modal self-adaptation during object recognition in an artificial cognitive system
title_full Multi-modal self-adaptation during object recognition in an artificial cognitive system
title_fullStr Multi-modal self-adaptation during object recognition in an artificial cognitive system
title_full_unstemmed Multi-modal self-adaptation during object recognition in an artificial cognitive system
title_short Multi-modal self-adaptation during object recognition in an artificial cognitive system
title_sort multi-modal self-adaptation during object recognition in an artificial cognitive system
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8904602/
https://www.ncbi.nlm.nih.gov/pubmed/35260603
http://dx.doi.org/10.1038/s41598-022-07424-9
work_keys_str_mv AT mirallesdavid multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT garrofeguillem multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT parescarlota multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT gonzalezalejandro multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT serragerard multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT sotoalberto multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT sevillanoxavier multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT debeeckhansop multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem
AT massonhaemylee multimodalselfadaptationduringobjectrecognitioninanartificialcognitivesystem