Cargando…
Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification
Expectation learning is a unsupervised learning process which uses multisensory bindings to enhance unisensory perception. For instance, as humans, we learn to associate a barking sound with the visual appearance of a dog, and we continuously fine-tune this association over time, as we learn, e.g.,...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7806099/ https://www.ncbi.nlm.nih.gov/pubmed/33501152 http://dx.doi.org/10.3389/frobt.2019.00137 |
_version_ | 1783636456341766144 |
---|---|
author | Barros, Pablo Eppe, Manfred Parisi, German I. Liu, Xun Wermter, Stefan |
author_facet | Barros, Pablo Eppe, Manfred Parisi, German I. Liu, Xun Wermter, Stefan |
author_sort | Barros, Pablo |
collection | PubMed |
description | Expectation learning is a unsupervised learning process which uses multisensory bindings to enhance unisensory perception. For instance, as humans, we learn to associate a barking sound with the visual appearance of a dog, and we continuously fine-tune this association over time, as we learn, e.g., to associate high-pitched barking with small dogs. In this work, we address the problem of developing a computational model that addresses important properties of expectation learning, in particular focusing on the lack of explicit external supervision other than temporal co-occurrence. To this end, we present a novel hybrid neural model based on audio-visual autoencoders and a recurrent self-organizing network for multisensory bindings that facilitate stimulus reconstructions across different sensory modalities. We refer to this mechanism as stimulus prediction across modalities and demonstrate that the proposed model is capable of learning concept bindings by evaluating it on unisensory classification tasks for audio-visual stimuli using the 43,500 Youtube videos from the animal subset of the AudioSet corpus. |
format | Online Article Text |
id | pubmed-7806099 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-78060992021-01-25 Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification Barros, Pablo Eppe, Manfred Parisi, German I. Liu, Xun Wermter, Stefan Front Robot AI Robotics and AI Expectation learning is a unsupervised learning process which uses multisensory bindings to enhance unisensory perception. For instance, as humans, we learn to associate a barking sound with the visual appearance of a dog, and we continuously fine-tune this association over time, as we learn, e.g., to associate high-pitched barking with small dogs. In this work, we address the problem of developing a computational model that addresses important properties of expectation learning, in particular focusing on the lack of explicit external supervision other than temporal co-occurrence. To this end, we present a novel hybrid neural model based on audio-visual autoencoders and a recurrent self-organizing network for multisensory bindings that facilitate stimulus reconstructions across different sensory modalities. We refer to this mechanism as stimulus prediction across modalities and demonstrate that the proposed model is capable of learning concept bindings by evaluating it on unisensory classification tasks for audio-visual stimuli using the 43,500 Youtube videos from the animal subset of the AudioSet corpus. Frontiers Media S.A. 2019-12-11 /pmc/articles/PMC7806099/ /pubmed/33501152 http://dx.doi.org/10.3389/frobt.2019.00137 Text en Copyright © 2019 Barros, Eppe, Parisi, Liu and Wermter. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Robotics and AI Barros, Pablo Eppe, Manfred Parisi, German I. Liu, Xun Wermter, Stefan Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification |
title | Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification |
title_full | Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification |
title_fullStr | Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification |
title_full_unstemmed | Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification |
title_short | Expectation Learning for Stimulus Prediction Across Modalities Improves Unisensory Classification |
title_sort | expectation learning for stimulus prediction across modalities improves unisensory classification |
topic | Robotics and AI |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7806099/ https://www.ncbi.nlm.nih.gov/pubmed/33501152 http://dx.doi.org/10.3389/frobt.2019.00137 |
work_keys_str_mv | AT barrospablo expectationlearningforstimuluspredictionacrossmodalitiesimprovesunisensoryclassification AT eppemanfred expectationlearningforstimuluspredictionacrossmodalitiesimprovesunisensoryclassification AT parisigermani expectationlearningforstimuluspredictionacrossmodalitiesimprovesunisensoryclassification AT liuxun expectationlearningforstimuluspredictionacrossmodalitiesimprovesunisensoryclassification AT wermterstefan expectationlearningforstimuluspredictionacrossmodalitiesimprovesunisensoryclassification |