Cargando…

“Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships

The development of computational artifacts to study cross-modal associations has been a growing research topic, as they allow new degrees of abstraction. In this context, we propose a novel approach to the computational exploration of relationships between music and abstract images, grounded by find...

Descripción completa

Detalles Bibliográficos
Autores principales: Rodrigues, Ana, Sousa, Bruna, Cardoso, Amílcar, Machado, Penousal
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777780/
https://www.ncbi.nlm.nih.gov/pubmed/36554111
http://dx.doi.org/10.3390/e24121706
_version_ 1784856191037341696
author Rodrigues, Ana
Sousa, Bruna
Cardoso, Amílcar
Machado, Penousal
author_facet Rodrigues, Ana
Sousa, Bruna
Cardoso, Amílcar
Machado, Penousal
author_sort Rodrigues, Ana
collection PubMed
description The development of computational artifacts to study cross-modal associations has been a growing research topic, as they allow new degrees of abstraction. In this context, we propose a novel approach to the computational exploration of relationships between music and abstract images, grounded by findings from cognitive sciences (emotion and perception). Due to the problem’s high-level nature, we rely on evolutionary programming techniques to evolve this audio–visual dialogue. To articulate the complexity of the problem, we develop a framework with four modules: (i) vocabulary set, (ii) music generator, (iii) image generator, and (iv) evolutionary engine. We test our approach by evolving a given music set to a corresponding set of images, steered by the expression of four emotions (angry, calm, happy, sad). Then, we perform preliminary user tests to evaluate if the user’s perception is consistent with the system’s expression. Results suggest an agreement between the user’s emotional perception of the music–image pairs and the system outcomes, favoring the integration of cognitive science knowledge. We also discuss the benefit of employing evolutionary strategies, such as genetic programming on multi-modal problems of a creative nature. Overall, this research contributes to a better understanding of the foundations of auditory–visual associations mediated by emotions and perception.
format Online
Article
Text
id pubmed-9777780
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-97777802022-12-23 “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships Rodrigues, Ana Sousa, Bruna Cardoso, Amílcar Machado, Penousal Entropy (Basel) Article The development of computational artifacts to study cross-modal associations has been a growing research topic, as they allow new degrees of abstraction. In this context, we propose a novel approach to the computational exploration of relationships between music and abstract images, grounded by findings from cognitive sciences (emotion and perception). Due to the problem’s high-level nature, we rely on evolutionary programming techniques to evolve this audio–visual dialogue. To articulate the complexity of the problem, we develop a framework with four modules: (i) vocabulary set, (ii) music generator, (iii) image generator, and (iv) evolutionary engine. We test our approach by evolving a given music set to a corresponding set of images, steered by the expression of four emotions (angry, calm, happy, sad). Then, we perform preliminary user tests to evaluate if the user’s perception is consistent with the system’s expression. Results suggest an agreement between the user’s emotional perception of the music–image pairs and the system outcomes, favoring the integration of cognitive science knowledge. We also discuss the benefit of employing evolutionary strategies, such as genetic programming on multi-modal problems of a creative nature. Overall, this research contributes to a better understanding of the foundations of auditory–visual associations mediated by emotions and perception. MDPI 2022-11-22 /pmc/articles/PMC9777780/ /pubmed/36554111 http://dx.doi.org/10.3390/e24121706 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Rodrigues, Ana
Sousa, Bruna
Cardoso, Amílcar
Machado, Penousal
“Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships
title “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships
title_full “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships
title_fullStr “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships
title_full_unstemmed “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships
title_short “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships
title_sort “found in translation”: an evolutionary framework for auditory–visual relationships
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777780/
https://www.ncbi.nlm.nih.gov/pubmed/36554111
http://dx.doi.org/10.3390/e24121706
work_keys_str_mv AT rodriguesana foundintranslationanevolutionaryframeworkforauditoryvisualrelationships
AT sousabruna foundintranslationanevolutionaryframeworkforauditoryvisualrelationships
AT cardosoamilcar foundintranslationanevolutionaryframeworkforauditoryvisualrelationships
AT machadopenousal foundintranslationanevolutionaryframeworkforauditoryvisualrelationships