Cargando…
“Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships
The development of computational artifacts to study cross-modal associations has been a growing research topic, as they allow new degrees of abstraction. In this context, we propose a novel approach to the computational exploration of relationships between music and abstract images, grounded by find...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777780/ https://www.ncbi.nlm.nih.gov/pubmed/36554111 http://dx.doi.org/10.3390/e24121706 |
_version_ | 1784856191037341696 |
---|---|
author | Rodrigues, Ana Sousa, Bruna Cardoso, Amílcar Machado, Penousal |
author_facet | Rodrigues, Ana Sousa, Bruna Cardoso, Amílcar Machado, Penousal |
author_sort | Rodrigues, Ana |
collection | PubMed |
description | The development of computational artifacts to study cross-modal associations has been a growing research topic, as they allow new degrees of abstraction. In this context, we propose a novel approach to the computational exploration of relationships between music and abstract images, grounded by findings from cognitive sciences (emotion and perception). Due to the problem’s high-level nature, we rely on evolutionary programming techniques to evolve this audio–visual dialogue. To articulate the complexity of the problem, we develop a framework with four modules: (i) vocabulary set, (ii) music generator, (iii) image generator, and (iv) evolutionary engine. We test our approach by evolving a given music set to a corresponding set of images, steered by the expression of four emotions (angry, calm, happy, sad). Then, we perform preliminary user tests to evaluate if the user’s perception is consistent with the system’s expression. Results suggest an agreement between the user’s emotional perception of the music–image pairs and the system outcomes, favoring the integration of cognitive science knowledge. We also discuss the benefit of employing evolutionary strategies, such as genetic programming on multi-modal problems of a creative nature. Overall, this research contributes to a better understanding of the foundations of auditory–visual associations mediated by emotions and perception. |
format | Online Article Text |
id | pubmed-9777780 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-97777802022-12-23 “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships Rodrigues, Ana Sousa, Bruna Cardoso, Amílcar Machado, Penousal Entropy (Basel) Article The development of computational artifacts to study cross-modal associations has been a growing research topic, as they allow new degrees of abstraction. In this context, we propose a novel approach to the computational exploration of relationships between music and abstract images, grounded by findings from cognitive sciences (emotion and perception). Due to the problem’s high-level nature, we rely on evolutionary programming techniques to evolve this audio–visual dialogue. To articulate the complexity of the problem, we develop a framework with four modules: (i) vocabulary set, (ii) music generator, (iii) image generator, and (iv) evolutionary engine. We test our approach by evolving a given music set to a corresponding set of images, steered by the expression of four emotions (angry, calm, happy, sad). Then, we perform preliminary user tests to evaluate if the user’s perception is consistent with the system’s expression. Results suggest an agreement between the user’s emotional perception of the music–image pairs and the system outcomes, favoring the integration of cognitive science knowledge. We also discuss the benefit of employing evolutionary strategies, such as genetic programming on multi-modal problems of a creative nature. Overall, this research contributes to a better understanding of the foundations of auditory–visual associations mediated by emotions and perception. MDPI 2022-11-22 /pmc/articles/PMC9777780/ /pubmed/36554111 http://dx.doi.org/10.3390/e24121706 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Rodrigues, Ana Sousa, Bruna Cardoso, Amílcar Machado, Penousal “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships |
title | “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships |
title_full | “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships |
title_fullStr | “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships |
title_full_unstemmed | “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships |
title_short | “Found in Translation”: An Evolutionary Framework for Auditory–Visual Relationships |
title_sort | “found in translation”: an evolutionary framework for auditory–visual relationships |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777780/ https://www.ncbi.nlm.nih.gov/pubmed/36554111 http://dx.doi.org/10.3390/e24121706 |
work_keys_str_mv | AT rodriguesana foundintranslationanevolutionaryframeworkforauditoryvisualrelationships AT sousabruna foundintranslationanevolutionaryframeworkforauditoryvisualrelationships AT cardosoamilcar foundintranslationanevolutionaryframeworkforauditoryvisualrelationships AT machadopenousal foundintranslationanevolutionaryframeworkforauditoryvisualrelationships |