Cargando…

Multi-Modal Song Mood Detection with Deep Learning †

The production and consumption of music in the contemporary era results in big data generation and creates new needs for automated and more effective management of these data. Automated music mood detection constitutes an active task in the field of MIR (Music Information Retrieval). The first appro...

Descripción completa

Detalles Bibliográficos
Autores principales: Pyrovolakis, Konstantinos, Tzouveli, Paraskevi, Stamou, Giorgos
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8838547/
https://www.ncbi.nlm.nih.gov/pubmed/35161804
http://dx.doi.org/10.3390/s22031065
_version_ 1784650154240901120
author Pyrovolakis, Konstantinos
Tzouveli, Paraskevi
Stamou, Giorgos
author_facet Pyrovolakis, Konstantinos
Tzouveli, Paraskevi
Stamou, Giorgos
author_sort Pyrovolakis, Konstantinos
collection PubMed
description The production and consumption of music in the contemporary era results in big data generation and creates new needs for automated and more effective management of these data. Automated music mood detection constitutes an active task in the field of MIR (Music Information Retrieval). The first approach to correlating music and mood was made in 1990 by Gordon Burner who researched the way that musical emotion affects marketing. In 2016, Lidy and Schiner trained a CNN for the task of genre and mood classification based on audio. In 2018, Delbouys et al. developed a multi-modal Deep Learning system combining CNN and LSTM architectures and concluded that multi-modal approaches overcome single channel models. This work will examine and compare single channel and multi-modal approaches for the task of music mood detection applying Deep Learning architectures. Our first approach tries to utilize the audio signal and the lyrics of a musical track separately, while the second approach applies a uniform multi-modal analysis to classify the given data into mood classes. The available data we will use to train and evaluate our models comes from the MoodyLyrics dataset, which includes 2000 song titles with labels from four mood classes, {happy, angry, sad, relaxed}. The result of this work leads to a uniform prediction of the mood that represents a music track and has usage in many applications.
format Online
Article
Text
id pubmed-8838547
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-88385472022-02-13 Multi-Modal Song Mood Detection with Deep Learning † Pyrovolakis, Konstantinos Tzouveli, Paraskevi Stamou, Giorgos Sensors (Basel) Article The production and consumption of music in the contemporary era results in big data generation and creates new needs for automated and more effective management of these data. Automated music mood detection constitutes an active task in the field of MIR (Music Information Retrieval). The first approach to correlating music and mood was made in 1990 by Gordon Burner who researched the way that musical emotion affects marketing. In 2016, Lidy and Schiner trained a CNN for the task of genre and mood classification based on audio. In 2018, Delbouys et al. developed a multi-modal Deep Learning system combining CNN and LSTM architectures and concluded that multi-modal approaches overcome single channel models. This work will examine and compare single channel and multi-modal approaches for the task of music mood detection applying Deep Learning architectures. Our first approach tries to utilize the audio signal and the lyrics of a musical track separately, while the second approach applies a uniform multi-modal analysis to classify the given data into mood classes. The available data we will use to train and evaluate our models comes from the MoodyLyrics dataset, which includes 2000 song titles with labels from four mood classes, {happy, angry, sad, relaxed}. The result of this work leads to a uniform prediction of the mood that represents a music track and has usage in many applications. MDPI 2022-01-29 /pmc/articles/PMC8838547/ /pubmed/35161804 http://dx.doi.org/10.3390/s22031065 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Pyrovolakis, Konstantinos
Tzouveli, Paraskevi
Stamou, Giorgos
Multi-Modal Song Mood Detection with Deep Learning †
title Multi-Modal Song Mood Detection with Deep Learning †
title_full Multi-Modal Song Mood Detection with Deep Learning †
title_fullStr Multi-Modal Song Mood Detection with Deep Learning †
title_full_unstemmed Multi-Modal Song Mood Detection with Deep Learning †
title_short Multi-Modal Song Mood Detection with Deep Learning †
title_sort multi-modal song mood detection with deep learning †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8838547/
https://www.ncbi.nlm.nih.gov/pubmed/35161804
http://dx.doi.org/10.3390/s22031065
work_keys_str_mv AT pyrovolakiskonstantinos multimodalsongmooddetectionwithdeeplearning
AT tzouveliparaskevi multimodalsongmooddetectionwithdeeplearning
AT stamougiorgos multimodalsongmooddetectionwithdeeplearning