Cargando…

Research on Audio Recognition Based on the Deep Neural Network in Music Teaching

Solfeggio is an important basic course for music majors, and audio recognition training is one of the important links. With the improvement of computer performance, audio recognition has been widely used in smart wearable devices. In recent years, the development of deep learning has accelerated the...

Descripción completa

Detalles Bibliográficos
Autores principales: Cui, Yun, Wang, Fu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9166999/
https://www.ncbi.nlm.nih.gov/pubmed/35669668
http://dx.doi.org/10.1155/2022/7055624
_version_ 1784720734261608448
author Cui, Yun
Wang, Fu
author_facet Cui, Yun
Wang, Fu
author_sort Cui, Yun
collection PubMed
description Solfeggio is an important basic course for music majors, and audio recognition training is one of the important links. With the improvement of computer performance, audio recognition has been widely used in smart wearable devices. In recent years, the development of deep learning has accelerated the research process of audio recognition. However, there is a lot of sound interference in music teaching environment, which leads to the performance of the audio classifier that cannot meet the actual demand. In order to solve this problem, an improved audio recognition system based on YOLO-v4 is proposed, which mainly improves the network structure. First, Mel frequency cepstrum number is used to process the original audio and extract the corresponding features. Then, try to apply the YOLO-v4 model in the field of deep learning to the field of audio recognition and improve it by combining with the spatial pyramid pool module to strengthen the generalization ability of data in different audio formats. Second, the stacking method in ensemble learning is used to fuse the independent submodels of two different channels. Experimental results show that compared with other deep learning technologies, the improved YOLO-v4 model can improve the performance of audio recognition, and it has better performance in processing data of different audio formats, which shows better generalization ability.
format Online
Article
Text
id pubmed-9166999
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-91669992022-06-05 Research on Audio Recognition Based on the Deep Neural Network in Music Teaching Cui, Yun Wang, Fu Comput Intell Neurosci Research Article Solfeggio is an important basic course for music majors, and audio recognition training is one of the important links. With the improvement of computer performance, audio recognition has been widely used in smart wearable devices. In recent years, the development of deep learning has accelerated the research process of audio recognition. However, there is a lot of sound interference in music teaching environment, which leads to the performance of the audio classifier that cannot meet the actual demand. In order to solve this problem, an improved audio recognition system based on YOLO-v4 is proposed, which mainly improves the network structure. First, Mel frequency cepstrum number is used to process the original audio and extract the corresponding features. Then, try to apply the YOLO-v4 model in the field of deep learning to the field of audio recognition and improve it by combining with the spatial pyramid pool module to strengthen the generalization ability of data in different audio formats. Second, the stacking method in ensemble learning is used to fuse the independent submodels of two different channels. Experimental results show that compared with other deep learning technologies, the improved YOLO-v4 model can improve the performance of audio recognition, and it has better performance in processing data of different audio formats, which shows better generalization ability. Hindawi 2022-05-27 /pmc/articles/PMC9166999/ /pubmed/35669668 http://dx.doi.org/10.1155/2022/7055624 Text en Copyright © 2022 Yun Cui and Fu Wang. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Cui, Yun
Wang, Fu
Research on Audio Recognition Based on the Deep Neural Network in Music Teaching
title Research on Audio Recognition Based on the Deep Neural Network in Music Teaching
title_full Research on Audio Recognition Based on the Deep Neural Network in Music Teaching
title_fullStr Research on Audio Recognition Based on the Deep Neural Network in Music Teaching
title_full_unstemmed Research on Audio Recognition Based on the Deep Neural Network in Music Teaching
title_short Research on Audio Recognition Based on the Deep Neural Network in Music Teaching
title_sort research on audio recognition based on the deep neural network in music teaching
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9166999/
https://www.ncbi.nlm.nih.gov/pubmed/35669668
http://dx.doi.org/10.1155/2022/7055624
work_keys_str_mv AT cuiyun researchonaudiorecognitionbasedonthedeepneuralnetworkinmusicteaching
AT wangfu researchonaudiorecognitionbasedonthedeepneuralnetworkinmusicteaching