Cargando…
Vowel speech recognition from rat electroencephalography using long short-term memory neural network
Over the years, considerable research has been conducted to investigate the mechanisms of speech perception and recognition. Electroencephalography (EEG) is a powerful tool for identifying brain activity; therefore, it has been widely used to determine the neural basis of speech recognition. In part...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9223328/ https://www.ncbi.nlm.nih.gov/pubmed/35737731 http://dx.doi.org/10.1371/journal.pone.0270405 |
_version_ | 1784733099504959488 |
---|---|
author | Ham, Jinsil Yoo, Hyun-Joon Kim, Jongin Lee, Boreom |
author_facet | Ham, Jinsil Yoo, Hyun-Joon Kim, Jongin Lee, Boreom |
author_sort | Ham, Jinsil |
collection | PubMed |
description | Over the years, considerable research has been conducted to investigate the mechanisms of speech perception and recognition. Electroencephalography (EEG) is a powerful tool for identifying brain activity; therefore, it has been widely used to determine the neural basis of speech recognition. In particular, for the classification of speech recognition, deep learning-based approaches are in the spotlight because they can automatically learn and extract representative features through end-to-end learning. This study aimed to identify particular components that are potentially related to phoneme representation in the rat brain and to discriminate brain activity for each vowel stimulus on a single-trial basis using a bidirectional long short-term memory (BiLSTM) network and classical machine learning methods. Nineteen male Sprague-Dawley rats subjected to microelectrode implantation surgery to record EEG signals from the bilateral anterior auditory fields were used. Five different vowel speech stimuli were chosen, /a/, /e/, /i/, /o/, and /u/, which have highly different formant frequencies. EEG recorded under randomly given vowel stimuli was minimally preprocessed and normalized by a z-score transformation to be used as input for the classification of speech recognition. The BiLSTM network showed the best performance among the classifiers by achieving an overall accuracy, f1-score, and Cohen’s κ values of 75.18%, 0.75, and 0.68, respectively, using a 10-fold cross-validation approach. These results indicate that LSTM layers can effectively model sequential data, such as EEG; hence, informative features can be derived through BiLSTM trained with end-to-end learning without any additional hand-crafted feature extraction methods. |
format | Online Article Text |
id | pubmed-9223328 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-92233282022-06-24 Vowel speech recognition from rat electroencephalography using long short-term memory neural network Ham, Jinsil Yoo, Hyun-Joon Kim, Jongin Lee, Boreom PLoS One Research Article Over the years, considerable research has been conducted to investigate the mechanisms of speech perception and recognition. Electroencephalography (EEG) is a powerful tool for identifying brain activity; therefore, it has been widely used to determine the neural basis of speech recognition. In particular, for the classification of speech recognition, deep learning-based approaches are in the spotlight because they can automatically learn and extract representative features through end-to-end learning. This study aimed to identify particular components that are potentially related to phoneme representation in the rat brain and to discriminate brain activity for each vowel stimulus on a single-trial basis using a bidirectional long short-term memory (BiLSTM) network and classical machine learning methods. Nineteen male Sprague-Dawley rats subjected to microelectrode implantation surgery to record EEG signals from the bilateral anterior auditory fields were used. Five different vowel speech stimuli were chosen, /a/, /e/, /i/, /o/, and /u/, which have highly different formant frequencies. EEG recorded under randomly given vowel stimuli was minimally preprocessed and normalized by a z-score transformation to be used as input for the classification of speech recognition. The BiLSTM network showed the best performance among the classifiers by achieving an overall accuracy, f1-score, and Cohen’s κ values of 75.18%, 0.75, and 0.68, respectively, using a 10-fold cross-validation approach. These results indicate that LSTM layers can effectively model sequential data, such as EEG; hence, informative features can be derived through BiLSTM trained with end-to-end learning without any additional hand-crafted feature extraction methods. Public Library of Science 2022-06-23 /pmc/articles/PMC9223328/ /pubmed/35737731 http://dx.doi.org/10.1371/journal.pone.0270405 Text en © 2022 Ham et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Ham, Jinsil Yoo, Hyun-Joon Kim, Jongin Lee, Boreom Vowel speech recognition from rat electroencephalography using long short-term memory neural network |
title | Vowel speech recognition from rat electroencephalography using long short-term memory neural network |
title_full | Vowel speech recognition from rat electroencephalography using long short-term memory neural network |
title_fullStr | Vowel speech recognition from rat electroencephalography using long short-term memory neural network |
title_full_unstemmed | Vowel speech recognition from rat electroencephalography using long short-term memory neural network |
title_short | Vowel speech recognition from rat electroencephalography using long short-term memory neural network |
title_sort | vowel speech recognition from rat electroencephalography using long short-term memory neural network |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9223328/ https://www.ncbi.nlm.nih.gov/pubmed/35737731 http://dx.doi.org/10.1371/journal.pone.0270405 |
work_keys_str_mv | AT hamjinsil vowelspeechrecognitionfromratelectroencephalographyusinglongshorttermmemoryneuralnetwork AT yoohyunjoon vowelspeechrecognitionfromratelectroencephalographyusinglongshorttermmemoryneuralnetwork AT kimjongin vowelspeechrecognitionfromratelectroencephalographyusinglongshorttermmemoryneuralnetwork AT leeboreom vowelspeechrecognitionfromratelectroencephalographyusinglongshorttermmemoryneuralnetwork |