Cargando…

Imagined Speech Classification Using EEG and Deep Learning

In this paper, we propose an imagined speech-based brain wave pattern recognition using deep learning. Multiple features were extracted concurrently from eight-channel electroencephalography (EEG) signals. To obtain classifiable EEG data with fewer sensors, we placed the EEG sensors on carefully sel...

Descripción completa

Detalles Bibliográficos
Autores principales: Abdulghani, Mokhles M., Walters, Wilbur L., Abed, Khalid H.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10295790/
https://www.ncbi.nlm.nih.gov/pubmed/37370580
http://dx.doi.org/10.3390/bioengineering10060649
_version_ 1785063505087430656
author Abdulghani, Mokhles M.
Walters, Wilbur L.
Abed, Khalid H.
author_facet Abdulghani, Mokhles M.
Walters, Wilbur L.
Abed, Khalid H.
author_sort Abdulghani, Mokhles M.
collection PubMed
description In this paper, we propose an imagined speech-based brain wave pattern recognition using deep learning. Multiple features were extracted concurrently from eight-channel electroencephalography (EEG) signals. To obtain classifiable EEG data with fewer sensors, we placed the EEG sensors on carefully selected spots on the scalp. To decrease the dimensions and complexity of the EEG dataset and to avoid overfitting during the deep learning algorithm, we utilized the wavelet scattering transformation. A low-cost 8-channel EEG headset was used with MATLAB 2023a to acquire the EEG data. The long-short term memory recurrent neural network (LSTM-RNN) was used to decode the identified EEG signals into four audio commands: up, down, left, and right. Wavelet scattering transformation was applied to extract the most stable features by passing the EEG dataset through a series of filtration processes. Filtration was implemented for each individual command in the EEG datasets. The proposed imagined speech-based brain wave pattern recognition approach achieved a 92.50% overall classification accuracy. This accuracy is promising for designing a trustworthy imagined speech-based brain–computer interface (BCI) future real-time systems. For better evaluation of the classification performance, other metrics were considered, and we obtained 92.74%, 92.50%, and 92.62% for precision, recall, and F1-score, respectively.
format Online
Article
Text
id pubmed-10295790
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102957902023-06-28 Imagined Speech Classification Using EEG and Deep Learning Abdulghani, Mokhles M. Walters, Wilbur L. Abed, Khalid H. Bioengineering (Basel) Article In this paper, we propose an imagined speech-based brain wave pattern recognition using deep learning. Multiple features were extracted concurrently from eight-channel electroencephalography (EEG) signals. To obtain classifiable EEG data with fewer sensors, we placed the EEG sensors on carefully selected spots on the scalp. To decrease the dimensions and complexity of the EEG dataset and to avoid overfitting during the deep learning algorithm, we utilized the wavelet scattering transformation. A low-cost 8-channel EEG headset was used with MATLAB 2023a to acquire the EEG data. The long-short term memory recurrent neural network (LSTM-RNN) was used to decode the identified EEG signals into four audio commands: up, down, left, and right. Wavelet scattering transformation was applied to extract the most stable features by passing the EEG dataset through a series of filtration processes. Filtration was implemented for each individual command in the EEG datasets. The proposed imagined speech-based brain wave pattern recognition approach achieved a 92.50% overall classification accuracy. This accuracy is promising for designing a trustworthy imagined speech-based brain–computer interface (BCI) future real-time systems. For better evaluation of the classification performance, other metrics were considered, and we obtained 92.74%, 92.50%, and 92.62% for precision, recall, and F1-score, respectively. MDPI 2023-05-26 /pmc/articles/PMC10295790/ /pubmed/37370580 http://dx.doi.org/10.3390/bioengineering10060649 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Abdulghani, Mokhles M.
Walters, Wilbur L.
Abed, Khalid H.
Imagined Speech Classification Using EEG and Deep Learning
title Imagined Speech Classification Using EEG and Deep Learning
title_full Imagined Speech Classification Using EEG and Deep Learning
title_fullStr Imagined Speech Classification Using EEG and Deep Learning
title_full_unstemmed Imagined Speech Classification Using EEG and Deep Learning
title_short Imagined Speech Classification Using EEG and Deep Learning
title_sort imagined speech classification using eeg and deep learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10295790/
https://www.ncbi.nlm.nih.gov/pubmed/37370580
http://dx.doi.org/10.3390/bioengineering10060649
work_keys_str_mv AT abdulghanimokhlesm imaginedspeechclassificationusingeeganddeeplearning
AT walterswilburl imaginedspeechclassificationusingeeganddeeplearning
AT abedkhalidh imaginedspeechclassificationusingeeganddeeplearning