Cargando…

Sign and Human Action Detection Using Deep Learning

Human beings usually rely on communication to express their feeling and ideas and to solve disputes among themselves. A major component required for effective communication is language. Language can occur in different forms, including written symbols, gestures, and vocalizations. It is usually essen...

Descripción completa

Detalles Bibliográficos
Autores principales: Dhulipala, Shivanarayna, Adedoyin, Festus Fatai, Bruno, Alessandro
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9316580/
https://www.ncbi.nlm.nih.gov/pubmed/35877636
http://dx.doi.org/10.3390/jimaging8070192
_version_ 1784754849980612608
author Dhulipala, Shivanarayna
Adedoyin, Festus Fatai
Bruno, Alessandro
author_facet Dhulipala, Shivanarayna
Adedoyin, Festus Fatai
Bruno, Alessandro
author_sort Dhulipala, Shivanarayna
collection PubMed
description Human beings usually rely on communication to express their feeling and ideas and to solve disputes among themselves. A major component required for effective communication is language. Language can occur in different forms, including written symbols, gestures, and vocalizations. It is usually essential for all of the communicating parties to be fully conversant with a common language. However, to date this has not been the case between speech-impaired people who use sign language and people who use spoken languages. A number of different studies have pointed out a significant gaps between these two groups which can limit the ease of communication. Therefore, this study aims to develop an efficient deep learning model that can be used to predict British sign language in an attempt to narrow this communication gap between speech-impaired and non-speech-impaired people in the community. Two models were developed in this research, CNN and LSTM, and their performance was evaluated using a multi-class confusion matrix. The CNN model emerged with the highest performance, attaining training and testing accuracies of 98.8% and 97.4%, respectively. In addition, the model achieved average weighted precession and recall of 97% and 96%, respectively. On the other hand, the LSTM model’s performance was quite poor, with the maximum training and testing performance accuracies achieved being 49.4% and 48.7%, respectively. Our research concluded that the CNN model was the best for recognizing and determining British sign language.
format Online
Article
Text
id pubmed-9316580
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-93165802022-07-27 Sign and Human Action Detection Using Deep Learning Dhulipala, Shivanarayna Adedoyin, Festus Fatai Bruno, Alessandro J Imaging Article Human beings usually rely on communication to express their feeling and ideas and to solve disputes among themselves. A major component required for effective communication is language. Language can occur in different forms, including written symbols, gestures, and vocalizations. It is usually essential for all of the communicating parties to be fully conversant with a common language. However, to date this has not been the case between speech-impaired people who use sign language and people who use spoken languages. A number of different studies have pointed out a significant gaps between these two groups which can limit the ease of communication. Therefore, this study aims to develop an efficient deep learning model that can be used to predict British sign language in an attempt to narrow this communication gap between speech-impaired and non-speech-impaired people in the community. Two models were developed in this research, CNN and LSTM, and their performance was evaluated using a multi-class confusion matrix. The CNN model emerged with the highest performance, attaining training and testing accuracies of 98.8% and 97.4%, respectively. In addition, the model achieved average weighted precession and recall of 97% and 96%, respectively. On the other hand, the LSTM model’s performance was quite poor, with the maximum training and testing performance accuracies achieved being 49.4% and 48.7%, respectively. Our research concluded that the CNN model was the best for recognizing and determining British sign language. MDPI 2022-07-11 /pmc/articles/PMC9316580/ /pubmed/35877636 http://dx.doi.org/10.3390/jimaging8070192 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Dhulipala, Shivanarayna
Adedoyin, Festus Fatai
Bruno, Alessandro
Sign and Human Action Detection Using Deep Learning
title Sign and Human Action Detection Using Deep Learning
title_full Sign and Human Action Detection Using Deep Learning
title_fullStr Sign and Human Action Detection Using Deep Learning
title_full_unstemmed Sign and Human Action Detection Using Deep Learning
title_short Sign and Human Action Detection Using Deep Learning
title_sort sign and human action detection using deep learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9316580/
https://www.ncbi.nlm.nih.gov/pubmed/35877636
http://dx.doi.org/10.3390/jimaging8070192
work_keys_str_mv AT dhulipalashivanarayna signandhumanactiondetectionusingdeeplearning
AT adedoyinfestusfatai signandhumanactiondetectionusingdeeplearning
AT brunoalessandro signandhumanactiondetectionusingdeeplearning