Cargando…

An integrated mediapipe-optimized GRU model for Indian sign language recognition

Sign language recognition is challenged by problems, such as accurate tracking of hand gestures, occlusion of hands, and high computational cost. Recently, it has benefited from advancements in deep learning techniques. However, these larger complex approaches cannot manage long-term sequential data...

Descripción completa

Detalles Bibliográficos
Autores principales: Subramanian, Barathi, Olimov, Bekhzod, Naik, Shraddha M., Kim, Sangchul, Park, Kil-Houm, Kim, Jeonghong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9279358/
https://www.ncbi.nlm.nih.gov/pubmed/35831393
http://dx.doi.org/10.1038/s41598-022-15998-7
_version_ 1784746379187322880
author Subramanian, Barathi
Olimov, Bekhzod
Naik, Shraddha M.
Kim, Sangchul
Park, Kil-Houm
Kim, Jeonghong
author_facet Subramanian, Barathi
Olimov, Bekhzod
Naik, Shraddha M.
Kim, Sangchul
Park, Kil-Houm
Kim, Jeonghong
author_sort Subramanian, Barathi
collection PubMed
description Sign language recognition is challenged by problems, such as accurate tracking of hand gestures, occlusion of hands, and high computational cost. Recently, it has benefited from advancements in deep learning techniques. However, these larger complex approaches cannot manage long-term sequential data and they are characterized by poor information processing and learning efficiency in capturing useful information. To overcome these challenges, we propose an integrated MediaPipe-optimized gated recurrent unit (MOPGRU) model for Indian sign language recognition. Specifically, we improved the update gate of the standard GRU cell by multiplying it by the reset gate to discard the redundant information from the past in one screening. By obtaining feedback from the resultant of the reset gate, additional attention is shown to the present input. Additionally, we replace the hyperbolic tangent activation in standard GRUs with exponential linear unit activation and SoftMax with Softsign activation in the output layer of the GRU cell. Thus, our proposed MOPGRU model achieved better prediction accuracy, high learning efficiency, information processing capability, and faster convergence than other sequential models.
format Online
Article
Text
id pubmed-9279358
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-92793582022-07-15 An integrated mediapipe-optimized GRU model for Indian sign language recognition Subramanian, Barathi Olimov, Bekhzod Naik, Shraddha M. Kim, Sangchul Park, Kil-Houm Kim, Jeonghong Sci Rep Article Sign language recognition is challenged by problems, such as accurate tracking of hand gestures, occlusion of hands, and high computational cost. Recently, it has benefited from advancements in deep learning techniques. However, these larger complex approaches cannot manage long-term sequential data and they are characterized by poor information processing and learning efficiency in capturing useful information. To overcome these challenges, we propose an integrated MediaPipe-optimized gated recurrent unit (MOPGRU) model for Indian sign language recognition. Specifically, we improved the update gate of the standard GRU cell by multiplying it by the reset gate to discard the redundant information from the past in one screening. By obtaining feedback from the resultant of the reset gate, additional attention is shown to the present input. Additionally, we replace the hyperbolic tangent activation in standard GRUs with exponential linear unit activation and SoftMax with Softsign activation in the output layer of the GRU cell. Thus, our proposed MOPGRU model achieved better prediction accuracy, high learning efficiency, information processing capability, and faster convergence than other sequential models. Nature Publishing Group UK 2022-07-13 /pmc/articles/PMC9279358/ /pubmed/35831393 http://dx.doi.org/10.1038/s41598-022-15998-7 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Subramanian, Barathi
Olimov, Bekhzod
Naik, Shraddha M.
Kim, Sangchul
Park, Kil-Houm
Kim, Jeonghong
An integrated mediapipe-optimized GRU model for Indian sign language recognition
title An integrated mediapipe-optimized GRU model for Indian sign language recognition
title_full An integrated mediapipe-optimized GRU model for Indian sign language recognition
title_fullStr An integrated mediapipe-optimized GRU model for Indian sign language recognition
title_full_unstemmed An integrated mediapipe-optimized GRU model for Indian sign language recognition
title_short An integrated mediapipe-optimized GRU model for Indian sign language recognition
title_sort integrated mediapipe-optimized gru model for indian sign language recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9279358/
https://www.ncbi.nlm.nih.gov/pubmed/35831393
http://dx.doi.org/10.1038/s41598-022-15998-7
work_keys_str_mv AT subramanianbarathi anintegratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT olimovbekhzod anintegratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT naikshraddham anintegratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT kimsangchul anintegratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT parkkilhoum anintegratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT kimjeonghong anintegratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT subramanianbarathi integratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT olimovbekhzod integratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT naikshraddham integratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT kimsangchul integratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT parkkilhoum integratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition
AT kimjeonghong integratedmediapipeoptimizedgrumodelforindiansignlanguagerecognition