Cargando…
Recognition of Non-Manual Content in Continuous Japanese Sign Language
The quality of recognition systems for continuous utterances in signed languages could be largely advanced within the last years. However, research efforts often do not address specific linguistic features of signed languages, as e.g., non-manual expressions. In this work, we evaluate the potential...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7582855/ https://www.ncbi.nlm.nih.gov/pubmed/33019608 http://dx.doi.org/10.3390/s20195621 |
_version_ | 1783599286720659456 |
---|---|
author | Brock, Heike Farag, Iva Nakadai, Kazuhiro |
author_facet | Brock, Heike Farag, Iva Nakadai, Kazuhiro |
author_sort | Brock, Heike |
collection | PubMed |
description | The quality of recognition systems for continuous utterances in signed languages could be largely advanced within the last years. However, research efforts often do not address specific linguistic features of signed languages, as e.g., non-manual expressions. In this work, we evaluate the potential of a single video camera-based recognition system with respect to the latter. For this, we introduce a two-stage pipeline based on two-dimensional body joint positions extracted from RGB camera data. The system first separates the data flow of a signed expression into meaningful word segments on the base of a frame-wise binary Random Forest. Next, every segment is transformed into image-like shape and classified with a Convolutional Neural Network. The proposed system is then evaluated on a data set of continuous sentence expressions in Japanese Sign Language with a variation of non-manual expressions. Exploring multiple variations of data representations and network parameters, we are able to distinguish word segments of specific non-manual intonations with 86% accuracy from the underlying body joint movement data. Full sentence predictions achieve a total Word Error Rate of 15.75%. This marks an improvement of 13.22% as compared to ground truth predictions obtained from labeling insensitive towards non-manual content. Consequently, our analysis constitutes an important contribution for a better understanding of mixed manual and non-manual content in signed communication. |
format | Online Article Text |
id | pubmed-7582855 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75828552020-10-28 Recognition of Non-Manual Content in Continuous Japanese Sign Language Brock, Heike Farag, Iva Nakadai, Kazuhiro Sensors (Basel) Article The quality of recognition systems for continuous utterances in signed languages could be largely advanced within the last years. However, research efforts often do not address specific linguistic features of signed languages, as e.g., non-manual expressions. In this work, we evaluate the potential of a single video camera-based recognition system with respect to the latter. For this, we introduce a two-stage pipeline based on two-dimensional body joint positions extracted from RGB camera data. The system first separates the data flow of a signed expression into meaningful word segments on the base of a frame-wise binary Random Forest. Next, every segment is transformed into image-like shape and classified with a Convolutional Neural Network. The proposed system is then evaluated on a data set of continuous sentence expressions in Japanese Sign Language with a variation of non-manual expressions. Exploring multiple variations of data representations and network parameters, we are able to distinguish word segments of specific non-manual intonations with 86% accuracy from the underlying body joint movement data. Full sentence predictions achieve a total Word Error Rate of 15.75%. This marks an improvement of 13.22% as compared to ground truth predictions obtained from labeling insensitive towards non-manual content. Consequently, our analysis constitutes an important contribution for a better understanding of mixed manual and non-manual content in signed communication. MDPI 2020-10-01 /pmc/articles/PMC7582855/ /pubmed/33019608 http://dx.doi.org/10.3390/s20195621 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Brock, Heike Farag, Iva Nakadai, Kazuhiro Recognition of Non-Manual Content in Continuous Japanese Sign Language |
title | Recognition of Non-Manual Content in Continuous Japanese Sign Language |
title_full | Recognition of Non-Manual Content in Continuous Japanese Sign Language |
title_fullStr | Recognition of Non-Manual Content in Continuous Japanese Sign Language |
title_full_unstemmed | Recognition of Non-Manual Content in Continuous Japanese Sign Language |
title_short | Recognition of Non-Manual Content in Continuous Japanese Sign Language |
title_sort | recognition of non-manual content in continuous japanese sign language |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7582855/ https://www.ncbi.nlm.nih.gov/pubmed/33019608 http://dx.doi.org/10.3390/s20195621 |
work_keys_str_mv | AT brockheike recognitionofnonmanualcontentincontinuousjapanesesignlanguage AT faragiva recognitionofnonmanualcontentincontinuousjapanesesignlanguage AT nakadaikazuhiro recognitionofnonmanualcontentincontinuousjapanesesignlanguage |