Cargando…
An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences
Sign language translation (SLT) is an important application to bridge the communication gap between deaf and hearing people. In recent years, the research on the SLT based on neural translation frameworks has attracted wide attention. Despite the progress, current SLT research is still in the initia...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7604584/ https://www.ncbi.nlm.nih.gov/pubmed/33163072 http://dx.doi.org/10.1155/2020/8816125 |
_version_ | 1783604169524903936 |
---|---|
author | Zheng, Jiangbin Zhao, Zheng Chen, Min Chen, Jing Wu, Chong Chen, Yidong Shi, Xiaodong Tong, Yiqi |
author_facet | Zheng, Jiangbin Zhao, Zheng Chen, Min Chen, Jing Wu, Chong Chen, Yidong Shi, Xiaodong Tong, Yiqi |
author_sort | Zheng, Jiangbin |
collection | PubMed |
description | Sign language translation (SLT) is an important application to bridge the communication gap between deaf and hearing people. In recent years, the research on the SLT based on neural translation frameworks has attracted wide attention. Despite the progress, current SLT research is still in the initial stage. In fact, current systems perform poorly in processing long sign sentences, which often involve long-distance dependencies and require large resource consumption. To tackle this problem, we propose two explainable adaptations to the traditional neural SLT models using optimized tokenization-related modules. First, we introduce a frame stream density compression (FSDC) algorithm for detecting and reducing the redundant similar frames, which effectively shortens the long sign sentences without losing information. Then, we replace the traditional encoder in a neural machine translation (NMT) module with an improved architecture, which incorporates a temporal convolution (T-Conv) unit and a dynamic hierarchical bidirectional GRU (DH-BiGRU) unit sequentially. The improved component takes the temporal tokenization information into consideration to extract deeper information with reasonable resource consumption. Our experiments on the RWTH-PHOENIX-Weather 2014T dataset show that the proposed model outperforms the state-of-the-art baseline up to about 1.5+ BLEU-4 score gains. |
format | Online Article Text |
id | pubmed-7604584 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-76045842020-11-05 An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences Zheng, Jiangbin Zhao, Zheng Chen, Min Chen, Jing Wu, Chong Chen, Yidong Shi, Xiaodong Tong, Yiqi Comput Intell Neurosci Research Article Sign language translation (SLT) is an important application to bridge the communication gap between deaf and hearing people. In recent years, the research on the SLT based on neural translation frameworks has attracted wide attention. Despite the progress, current SLT research is still in the initial stage. In fact, current systems perform poorly in processing long sign sentences, which often involve long-distance dependencies and require large resource consumption. To tackle this problem, we propose two explainable adaptations to the traditional neural SLT models using optimized tokenization-related modules. First, we introduce a frame stream density compression (FSDC) algorithm for detecting and reducing the redundant similar frames, which effectively shortens the long sign sentences without losing information. Then, we replace the traditional encoder in a neural machine translation (NMT) module with an improved architecture, which incorporates a temporal convolution (T-Conv) unit and a dynamic hierarchical bidirectional GRU (DH-BiGRU) unit sequentially. The improved component takes the temporal tokenization information into consideration to extract deeper information with reasonable resource consumption. Our experiments on the RWTH-PHOENIX-Weather 2014T dataset show that the proposed model outperforms the state-of-the-art baseline up to about 1.5+ BLEU-4 score gains. Hindawi 2020-10-23 /pmc/articles/PMC7604584/ /pubmed/33163072 http://dx.doi.org/10.1155/2020/8816125 Text en Copyright © 2020 Jiangbin Zheng et al. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Zheng, Jiangbin Zhao, Zheng Chen, Min Chen, Jing Wu, Chong Chen, Yidong Shi, Xiaodong Tong, Yiqi An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences |
title | An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences |
title_full | An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences |
title_fullStr | An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences |
title_full_unstemmed | An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences |
title_short | An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences |
title_sort | improved sign language translation model with explainable adaptations for processing long sign sentences |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7604584/ https://www.ncbi.nlm.nih.gov/pubmed/33163072 http://dx.doi.org/10.1155/2020/8816125 |
work_keys_str_mv | AT zhengjiangbin animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT zhaozheng animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT chenmin animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT chenjing animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT wuchong animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT chenyidong animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT shixiaodong animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT tongyiqi animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT zhengjiangbin improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT zhaozheng improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT chenmin improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT chenjing improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT wuchong improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT chenyidong improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT shixiaodong improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences AT tongyiqi improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences |