Cargando…

An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences

Sign language translation (SLT) is an important application to bridge the communication gap between deaf and hearing people. In recent years, the research on the SLT based on neural translation frameworks has attracted wide attention. Despite the progress, current SLT research is still in the initia...

Descripción completa

Detalles Bibliográficos
Autores principales: Zheng, Jiangbin, Zhao, Zheng, Chen, Min, Chen, Jing, Wu, Chong, Chen, Yidong, Shi, Xiaodong, Tong, Yiqi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7604584/
https://www.ncbi.nlm.nih.gov/pubmed/33163072
http://dx.doi.org/10.1155/2020/8816125
_version_ 1783604169524903936
author Zheng, Jiangbin
Zhao, Zheng
Chen, Min
Chen, Jing
Wu, Chong
Chen, Yidong
Shi, Xiaodong
Tong, Yiqi
author_facet Zheng, Jiangbin
Zhao, Zheng
Chen, Min
Chen, Jing
Wu, Chong
Chen, Yidong
Shi, Xiaodong
Tong, Yiqi
author_sort Zheng, Jiangbin
collection PubMed
description Sign language translation (SLT) is an important application to bridge the communication gap between deaf and hearing people. In recent years, the research on the SLT based on neural translation frameworks has attracted wide attention. Despite the progress, current SLT research is still in the initial stage. In fact, current systems perform poorly in processing long sign sentences, which often involve long-distance dependencies and require large resource consumption. To tackle this problem, we propose two explainable adaptations to the traditional neural SLT models using optimized tokenization-related modules. First, we introduce a frame stream density compression (FSDC) algorithm for detecting and reducing the redundant similar frames, which effectively shortens the long sign sentences without losing information. Then, we replace the traditional encoder in a neural machine translation (NMT) module with an improved architecture, which incorporates a temporal convolution (T-Conv) unit and a dynamic hierarchical bidirectional GRU (DH-BiGRU) unit sequentially. The improved component takes the temporal tokenization information into consideration to extract deeper information with reasonable resource consumption. Our experiments on the RWTH-PHOENIX-Weather 2014T dataset show that the proposed model outperforms the state-of-the-art baseline up to about 1.5+ BLEU-4 score gains.
format Online
Article
Text
id pubmed-7604584
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-76045842020-11-05 An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences Zheng, Jiangbin Zhao, Zheng Chen, Min Chen, Jing Wu, Chong Chen, Yidong Shi, Xiaodong Tong, Yiqi Comput Intell Neurosci Research Article Sign language translation (SLT) is an important application to bridge the communication gap between deaf and hearing people. In recent years, the research on the SLT based on neural translation frameworks has attracted wide attention. Despite the progress, current SLT research is still in the initial stage. In fact, current systems perform poorly in processing long sign sentences, which often involve long-distance dependencies and require large resource consumption. To tackle this problem, we propose two explainable adaptations to the traditional neural SLT models using optimized tokenization-related modules. First, we introduce a frame stream density compression (FSDC) algorithm for detecting and reducing the redundant similar frames, which effectively shortens the long sign sentences without losing information. Then, we replace the traditional encoder in a neural machine translation (NMT) module with an improved architecture, which incorporates a temporal convolution (T-Conv) unit and a dynamic hierarchical bidirectional GRU (DH-BiGRU) unit sequentially. The improved component takes the temporal tokenization information into consideration to extract deeper information with reasonable resource consumption. Our experiments on the RWTH-PHOENIX-Weather 2014T dataset show that the proposed model outperforms the state-of-the-art baseline up to about 1.5+ BLEU-4 score gains. Hindawi 2020-10-23 /pmc/articles/PMC7604584/ /pubmed/33163072 http://dx.doi.org/10.1155/2020/8816125 Text en Copyright © 2020 Jiangbin Zheng et al. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Zheng, Jiangbin
Zhao, Zheng
Chen, Min
Chen, Jing
Wu, Chong
Chen, Yidong
Shi, Xiaodong
Tong, Yiqi
An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences
title An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences
title_full An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences
title_fullStr An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences
title_full_unstemmed An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences
title_short An Improved Sign Language Translation Model with Explainable Adaptations for Processing Long Sign Sentences
title_sort improved sign language translation model with explainable adaptations for processing long sign sentences
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7604584/
https://www.ncbi.nlm.nih.gov/pubmed/33163072
http://dx.doi.org/10.1155/2020/8816125
work_keys_str_mv AT zhengjiangbin animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT zhaozheng animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT chenmin animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT chenjing animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT wuchong animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT chenyidong animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT shixiaodong animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT tongyiqi animprovedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT zhengjiangbin improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT zhaozheng improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT chenmin improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT chenjing improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT wuchong improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT chenyidong improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT shixiaodong improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences
AT tongyiqi improvedsignlanguagetranslationmodelwithexplainableadaptationsforprocessinglongsignsentences