Cargando…
Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort
Translation Environment Tools make translators’ work easier by providing them with term lists, translation memories and machine translation output. Ideally, such tools automatically predict whether it is more effortful to post-edit than to translate from scratch, and determine whether or not to prov...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5539081/ https://www.ncbi.nlm.nih.gov/pubmed/28824482 http://dx.doi.org/10.3389/fpsyg.2017.01282 |
_version_ | 1783254412435652608 |
---|---|
author | Daems, Joke Vandepitte, Sonia Hartsuiker, Robert J. Macken, Lieve |
author_facet | Daems, Joke Vandepitte, Sonia Hartsuiker, Robert J. Macken, Lieve |
author_sort | Daems, Joke |
collection | PubMed |
description | Translation Environment Tools make translators’ work easier by providing them with term lists, translation memories and machine translation output. Ideally, such tools automatically predict whether it is more effortful to post-edit than to translate from scratch, and determine whether or not to provide translators with machine translation output. Current machine translation quality estimation systems heavily rely on automatic metrics, even though they do not accurately capture actual post-editing effort. In addition, these systems do not take translator experience into account, even though novices’ translation processes are different from those of professional translators. In this paper, we report on the impact of machine translation errors on various types of post-editing effort indicators, for professional translators as well as student translators. We compare the impact of MT quality on a product effort indicator (HTER) with that on various process effort indicators. The translation and post-editing process of student translators and professional translators was logged with a combination of keystroke logging and eye-tracking, and the MT output was analyzed with a fine-grained translation quality assessment approach. We find that most post-editing effort indicators (product as well as process) are influenced by machine translation quality, but that different error types affect different post-editing effort indicators, confirming that a more fine-grained MT quality analysis is needed to correctly estimate actual post-editing effort. Coherence, meaning shifts, and structural issues are shown to be good indicators of post-editing effort. The additional impact of experience on these interactions between MT quality and post-editing effort is smaller than expected. |
format | Online Article Text |
id | pubmed-5539081 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-55390812017-08-18 Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort Daems, Joke Vandepitte, Sonia Hartsuiker, Robert J. Macken, Lieve Front Psychol Psychology Translation Environment Tools make translators’ work easier by providing them with term lists, translation memories and machine translation output. Ideally, such tools automatically predict whether it is more effortful to post-edit than to translate from scratch, and determine whether or not to provide translators with machine translation output. Current machine translation quality estimation systems heavily rely on automatic metrics, even though they do not accurately capture actual post-editing effort. In addition, these systems do not take translator experience into account, even though novices’ translation processes are different from those of professional translators. In this paper, we report on the impact of machine translation errors on various types of post-editing effort indicators, for professional translators as well as student translators. We compare the impact of MT quality on a product effort indicator (HTER) with that on various process effort indicators. The translation and post-editing process of student translators and professional translators was logged with a combination of keystroke logging and eye-tracking, and the MT output was analyzed with a fine-grained translation quality assessment approach. We find that most post-editing effort indicators (product as well as process) are influenced by machine translation quality, but that different error types affect different post-editing effort indicators, confirming that a more fine-grained MT quality analysis is needed to correctly estimate actual post-editing effort. Coherence, meaning shifts, and structural issues are shown to be good indicators of post-editing effort. The additional impact of experience on these interactions between MT quality and post-editing effort is smaller than expected. Frontiers Media S.A. 2017-08-02 /pmc/articles/PMC5539081/ /pubmed/28824482 http://dx.doi.org/10.3389/fpsyg.2017.01282 Text en Copyright © 2017 Daems, Vandepitte, Hartsuiker and Macken. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Psychology Daems, Joke Vandepitte, Sonia Hartsuiker, Robert J. Macken, Lieve Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort |
title | Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort |
title_full | Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort |
title_fullStr | Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort |
title_full_unstemmed | Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort |
title_short | Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort |
title_sort | identifying the machine translation error types with the greatest impact on post-editing effort |
topic | Psychology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5539081/ https://www.ncbi.nlm.nih.gov/pubmed/28824482 http://dx.doi.org/10.3389/fpsyg.2017.01282 |
work_keys_str_mv | AT daemsjoke identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort AT vandepittesonia identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort AT hartsuikerrobertj identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort AT mackenlieve identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort |