Cargando…

Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort

Translation Environment Tools make translators’ work easier by providing them with term lists, translation memories and machine translation output. Ideally, such tools automatically predict whether it is more effortful to post-edit than to translate from scratch, and determine whether or not to prov...

Descripción completa

Detalles Bibliográficos
Autores principales: Daems, Joke, Vandepitte, Sonia, Hartsuiker, Robert J., Macken, Lieve
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5539081/
https://www.ncbi.nlm.nih.gov/pubmed/28824482
http://dx.doi.org/10.3389/fpsyg.2017.01282
_version_ 1783254412435652608
author Daems, Joke
Vandepitte, Sonia
Hartsuiker, Robert J.
Macken, Lieve
author_facet Daems, Joke
Vandepitte, Sonia
Hartsuiker, Robert J.
Macken, Lieve
author_sort Daems, Joke
collection PubMed
description Translation Environment Tools make translators’ work easier by providing them with term lists, translation memories and machine translation output. Ideally, such tools automatically predict whether it is more effortful to post-edit than to translate from scratch, and determine whether or not to provide translators with machine translation output. Current machine translation quality estimation systems heavily rely on automatic metrics, even though they do not accurately capture actual post-editing effort. In addition, these systems do not take translator experience into account, even though novices’ translation processes are different from those of professional translators. In this paper, we report on the impact of machine translation errors on various types of post-editing effort indicators, for professional translators as well as student translators. We compare the impact of MT quality on a product effort indicator (HTER) with that on various process effort indicators. The translation and post-editing process of student translators and professional translators was logged with a combination of keystroke logging and eye-tracking, and the MT output was analyzed with a fine-grained translation quality assessment approach. We find that most post-editing effort indicators (product as well as process) are influenced by machine translation quality, but that different error types affect different post-editing effort indicators, confirming that a more fine-grained MT quality analysis is needed to correctly estimate actual post-editing effort. Coherence, meaning shifts, and structural issues are shown to be good indicators of post-editing effort. The additional impact of experience on these interactions between MT quality and post-editing effort is smaller than expected.
format Online
Article
Text
id pubmed-5539081
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-55390812017-08-18 Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort Daems, Joke Vandepitte, Sonia Hartsuiker, Robert J. Macken, Lieve Front Psychol Psychology Translation Environment Tools make translators’ work easier by providing them with term lists, translation memories and machine translation output. Ideally, such tools automatically predict whether it is more effortful to post-edit than to translate from scratch, and determine whether or not to provide translators with machine translation output. Current machine translation quality estimation systems heavily rely on automatic metrics, even though they do not accurately capture actual post-editing effort. In addition, these systems do not take translator experience into account, even though novices’ translation processes are different from those of professional translators. In this paper, we report on the impact of machine translation errors on various types of post-editing effort indicators, for professional translators as well as student translators. We compare the impact of MT quality on a product effort indicator (HTER) with that on various process effort indicators. The translation and post-editing process of student translators and professional translators was logged with a combination of keystroke logging and eye-tracking, and the MT output was analyzed with a fine-grained translation quality assessment approach. We find that most post-editing effort indicators (product as well as process) are influenced by machine translation quality, but that different error types affect different post-editing effort indicators, confirming that a more fine-grained MT quality analysis is needed to correctly estimate actual post-editing effort. Coherence, meaning shifts, and structural issues are shown to be good indicators of post-editing effort. The additional impact of experience on these interactions between MT quality and post-editing effort is smaller than expected. Frontiers Media S.A. 2017-08-02 /pmc/articles/PMC5539081/ /pubmed/28824482 http://dx.doi.org/10.3389/fpsyg.2017.01282 Text en Copyright © 2017 Daems, Vandepitte, Hartsuiker and Macken. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Psychology
Daems, Joke
Vandepitte, Sonia
Hartsuiker, Robert J.
Macken, Lieve
Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort
title Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort
title_full Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort
title_fullStr Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort
title_full_unstemmed Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort
title_short Identifying the Machine Translation Error Types with the Greatest Impact on Post-editing Effort
title_sort identifying the machine translation error types with the greatest impact on post-editing effort
topic Psychology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5539081/
https://www.ncbi.nlm.nih.gov/pubmed/28824482
http://dx.doi.org/10.3389/fpsyg.2017.01282
work_keys_str_mv AT daemsjoke identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort
AT vandepittesonia identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort
AT hartsuikerrobertj identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort
AT mackenlieve identifyingthemachinetranslationerrortypeswiththegreatestimpactonposteditingeffort