• Corpus ID: 2097570

Analysing Post-Editing Performance: Correlations with Years of Translation Experience

  title={Analysing Post-Editing Performance: Correlations with Years of Translation Experience},
  author={Giselle de Almeida and Sharon O’Brien},
  booktitle={European Association for Machine Translation Conferences/Workshops},
Post-editing (PE) is still a new activity for many translators. The lack of training, clear and consistent guidelines and international standards may cause difficulties in the transition from translation to PE. Aiming to gain a better understanding of these difficulties and using data gathered from a pilot project, this paper explores possible correlations between PE performance and previous translation experience. We test a combination of the LISA QA Model and the GALE postediting guidelines… 

Tables from this paper

Enhancement of post-editing performance: introducing machine translation post-editing in translator training

The dissertation contributes to the definition of the scope of post-editors’ professional expertise, offers a scalable training model and describes to what extent such model may enhance post-EDiting performance in undergraduate translation students.

Testing post-editing guidelines: how translation trainees interpret them and how to tailor them for translator training purposes

A qualitative study that investigates how trainee translators on an MA course, aimed at preparing the trainees for the translation industry, interpret these PE guidelines for publishable quality, suggests trainees have difficulties interpreting the guidelines.

Towards acquiring post-editing abilities through research-informed practical tasks

The integration of machine translation (MT) in professional workflows, facilitated by recent gains in quality achieved by neural MT systems, means that translators are increasingly required to do

Comparison of post-editing productivity between professional translators and lay users

Results suggest that overall, post-editing increases translation throughput for both translators and users, although the latter seem to benefit more from the MT output.

Translators’ perceptions of literary post-editing using statistical and neural machine translation

In the context of recent improvements in the quality of machine translation (MT) output and new use cases being found for that output, this article reports on an experiment using statistical and

Towards predicting post-editing productivity

Correlations between general text matcher and translation edit rate and post-editing productivity and segments with high GTM and TER scores and cognitive measures of effort are investigated.

Cognitive Effort in Post-Editing of Machine Translation

This thesis presents an empirical study that examines MT post-editing by contrasting the cognitive effort required by this activity with a number of its key elements, including characteristics of the source text and of the MT output, post-edsitors’ individual traits, and the quality of the post-edited text assessed by human evaluators.

What Do Professional Translators Do when Post-Editing for the First Time? First Insight into the Spanish-Basque Language Pair

This work provides a first attempt to uncover what activity professional translators carry out when working from Spanish into Basque and reveals that when working with moderate machine translation output post-editing shifts from the task of identifying and fixing errors, to that of “patchwork” wherePost-editors identify the machine translated elements to reuse and connect them using their own contributions.

Quality and productivity : a comparative analysis of human translation and post-editing with Malay learners of Arabic and English

It was revealed that the translation technologies helped improve the non-native trainee translators’ speed and quality and the study discovered that syntactic and lexical errors are the most problematic in the PE tasks.

Teaching Computer-assisted Translation: A Proposal for Course Content

  • Marija Brkic Bakaric
  • Computer Science
    2019 42nd International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO)
  • 2019
This paper argues for the inclusion of the post-editors, pre-editing, and machine translation content into the course on computer-assisted translation and gives proposal for the course content.



Repairing Texts: Empirical Investigations of Machine Translation Post-Editing Processes

In Repairing Texts, Hans P. Krings challenges the idea that, given the effectiveness of machine translation, major costs could be reduced by using monolingual staff to post -- edit translations.

Pauses as Indicators of Cognitive Effort in Post-editing Machine Translation Output

This paper presents data from a research project which includes an analysis of pauses in post-editing, triangulated with the Choice Network Analysis method and Translog, and suggests that while pauses provide some indication of cognitive processing, supplementary methods are required to give a fuller picture.

Domain specific MT in use

This paper reports on the feedback of post-editors using fluency/adequacy evaluation and the evaluation metric 'Usability', understood in this context as where users on a three point scale evaluate the sentence from the point of view of the post-editor.

Post-editing on-screen: machine translation from Spanish into English

The end of 1986 marked seven years of ongoing machine translation (MT) production at the Pan American Health Organization (PAHO) in the Spanish-English combination. SPANAM, the Spanish-English MT

TransType2 : The Last Word

The results of the usability evaluations that were conducted within TransType2, an international R&D project the goal of which was to develop a novel approach to interactive machine translation are presented.

Interactive Assistance to Human Translators using Statistical Machine Translation Methods

Caitra is a tool that makes suggestions for sentence completion, shows word and phrase translation options, and allows postediting of machine translation output, which validates the types of assistance and provides insight into the human translation process.

Quantitative evaluation of machine translation systems: sentence level

This paper pursues a more consistent, fine-grained and comprehensive classification of possible translation errors and proposes metrics for sentence level errors, specifically lexical and syntactic errors.

Error Classification for MT Evaluation

A classification system for errors in machine translation (MT) output is presented and the results of CompuServe's use of the error classification methodology are presented as a case study.

Available from: http://www.lisa.org/ [Accessed 31

  • Available from: http://www.lisa.org/ [Accessed 31
  • 2009

The Localization Industry Standards Association. (Homepage

  • The Localization Industry Standards Association. (Homepage