Machine translation for institutional academic texts: Output quality, terminology translation and post-editor trust

Scansani, Randy (2020) Machine translation for institutional academic texts: Output quality, terminology translation and post-editor trust, [Dissertation thesis], Alma Mater Studiorum Università di Bologna. Dottorato di ricerca in Traduzione, interpretazione e interculturalità, 32 Ciclo. DOI 10.6092/unibo/amsdottorato/9191.
Documenti full-text disponibili:
[img] Documento PDF (English) - Richiede un lettore di PDF come Xpdf o Adobe Acrobat Reader
Disponibile con Licenza: Creative Commons Attribution Non-commercial ShareAlike 4.0 (CC BY-NC-SA 4.0) .
Download (1MB)

Abstract

The present work is a feasibility study on the application of Machine Translation (MT) to institutional academic texts, specifically course catalogues, for Italian-English and German-English. The first research question of this work focuses on the feasibility of profitably applying MT to such texts. Since the benefits of a good quality MT might be counteracted by preconceptions of translators towards the output, the second research question examines translator trainees' trust towards an MT output as compared to a human translation (HT). Training and test sets are created for both language combinations in the institutional academic domain. MT systems used are ModernMT and Google Translate. Overall evaluations of the output quality are carried out using automatic metrics. Results show that applying neural MT to institutional academic texts can be beneficial even when bilingual data are not available. When small amounts of sentence pairs become available, MT quality improves. Then, a gold standard data set with manual annotations of terminology (MAGMATic) is created and used for an evaluation of the output focused on terminology translation. The gold standard was publicly released to stimulate research on terminology assessment. The assessment proves that domain-adaptation improves the quality of term translation. To conclude, a method to measure trust in a post-editing task is proposed and results regarding translator trainees trust towards MT are outlined. All participants are asked to work on the same text. Half of them is told that it is an MT output to be post-edited, and the other half that it is a HT needing revision. Results prove that there is no statistically significant difference between post-editing and HT revision in terms of number of edits and temporal effort. Results thus suggest that a new generation of translators that received training on MT and post-editing is not influenced by preconceptions against MT.

Abstract
Tipologia del documento
Tesi di dottorato
Autore
Scansani, Randy
Supervisore
Dottorato di ricerca
Ciclo
32
Coordinatore
Settore disciplinare
Settore concorsuale
Parole chiave
Machine Translation, post-editing, MT, neural machine translation, translation, terminology, course catalogues, institutional academic texts
URN:NBN
DOI
10.6092/unibo/amsdottorato/9191
Data di discussione
30 Marzo 2020
URI

Altri metadati

Statistica sui download

Gestione del documento: Visualizza la tesi

^