The taraXÜ corpus of human-annotated machine translations
Eleftherios Avramidis, Aljoscha Burchardt, Sabine Hunsicker, Maja Popović, Cindy Tscherwinka, David Vilar, Hans Uszkoreit
Abstract
Human translators are the key to evaluating machine translation (MT) quality and also to addressing the so far unanswered question when and how to use MT in professional translation workflows. This paper describes the corpus developed as a result of a detailed large scale human evaluation consisting of three tightly connected tasks: ranking, error classification and post-editing.- Anthology ID:
- L14-1347
- Volume:
- Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)
- Month:
- May
- Year:
- 2014
- Address:
- Reykjavik, Iceland
- Venue:
- LREC
- SIG:
- Publisher:
- European Language Resources Association (ELRA)
- Note:
- Pages:
- 2679–2682
- Language:
- URL:
- http://www.lrec-conf.org/proceedings/lrec2014/pdf/401_Paper.pdf
- DOI:
- PDF:
- http://www.lrec-conf.org/proceedings/lrec2014/pdf/401_Paper.pdf