Towards human linguistic machine translation evaluation
View/Open
Cita com:
hdl:2117/104808
Document typeArticle
Defense date2013-12-01
Rights accessOpen Access
Except where otherwise noted, content on this work
is licensed under a Creative Commons license
:
Attribution-NonCommercial-NoDerivs 3.0 Spain
Abstract
When evaluating machine translation outputs, linguistics is usually taken into account implicitly. Annotators have to decide whether a sentence is better than another or not, using, for example, adequacy and fluency criteria or, as recently proposed, editing the translation output so that it has the same meaning as a reference translation, and it is understandable. Therefore, the important fields of linguistics of meaning (semantics) and grammar (syntax) are indirectly considered. In this study, we propose to go one step further towards a linguistic human evaluation. The idea is to introduce linguistics implicitly by formulating precise guidelines. These guidelines strictly mark the difference between the sub-fields of linguistics such as: morphology, syntax, semantics, and orthography. We show our guidelines have a high inter-annotation agreement and wide-error coverage. Additionally, we examine how the linguistic human evaluation data correlate with: among different types of machine translation systems (rule and statistical-based); and with adequacy and fluency.
CitationRuiz, M., Farrus, M. Towards human linguistic machine translation evaluation. "Literacy and linguistic computing", 1 Desembre 2013, vol. 28, núm. 4, p. 157-166.
ISSN0268-1145
Files | Description | Size | Format | View |
---|---|---|---|---|
llc-guidelines.pdf | 205,2Kb | View/Open |