Linguistic measures for automatic machine translation evaluation
Rights accessRestricted access - publisher's policy
European Commisision's projectFAUST - Feedback Analysis for User adaptive Statistical Translation (EC-FP7-247762)
Assessing the quality of candidate translations involves diverse linguistic facets. However, most automatic evaluation methods in use today rely on limited quality assumptions, such as lexical similarity. This introduces a bias in the development cycle which in some cases has been reported to carry very negative consequences. In order to tackle this methodological problem, we explore a novel path towards heterogeneous automatic Machine Translation evaluation. We have compiled a rich set of specialized similarity measures operating at different linguistic dimensions and analyzed their individual and collective behaviour over a wide range of evaluation scenarios. Results show that measures based on syntactic and semantic information are able to provide more reliable system rankings than lexical measures, especially when the systems under evaluation are based on different paradigms. At the sentence level, while some linguistic measures perform better than most lexical measures, some others perform substantially worse, mainly due to parsing problems. Their scores are, however, suitable for combination, yielding a substantially improved evaluation quality.
CitationGimenez, J.; Màrquez, L. Linguistic measures for automatic machine translation evaluation. "Machine translation", Desembre 2010, vol. 24, núm. 3-4, p. 209-240.
|Linguistic meas ... translation evaluation.pdf||347.8Kb||Restricted access|