Rights accessRestricted access - publisher's policy
Assessing the quality of candidate translations involves diverse linguistic
facets. However, most automatic evaluation methods in use today rely on limited
quality assumptions, such as lexical similarity. This introduces a bias in the development cycle which in some cases has been reported to carry very negative consequences.
In order to tackle this methodological problem, we explore a novel path towards heterogeneous automatic Machine Translation evaluation. We have compiled a rich set of specialized similarity measures operating at different linguistic dimensions and analyzed their individual and collective behaviour over a wide range of evaluation scenarios. Results show that measures based on syntactic and semantic information are able to provide more reliable system rankings than lexical measures, especially when the systems under evaluation are based on different paradigms. At the sentence level, while some linguistic measures perform better than most lexical measures, some others perform substantially worse, mainly due to parsing problems.
Their scores are, however, suitable for combination, yielding a substantially improved evaluation quality.
CitationGimenez, J.; Màrquez, L. Linguistic measures for automatic machine translation evaluation. "Machine translation", Desembre 2010, vol. 24, núm. 3-4, p. 209-240.
All rights reserved. This work is protected by the corresponding intellectual and industrial property rights. Without prejudice to any existing legal exemptions, reproduction, distribution, public communication or transformation of this work are prohibited without permission of the copyright holder. If you wish to make any use of the work not provided for in the law, please contact: firstname.lastname@example.org