LEPOR: A Robust Evaluation Metric for Machine Translation with Augmented Factors

Lifeng Han, Derek F. Wong, Lidia S. Chao

Research output: Chapter in Book/Conference proceedingConference contributionpeer-review

Abstract

In the conventional evaluation metrics of machine translation, considering less information about the translations usually makes the result not reasonable and low correlation with human judgments. On the other hand, using many external linguistic resources and tools (e.g. Part-ofspeech tagging, morpheme, stemming, and synonyms) makes the metrics complicated, timeconsuming and not universal due to that different languages have the different linguistic features. This paper proposes a novel evaluation metric employing rich and augmented factors without relying on any additional resource or tool. Experiments show that this novel metric yields the state-of-the-art correlation with human judgments compared with classic metrics BLEU, TER, Meteor-1.3 and two latest metrics (AMBER and MP4IBM1), which proves it a robust one by employing a feature-rich and model-independent approach.
Original languageEnglish
Title of host publicationProceedings of the 24th International Conference on Computational Linguistics (COLING 2012)
Place of PublicationMumbai, India
PublisherThe COLING 2012 Organizing Committee
Pages441-450
Number of pages10
Publication statusPublished - 2012

Keywords

  • Machine translation
  • evaluation metric
  • context-dependent n-gram alignment
  • modified length penalty
  • precision
  • recall

Fingerprint

Dive into the research topics of 'LEPOR: A Robust Evaluation Metric for Machine Translation with Augmented Factors'. Together they form a unique fingerprint.

Cite this