Sciweavers

2008 search results - page 4 / 402
» MUC-4 evaluation metrics
Sort
View
ACL
2009
13 years 5 months ago
Correlating Human and Automatic Evaluation of a German Surface Realiser
We examine correlations between native speaker judgements on automatically generated German text against automatic evaluation metrics. We look at a number of metrics from the MT a...
Aoife Cahill
ACL
2009
13 years 5 months ago
The Contribution of Linguistic Features to Automatic Machine Translation Evaluation
A number of approaches to Automatic MT Evaluation based on deep linguistic knowledge have been suggested. However, n-gram based metrics are still today the dominant approach. The ...
Enrique Amigó, Jesús Giménez,...
ACL
2009
13 years 5 months ago
Robust Machine Translation Evaluation with Entailment Features
Existing evaluation metrics for machine translation lack crucial robustness: their correlations with human quality judgments vary considerably across languages and genres. We beli...
Sebastian Padó, Michel Galley, Daniel Juraf...
NAACL
2010
13 years 5 months ago
Extending the METEOR Machine Translation Evaluation Metric to the Phrase Level
This paper presents METEOR-NEXT, an extended version of the METEOR metric designed to have high correlation with postediting measures of machine translation quality. We describe c...
Michael J. Denkowski, Alon Lavie
EMNLP
2010
13 years 5 months ago
Automatic Evaluation of Translation Quality for Distant Language Pairs
Automatic evaluation of Machine Translation (MT) quality is essential to developing highquality MT systems. Various evaluation metrics have been proposed, and BLEU is now used as ...
Hideki Isozaki, Tsutomu Hirao, Kevin Duh, Katsuhit...