System outputs ready to download | May 10th, 2019 |
Start of manual evaluation period | May 15th, 2019 |
Paper submission deadline | May 17th, 2019 (indeed earlier than the final submission of your scores) |
Submission deadline for metrics task | May 25th, 2019 (AoE) |
End of manual evaluation | May 27th, 2019 (or longer if struggling for confidence) |
Notification of acceptance | June 7th, 2019 |
Camera-ready deadline | June 17th, 2019 |
Conference in Florence | August 1st—2nd, 2019 |
This shared task will examine automatic evaluation metrics for machine translation. We will provide you with all of the translations produced in the translation task along with the human reference translations. You will return your automatic metric scores for translations at the system-level and/or at the sentence-level. We will calculate the system-level and sentence-level correlations of your scores with WMT19 human judgements once the manual evaluation has been completed.
The goals of the shared metrics task are:
Submissions to this year's metrics task should include in each submission:
Since 2016, the system-level evaluation includes evaluation of metrics against large sets of references (10k synthetic, "hybrid" MT systems). If your system-level metric is not a simple arithmetic average of segment-level scores and it is not terribly computationally expensive, please provide also your scores for the 10k hybrid MT systems.
This year, there are no explicitly labelled additional domain but the German-French pair is targeted at EU elections and the "Test suites" part of newstest2019 does contain various domains for some language pairs.
We will provide you with the output of machine translation systems and reference translations for the following language pairs:
Additionally, Quality Estimation Task 3 "QE as a metric" runs joinly with the metrics task. For "QE as a metric", you need to provide the same outputs as standard metrics participants (see below) but you must not make use of the references.
You will compute scores for each of the outputs at the system-level and/or the sentence-level. If your automatic metric does not produce sentence-level scores, you can participate in just the system-level ranking. If your automatic metric uses linguistic annotation and supports only some language pairs, you are free to assign scores only where you can.
We will assess automatic evaluation metrics in the following ways:
System-level correlation: We will use absolute Pearson correlation coefficient to measure the correlation of the automatic metric scores with official human scores as computed in the translation task. Direct Assessment will be the official human evaluation, see last year's results for further details.
Sentence-level correlation: "Direct Assessment" will use the Pearson correlation of your scores with human judgements of translation quality. (Fallback to Kendall's tau on "relative ranking" implied from direct assessments may be necessary for some language pairs, as done in 2017.)
Document-level correlation (TBC): For some language pairs, document-level evaluation may be also available. If your metric produces document-level scores which are not a simple arithmetic average of your segment-level scores, please get in touch with us.
The following table summarizes the planned evaluation methods and text domains of each evaluation track.
Track | Text Domain | Level | Golden Truth Source |
---|---|---|---|
DAsys | news, from WMT19 news task | system-level | direct assessment |
DAseg | news, from WMT19 news task | segment-level | direct assessment |
If you participate in the metrics task, we ask you to commit about 8 hours of time to do the manual evaluation. You are also invited to submit a paper describing your metric.
The evaluation will be done with an online tool, details will be posted here.
You are invited to submit a short paper (4 to 6 pages) describing your automatic evaluation metric. You are not required to submit a paper if you do not want to. If you don't, we ask that you give an appropriate reference describing your metric that we can cite in the overview paper.
WMT19 metrics task test sets are ready, apologies for the delay.
There are three subsets of outputs that we would like you to evaluate:
The package of inputs for you to evaluate this year comes in three versions (the directory layout and naming convention this year match the main translation task):
Here is a bash script that you may want to run around your scorer to process everything:
wget http://ufallab.ms.mff.cuni.cz/~bojar/wmt19/wmt19-submitted-data-v3-txt.tgz tar xzf wmt19-submitted-data-v3-txt.tgz cd wmt19-submitted-data-v3/txt-ts for testset in `ls -d system-outputs/* | cut -d/ -f2`; do for lp in `ls -d system-outputs/$testset/* | cut -d/ -f3`; do echo "PROCESSING TESTSET $testset, LANGUAGE_PAIR $lp" ref=references/$testset-${lp:0:2}${lp:3:5}-ref-ts.${lp:3:5} src=sources/$testset-${lp:0:2}${lp:3:5}-src-ts.${lp:0:2} echo " REF: $ref SRC: $src" for hyp in system-outputs/$testset/$lp/*; do echo " EVALUATING $hyp" <YOUR EVALUATION TOOL> --reference=$ref --hypothesis=$hyp --source=$src done done done
You may want to use some of the following data to tune or train your metric.
For system-level, see the results from the previous years:
For segment-level, the following datasets are available:
Each dataset contains:
Although RR is no longer the manual evaluation employed in the metrics task, human judgments from the previous year's data sets may still prove useful:
You can use any past year's data to tune your metric's free parameters if it has any for this year's submission. Additionally, you can use any past data as a test set to compare the performance of your metric against published results from past years metric participants.
Last year's data contains all of the system's translations, the source documents and human reference translations and the human judgments of the translation quality.
The output of your software should produce scores for the translations either at the system-level or the segment-level (or preferably both).
Since we assume that your metrics are mostly simple arithmetic averages of segment-level scores, your system-level outputs serve primarily as a sanity check if we get the exact same averages.
The output files for system-level rankings should be called YOURMETRIC.sys.score.gz
and formatted in the following way:
<METRIC NAME> <LANG-PAIR> <TEST SET> <SYSTEM> <SYSTEM LEVEL SCORE> <ENSEMBLE> <AVAILABLE>Where:
METRIC NAME
is the name of your automatic evaluation metric.LANG-PAIR
is the language pair using two letter abbreviations for the languages (de-en
for German-English, for example).
TEST SET
is the ID of the test set optionally including the evaluation track (DAsys+newstest2019
for example).SYSTEM
is the ID of system being scored (given by the part of the filename for the plain text file, uedin-syntax.3866
for example).SYSTEM LEVEL SCORE
is the overall system level score that your metric is predicting.
ENSEMBLE
information about whether or not your metric employs any other existing metric or not (ensemble
if yes, non-ensemble
if not).
AVAILABLE
public availability information for your metric (the appropriate url, https://github.com/jhclark/multeval
for example or no
if if it's not available yet.
(This year, we no longer collect the timing information.)
The output files for segment-level rankings should be called YOURMETRIC.seg.score.gz
and formatted in the following way:
<METRIC NAME> <LANG-PAIR> <TEST SET> <SYSTEM> <SEGMENT NUMBER> <SEGMENT SCORE> <ENSEMBLE> <AVAILABLE>Where:
METRIC NAME
is the name of your automatic evaluation metric.LANG-PAIR
is the language pair using two letter abbreviations for the languages (de-en
for German-English, for example).
TEST SET
is the ID of the test set optionally including the evaluation track (DAsegNews+newstest2019
for example).SYSTEM
is the ID of system being scored (given by the part of the filename for the plain text file, uedin-syntax.3866
for example).SEGMENT NUMBER
is the line number starting from 1 of the plain text input files.SEGMENT SCORE
is the score your metric predicts for the particular segment.ENSEMBLE
information about whether or not your metric employs any other existing metric or not (ensemble
if yes, non-ensemble
if not).
AVAILABLE
public availability information for your metric (the appropriate url, https://github.com/jhclark/multeval
for example or no
if if it's not available yet.
Note: fields ENSEMBLE
and AVAILABLE
should be filled with the same value in every line of
the submission file for a given metric. Inclusion in this format involves some redundancy but avoids adding extra files
to the submission requirements.
Submissions should be sent as an e-mail to wmt-metrics-submissions@googlegroups.com.
As a sanity check, please enter yourself to this shared spreadsheet.
In case the above e-mail doesn't work for you (Google seems to prevent non-member postings despite we set it so), please contact us directly.