The recurring translation task of the WMT workshops focuses on European language pairs. Translation quality will be evaluated on a shared, unseen test set of news stories. We provide a parallel corpus as training data, a baseline system, and additional resources for download. Participants may augment the baseline system or use their own system.
The goals of the shared translation task are:
We provide training data for five European language pairs, and a common framework (including a baseline system). The task is to improve methods current methods. This can be done in many ways. For instance participants could try to:
You may participate in any or all of the following language pairs:
We also strongly encourage your participation, if you use your own training corpus, your own sentence alignment, your own language model, or your own decoder.
If you use additional training data or existing translation systems, you must flag that your system uses additional data. We will distinguish system submissions that used the provided training data (constrained) from submissions that used significant additional data resources. Note that basic linguistic tools such as taggers, parsers, or morphological analyzers are allowed in the constrained condition.
Your submission report should highlight in which ways your own methods and data differ from the standard task. We may break down submitted results in different tracks, based on what resources were used. We are mostly interested in submission that are constraint to the provided training data, so that the comparison is focused on the methods, not on the data used. You may submit contrastive runs to demonstrate the benefit of additional training data.
The provided data is mainly taken from version 7 of the Europarl corpus, which is freely available. Please click on the links below to download the sentence-aligned data, or go to the Europarl website for the source release. Note that this the same data as last year, since Europarl is not anymore translted across all 23 official European languages.
Additional training data is taken from the new News Commentary corpus. There are about 50 million words of training data per language from the Europarl corpus and 3 million words from the News Commentary corpus.
A new data resource this year is the Common Crawl corpus which was collected from web sources. Each parallel corpus comes with a annotation file that gives the source of each sentence pair.
You may also use the following monolingual corpora released by the LDC:
Note that the released data is not tokenized and includes sentences of any length (including empty sentences). All data is in Unicode (UTF-8) format. The following tools allow the processing of the training data into tokenized format:
tokenizer.perl
detokenizer.perl
lowercase.perl
wrap-xml.perl
To evaluate your system during development, we suggest using the 2012 test set. The data is provided in raw text format and in an SGML format that suits the NIST scoring tool. We also release other test sets from previous years.
News news-test2008
2051 sentences. |
News news-test2009
|
News news-test2010
|
News news-test2011
|
News news-test2012
|
The news-test2011 set has three additional Czech translations that you may want to use. You can download them from Charles University.
File | Size | CS-EN | DE-EN | ES-EN | FR-EN | RU-EN | Notes |
---|---|---|---|---|---|---|---|
Europarl v7 | 628MB | ✓ | ✓ | ✓ | ✓ | corpus home page, same as previous year | |
876MB | ✓ | ✓ | ✓ | ✓ | ✓ | Updated 2013-02-25 to fix character encoding issues | |
UN corpus | 2.3GB | ✓ | ✓ | corpus home page | |||
News Commentary | 89MB | ✓ | ✓ | ✓ | ✓ | ✓ | |
2.3 GB | ✓ | ||||||
CzEng 1.0 | 115MB | ✓ | corpus home page | ||||
128MB | ✓ | corpus home page; lowercased | |||||
7.2MB | ✓ | provided by CMU |
File | Size | CS | DE | EN | ES | FR | RU | Notes |
---|---|---|---|---|---|---|---|---|
Europarl v7 | 446MB | ✓ | ✓ | ✓ | ✓ | ✓ | Europarl home page | |
News Commentary | 78MB | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | |
302MB | ✓ | ✓ | ✓ | ✓ | ✓ |
Extracted article text from various online news publications. NB the data this year is slightly different from last year's in that we didn't try to re-balance quotes. It does contain lines with "an odd number of quote signs. |
||
News Crawl (articles from 2008) | 1.5GB | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | |
News Crawl (articles from 2009) | 1.6GB | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | |
News Crawl (articles from 2010) | 727MB | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | |
News Crawl (articles from 2011) | 3.1GB | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | |
News Crawl (articles from 2012) | 3.1GB | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
To submit your results, please first convert into into SGML format as required by the NIST BLEU scorer, and then upload it to the website matrix.statmt.org.
Each submitted file has to be in a format that is used by standard scoring scripts such as NIST BLEU or TER.
This format is similar to the one used in the source test set files that were released, except for:
<tstset trglang="en" setid="newstest2013"
srclang="any">
, with trglang set to
either en
, de
, fr
, es
,
cs
or ru
. Important: srclang is
always any
.
sysid="uedin"
.
</tstset>
The script wrap-xml.perl makes the conversion of a output file in one-segment-per-line format into the required SGML file very easy:
Format: wrap-xml.perl LANGUAGE SRC_SGML_FILE SYSTEM_NAME < IN > OUT
Example: wrap-xml.perl en newstest2013-src.de.sgm Google < decoder-output > decoder-output.sgm
Upload happens in three easy steps:
If you are submitting contrastive runs, please submit your primary system first and mark it clearly as the primary submission.
Evaluation will be done both automatically as well as by human judgement.
Release of translation task training data | Early February 2013 |
Release of translation task test set | April 29, 2013 |
Submission deadline for translation task | May 3 2013 |
Release of quality estimation task training data | February 28 2013 |
Release of quality estimation task test data | May 25, 2013 |
Release of system outputs for metrics task | May 10, 2013 |
Submission deadline for metrics task and quality estimation task | May 31, 2013 |
Manual evaluation period | May 17, 2013 - June 7, 2013 |
Supported by the European Commision
under the
project (grant number 288487)