❗
The results of the 2023 edition are available.
❗
Test data is now available on github.
❗
Codalab is now ready! See links and information below.
❗
To easily stay up-to-date you can also follow our Twitter account.
❗
Please also register to the QE google-group in order to be able to receive updates and announcements immediately, and ask questions.
Task 1 - sentence-level: https://codalab.lisn.upsaclay.fr/competitions/15043 Task 1 - word-level: https://codalab.lisn.upsaclay.fr/competitions/15044 Task 2 - error span detection: https://codalab.lisn.upsaclay.fr/competitions/15045
Once you have created your account and registered to a competition, please make sure to:
Make only 1 registration for each team on Codalab;
Ensure you have the latest test data: especially the tokenised translations for Task 1 - word level which have been updated;
Follow the submission format indicated for each competition. Note that this year we do not have individual phases for each language pair. Instead, you can submit your predictions all together (in any order of language pairs), i.e. one submission may include all language pairs of a task or just the ones you want to focus on. To know how your submission performed, due to the leaderboard being blind, you need to manually download the scores via “View scoring error log”. Note: you will see the macro-average only if you submit to all language pairs.
Each participant is given a quota of 10 submissions overall, with a limit of 5 per day;
This year we are keeping the leaderboards private (or blind) during the “competition” phase (ending August 17th noon UTC or 23:59 AoE). After that deadline, the “post-competition” phase of each Codalab will become active and you will be able to make as many submissions as you want (still in a limit of 5 a day). Note: only submissions received during the “competition” phase will be used to determine the winner(s) of each task, and thus will be considered for the findings paper.
This shared task focuses on automatic methods for estimating the quality of neural machine translation output at run-time, without relying on reference translations. It will cover estimation at sentence and word levels and critical error detection. This year we put emphasis on:
This year we will release test-sets on the following language-pairs:
See more details in Task 1 and Task 2 description.
Date | |
---|---|
Release of extra training data | 3rd July, 2023 |
Release of dev data | 3rd July 2023 |
Release of test data | 1st August, 2023 |
Submission of test predictions deadline | |
Announcement of results | 3rd September, 2023 |
System description submission deadline | 31st August 2023 |
Paper submission deadline to WMT | 8th September, 2023 |
WMT Notification of acceptance | 6th October, 2023 |
WMT Camera-ready deadline | 18th October, 2023 |
WMT Conference | 6-7 December, 2023 |
In addition to generally advancing the state of the art in quality estimation, our specific goals are:
For all tasks, the datasets and NMT models that generated the translations will be made publicly available.
Participants are also allowed to explore any additional data and resources deemed relevant. Below are the three QE tasks addressing these goals.
The shared task competition will take place on CODALAB (link in July).
Please register with one account per team.
Here are some open source software for QE that might be useful for participants: