This repository is a result of our participation in the shared task.
We went through the process of building, analyzing, and improving the neural machine translation system.
Poster: link
The shared task was for Estonian-English language pair. It included working with ~19.000.000 sentence pairs.
Shared task main page: link
Shared task on course page: link
Sections below summarize key milestones we went through.
- Our baseline system was 2-layer LSTM with 500 hidden units on both the encoder/decoder with vocabulary size of 20000.
- As a result, we got 21.14 BLEU points on the shared dev set.
More details: link
- We manually analyzed 60 baseline translations.
- Our main observation was that the next problems exist:
- Grammatical problems:
1.1 Passive voice lost
1.2 Wrong abreviation used. - Sentence meaning problems:
2.1 Reversed meaning used
2.2 Important words are lost
2.3 Meaning totally changed
2.4 Losts in meaning of relations between subject in sentence.
- Take a look at our motivating example produced by baseline system. You can see relevant examples in relevant numbers, and bold color for current problematic part:
1.1 "Even traditional websites die, which has killed a continuous flow of information from social networks."
1.2 "In order to understand the size of this figure, 1 TEN is about this energy, which is used by flying mosquito." (In original sentence 1 TeV as unit mentioned).
2.1 "The economy has good times, partly because the EU countries conclude customs duties on trade." (where in right translation custome duties are stoped).
2.1 One more example: "Thanks to this, every student and other student are aware of Erasmus, Leonardo and other cross-border partnerships." (in right translation students like student programs).
2.2 "Since pollution does not feel borders, the EU has been collectively acting in this area for years". (In original sentence, not EU itself, but its members).
2.2 "Two sides discussed the views of the international climate change debate in Copenhagen." (In original example, the debate took place in December).
2.3 "Mr Maris Selga, who is represented in the Arab Republic of Egypt, is represented by Latvia."
2.4 "You are one of our biggest customers and we appreciate the commercial transactions you have done." (where in original sentence they appreciate transactions with them).
More details: link
- In order to address translation issues found after our manual evaluation we
- increased the vocabulary size (from 20000 to 50000)
- Dropout tunning.
- Sockeye model training and tuning.
- The trained system gave us 24.56 BLEU points on the shared dev set that means we have satisfactory increase (3.42) over the baseline. (Including that sockeye model has not been evaluated for technical reasons, details below).
More details: link
- Generally speaking, new model shows significantly better translation with less meaning loss and grammatically more correct.
- As a result you can see that our final system did progress in solving previously mentioned issues (numbered as problems above):
1.1 "Even the traditional website dies , which has been killed by the constant force of the continuous information voiced by social networks."
1.2 "In order to understand the size of this size , 1 TEV is about the energy that is used to move a mosquito."
2.1 "There are good times in the economy, partly because EU countries will end customs duties on customs collection"
2.1 Second example: "Thanks to this , every student and other learner know the Erasmus , Leonardo and the other cross-border partnerships."
2.2 "Since pollution does not feel borders , EU countries have been acting together for years."
2.2 "Two sides discussed the views of the international climate change debate in Copenhagen in December."
2.3 "Latvia is represented by Maris Selga since the Arab Republic of Egypt."
2.4 "You are one of our biggest clients , and we appreciate business transactions with you."
Do not forget to check our poster: [POSTER_LINK]
Firstly we have tried increasing of vocablulary size and three different dropouts: 0.3, 0.5, 0.7. As dropout:0.3 gave the best results, we have chosen this one, and will not include results of models with other dropout (0.5, 0.7).
Model with dropout 0.5 gave score 23.35
Model with dropout 0.7 gave score 18.55
Lastly, we have also trained Sockeye model (1-layer bi-LSTM encoder, 1-layer LSTM decoder with dot attention), but right now translation of this model is in queue on Rocket, so we will include translation results, when script will be exexcuted.
- 1st_sentence: your_results_on_the_test_set_and_interpretation(feel_later)_
- The main dificulties are connected to the hardware we have to work with (rocket.hpc.ut.ee) as we always suffered from being queued for a long time, which then resulted in time losses for model traning and evaluating of generated translations.
- The more data you feed to the model - the better results you achive. We are not claming that model tuning process is insufficient but it is always has high timecost which often is not payed with appropriate improvements. So we would better work on data extension first and then try to play with existing tutings settings for the model and regenerating features for the "new" model.
@yevhen-tyshchenko
@dimatkachuk
@BreakINbaDs
Project board: link