I’m back here with a question, after being elsewhere for a long time. Of course, I certainly lack some fresh informations about the details of recent evolutions of ONMT.
Do you have a link explaining how is working your Infinite Training ?
Would it be possible to link 2 models with infinite training, continuously producing back-translations for each other, using large monolingual data sets ?
It could be a very nice way to solve the problem of learning languages with very low available parallel data.
PS : if the monolingual data sets are known at the starting point, it’s also possible to provide with the good large vocab dicts right at the beginning, even if the very first parallel data set contains only a small part of it.