I am using the Dynamic Dataset feature to train an MT model from a large parallel corpus. During training I had an OutOfMemory error and I had to resume the training (I think I avoided the error in the second run by reducing the maximum sentence size). Anyway, two questions arose when resuming the training:
- As Dynamic Dataset takes random samples from the training data, I guess that it is completely safe to resume a training, right?
- But then, even if the training does not crash, is it possible that, due to the random sampling, some of the sentences of the training set are not explored during training and other ones are overrepresented? Should I expect the same translation quality in a system trained with “partition sampling” and in a system trained with “uniform sampling” (keeping all the other parameters unchanged, of course)?
Thanks in advance