I am getting a good translation accuracy of about 82%, but this when I use the n-best 100 and beam 100, but the top-1, top2, …, top5 accuracies are very low.
My questions are:
why?
how can I push this accuracy to the top of my predictions.
any suggestions would be appreciated.
Thank you very much Yasmin.
My training dataset size is 76184 samples. I can’t add more data, it is a specific domain. Do you think that because of early_stopping criteria, the model is not learning enough? I am asking because most of the models in the literature did not use this criteria and they were trained till the end of train_steps of 250000 at least.
Best
Finally, it turns out that we can even generate purely synthetic data (both the source and target), based on the small in-domain data available, as in Domain-Specific Text Generation for Machine Translation (Moslem et al., 2022). You need to first experiment with the two previous approaches, which can be applied independently or together.
Well, early stopping simply means the model is not learning (enough) anymore. However, if you increase its value or remove it, you can test the results on the two checkpoints, one with early stopping and one after further steps.