LSTM or Transformer?

I am building a machine translator. My dataset has 90,000 sentence pairs. Should I use Transformer to train the machine translator with such a small dataset? Will it be better than using LSTM?
Thank you.

If this is a research work, I assume you can try both. Consider also the possibility of fine-tuning a (pretrained) multilingual model, especially if your language pair is similar to some high-resource languages.

Check also some posts on the forum about low-resource languages, including this one.

Kind regards,

1 Like