Best parameters for long sequence

I am working with sequence to sequence model with long sequence at the source language
i.e source language, each sample has length = 200 tokens

How can I choose best parameters to train well this NMT system?

Hello, there is no special parameters of the model for longer sentences, depending on your GPU you might need to adapt the batch size.