Is there any specifics / benchmark in terms of GPU Ram size wrt the vocabulary size ?
I saw in the tuto / guide that 40k or 50k is mostly retained as the vocab size basis.
Is there any experience with a much larger size ?
If not how do you handle oov beyond that limit ?
is it possible to rescore some kind of n-best list with an extra larger Language model ?