I’ve trained an OpenNMT model which is working perfectly. However, after being exported to CTranslate2, I’m having a memory issue on prediction (GPU) when the sentence has the <unk> token (also happens if it contains a non-existing token)
The token <unk> exists in the vocabulary, and the tokenization sent to the model is correct (replaces non-existing text by <unk>)
The specific message received is: “Segmentation fault (core dumped)”
I quickly tried with the pretrained English-German and using <unk> or OOV tokens did not make any error. So I would need more information to understand what is going on:
Which CTranslate2 version are you using?
Is the model coming from OpenNMT-py or OpenNMT-tf?
Can you post the tokenized input that makes the error?
We trained other similar models not having this issue.
Actually, we tried the same problematic input on another CTranslate2 model with the same source but different target lang, and it worked.