I’ve trained an OpenNMT model which is working perfectly. However, after being exported to CTranslate2, I’m having a memory issue on prediction (GPU) when the sentence has the
<unk> token (also happens if it contains a non-existing token)
<unk> exists in the vocabulary, and the tokenization sent to the model is correct (replaces non-existing text by
The specific message received is: “Segmentation fault (core dumped)”
CTranslate2 parameters used:
translator: inter_threads: 2 intra_threads: 4 translate_batch: max_batch_size: 1024 batch_type: tokens beam_size: 5 length_penalty: 0.6 max_decoding_length: 256
Any idea of what could be happening?
Thanks for your help,