Hello I was just wondering if using word2vec embeddings is exactly the same process as with glove.
my word2vec text file has one line that states vocab and vector size. Then it is followed by lines of vectors (of size 100 in this case). Also I have only created embeddings for the target language.
So far so good, however do I need to do anything extra to this file or can I just do this:
hi ,
I am working on machine translation task.So i want to use two different word2vec models on two different vocabulary .But opennmt-py generates only one vocab.pt how can i convert two different vocab to vector using two different word2vec