Hey everyone (eg @guillaumekln )
What sort of generalizations do we expect NMT to make:
- If we train an NMT on 1M sentences but the word ‘milk’ is only shown in 3x sentences (any 3 diverse sentences) will these 3x non-similar but simple sentences be (likely) translated correctly?
b. Milk, bread & water
c. I like milk
Do we expect that? If yes. Cool. And how?
If not, why not, & how do we expect NMT to work then? We can’t show every word in every sentence?
If we train on 1 million sentences and add an OOV like eg milk = milch as a one word sentence, will it learn to translate milk in real sentences?
If yes, why not supplement the training sentences with 1 word sentence pairs (from a dictionary) to increase the vocab?