Pruning useless weights (4)
Context-sensitive spell checking (2)
NMT's vocabulary (3)
Is there any efficient reference set(English) for bleu scoring? (3)
Word-level Distillation (2)
What is the use of Monolingual Corpora in SMT (3)
Training with out-of-domain data (1)
How is google's transformer translate between Chinese and English? (1)
Multiple translations open nmt (2)
Quality estimation / pred_score (4)
Paragraph vs sentence segmentation (3)
What is the best way to generate parallel corpus as per your past experience? (1)
In parallel corpus what should I keep, single sentence in each line or paragraph for MT? (3)
Metrics (Bleu, ppl, gold ppl, pred ....) (8)
About mask in attention layer and decoder (5)
How to create models that translate very quickly (8)
Exposure bias during training (4)
Improve BLEU by Coverage and Context Gate (7)
Use of Monolingual corpora in OPENNMT (8)
Using features for domain/client/subject adaptation (7)
Has anyone experimented with the dense bridge? (2)
Did someone test RNN with larger recurrence? (6)
Training chatbot with multiple inputs (to add context) (5)
Corpus compactization (3)
Incremental vocab (7)
Automatic training corpus filtering (16)
Implementing hashing to massively improve performance (by 95%) (1)
Speech-to-text using Convolutional LSTM layers (2)
Sentence length affect perplexity decrease (3)
Simple combination between SMT and NMT (1)