I am using OpenNMT Transformer model for summarization task, and I have modified multi_headed_attn.py module for my research. I just add one more linear layer in multi_headed_attn.py.
It can normally train and predict. However, it will predict some empty lines. Most of outputs are fine, but few of outputs are empty. The inputs are all normal sentence, not empty.
Thank you! It is useful for me.
After Setting [min_length], it won’t predict empty line anymore.
The average ROUGE score of generated summary is a little lower than before(the average ROUGE score computed after delete empty line). But not too bad!