I am trying to load a Transformer model checkpoint in order to get the outputs of Transformer Encoder. I tried with the script
OpenNMT-tf/examples/library/minimal_transformer_training.py and also to add a function in
runner.py. Both got a error of
Key encoder/LayerNorm/beta not found in checkpoint with
SelfAttentionEncoder.encode when loading the checkpoint.
Both OpenNMT-tf-1.20.1 with tf-1.4 and OpenNMT-tf-1.22.0 with tf-1.13.1 did not work.
Any help to solve the problem?