I am trying to do continued training (fine-tuning) on an in-domain dataset and would like to freeze the embedding layers of the encoder and decoder in the pre-trained model. It can be done in the config file:
# this example is from the documentation # (optional) List of layer to not optimize. freeze_layers: - "encoder/layers/0" - "decoder/output_layer"
However, where can I get the right names of the embedding layers to specify in the config file (like those
"decoder/output_layer" in the example above)?