I am trying to fine-tune the OpenNMT model on domain-specific data. I have used sentence piece to tokenize data with a subword segmentation model.
There are 260 trained layers, and the untrained layer is 0. I want to freeze 5 to 10 layers to fine-tune the model. How can I freeze a few layers of both the embeddings?