Hi @guillaumekln ,
I have been using the custom
DualSourceTransformer from the examples in the repository, but after updating the config.yaml with the necessary additions and vocal files for multiple sources, I’m having issues running this, mainly because I’m getting an error when I run:
CUDA_VISIBLE_DEVICES=0 onmt-main --model_type DualSourceTransformer --config config.yaml --auto_config train mainly because
DualSourceTransformer isn’t a registered model. I tried to use
Transformer but I get the error:
ValueError: Missing field 'source_vocabulary' in the data configuration
I wanted to understand how I could go about using the onmt-main command in this case, what would be the --model_type in this case?