diff --git a/configs/train-conll03-context.yaml b/configs/train-conll03-context.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f5bd7b4ad089ca297bf6deff661f6f33aa700293 --- /dev/null +++ b/configs/train-conll03-context.yaml @@ -0,0 +1,24 @@ +# Test how the training strategy affect production. +data_train: data/coNLL-2003/train.txt +data_tune: data/coNLL-2003/valid.txt +data_test: data/coNLL-2003/test.txt +pretrained_path: xlm-roberta-large +max_seq_length: 256 +num_train_epochs: 20 +seed: [101, 102, 103, 104, 105] +warmup_proportion: 0.0 +learning_rate: 5e-6 +eval_batch_size: 16 +train_batch_size: 16 +dropout: 0.2 +wandb: conll03_en_context +#freeze_model: +#sequence_generator: context-right +#sequence_generator: context-right-mix +#sequence_generator: single +#sequence_generator: merged +sequence_generator: union +sequence_generator_for_eval: context-window +output_dir: ../poldeepner2_models/dev/squeeze_research/conll_context/model_{sequence_generator}_{max_seq_length}_{seed}_v +hidden_size: 1024 +