| | |
| | task_type: MrcTaggingTask |
| | task_name: debug-Mirror_W2_MSRAv2_NER_FreezeBertEmbAnd0-3_bs64 |
| | comment: 'bert mrc w/ w2ner for NER' |
| |
|
| | |
| | max_seq_len: 300 |
| | negative_sample_prob: 1.0 |
| | debug_mode: false |
| | mode: w2 |
| |
|
| | |
| | base_model_path: outputs/RobertaBase_data20230314v2/ckpt/MrcGlobalPointerModel.best.pth |
| | plm_dir: hfl/chinese-roberta-wwm-ext |
| | data_dir: resources/NER/MSRA_v2/formatted |
| | output_dir: outputs |
| | task_dir: ${output_dir}/${task_name} |
| | train_filepath: ${data_dir}/train.char.bmes.jsonl |
| | dev_filepath: ${data_dir}/dev.char.bmes.jsonl |
| | test_filepath: ${data_dir}/test.char.bmes.jsonl |
| | ent_type2query_filepath: ${data_dir}/query.json |
| | dump_cache_dir: ${task_dir}/cache |
| | regenerate_cache: true |
| |
|
| | |
| | random_seed: 1227 |
| | eval_on_data: [dev, test] |
| | select_best_on_data: dev |
| | select_best_by_key: metric |
| | best_metric_field: micro.f1 |
| | final_eval_on_test: true |
| |
|
| | warmup_proportion: 0.1 |
| | num_epochs: 5 |
| | epoch_patience: 5 |
| | train_batch_size: 64 |
| | eval_batch_size: 128 |
| | learning_rate: !!float 5e-5 |
| | other_learning_rate: !!float 1e-4 |
| | max_grad_norm: 1.0 |
| | weight_decay: 0.1 |
| |
|
| | |
| | dropout: 0.3 |
| | biaffine_size: 512 |
| |
|