File size: 2,081 Bytes
4350817 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 |
{
"model_wrapper": {
"type": "token_classification",
"num_labels": 3
},
"label_mapper": {
"type": "answer-extraction"
},
"args": {
"type": "default",
"disable_tqdm": true,
"do_eval": true,
"do_train": true,
"eval_steps": 1824,
"evaluation_strategy": "steps",
"fp16": true,
"fp16_backend": "amp",
"gradient_accumulation_steps": 1,
"greater_is_better": true,
"label_names": [
"labels"
],
"logging_dir": "/opt/ml/checkpoints/logs",
"logging_first_step": true,
"logging_steps": 1824,
"lr_scheduler_type": "linear",
"metric_for_best_model": "eval_ANS_f1",
"no_cuda": false,
"num_train_epochs": 3,
"output_dir": "/opt/ml/checkpoints",
"per_device_eval_batch_size": 10,
"per_device_train_batch_size": 4,
"result_dir": "/opt/ml/model",
"save_steps": 1824,
"save_total_limit": null,
"warmup_steps": 985
},
"compute_metrics": {
"type": "seqeval"
},
"dataset_loader": {
"dataset_reader": {
"path": "squad_qa_test_fixture"
},
"data_processor": {
"type": "squad-answer-extraction"
},
"data_adapter": {
"type": "answer-extraction"
}
},
"dev_file_path": "/opt/ml/input/data/validation/dev.json",
"optimizer": {
"type": "huggingface_adamw",
"eps": 1e-06,
"lr": 3e-05,
"parameter_groups": [
[
[
"bias",
"LayerNorm\\\\.weight",
"layer_norm\\\\.weight"
],
{
"weight_decay": 0
}
]
],
"weight_decay": 0.01
},
"pretrained_model_name_or_path": "roberta-large",
"tokenizer_wrapper": {
"type": "answer-extraction"
},
"train_file_path": "/opt/ml/input/data/training/train.json"
} |