Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- checkpoint-1185/config.json +149 -0
- checkpoint-1185/model.safetensors +3 -0
- checkpoint-1185/optimizer.pt +3 -0
- checkpoint-1185/rng_state.pth +3 -0
- checkpoint-1185/scheduler.pt +3 -0
- checkpoint-1185/special_tokens_map.json +7 -0
- checkpoint-1185/tokenizer.json +0 -0
- checkpoint-1185/tokenizer_config.json +56 -0
- checkpoint-1185/trainer_state.json +234 -0
- checkpoint-1185/training_args.bin +3 -0
- checkpoint-1185/vocab.txt +0 -0
- checkpoint-237/config.json +149 -0
- checkpoint-237/model.safetensors +3 -0
- checkpoint-237/optimizer.pt +3 -0
- checkpoint-237/rng_state.pth +3 -0
- checkpoint-237/scheduler.pt +3 -0
- checkpoint-237/special_tokens_map.json +7 -0
- checkpoint-237/tokenizer.json +0 -0
- checkpoint-237/tokenizer_config.json +56 -0
- checkpoint-237/trainer_state.json +69 -0
- checkpoint-237/training_args.bin +3 -0
- checkpoint-237/vocab.txt +0 -0
- checkpoint-474/config.json +149 -0
- checkpoint-474/model.safetensors +3 -0
- checkpoint-474/optimizer.pt +3 -0
- checkpoint-474/rng_state.pth +3 -0
- checkpoint-474/scheduler.pt +3 -0
- checkpoint-474/special_tokens_map.json +7 -0
- checkpoint-474/tokenizer.json +0 -0
- checkpoint-474/tokenizer_config.json +56 -0
- checkpoint-474/trainer_state.json +112 -0
- checkpoint-474/training_args.bin +3 -0
- checkpoint-474/vocab.txt +0 -0
- checkpoint-711/config.json +149 -0
- checkpoint-711/model.safetensors +3 -0
- checkpoint-711/optimizer.pt +3 -0
- checkpoint-711/rng_state.pth +3 -0
- checkpoint-711/scheduler.pt +3 -0
- checkpoint-711/special_tokens_map.json +7 -0
- checkpoint-711/tokenizer.json +0 -0
- checkpoint-711/tokenizer_config.json +56 -0
- checkpoint-711/trainer_state.json +155 -0
- checkpoint-711/training_args.bin +3 -0
- checkpoint-711/vocab.txt +0 -0
- checkpoint-948/config.json +149 -0
- checkpoint-948/model.safetensors +3 -0
- checkpoint-948/optimizer.pt +3 -0
- checkpoint-948/rng_state.pth +3 -0
- checkpoint-948/scheduler.pt +3 -0
- checkpoint-948/special_tokens_map.json +7 -0
checkpoint-1185/config.json
ADDED
|
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "bert-base-uncased",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"BertForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"gradient_checkpointing": false,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"id2label": {
|
| 13 |
+
"0": "LABEL_0",
|
| 14 |
+
"1": "LABEL_1",
|
| 15 |
+
"2": "LABEL_2",
|
| 16 |
+
"3": "LABEL_3",
|
| 17 |
+
"4": "LABEL_4",
|
| 18 |
+
"5": "LABEL_5",
|
| 19 |
+
"6": "LABEL_6",
|
| 20 |
+
"7": "LABEL_7",
|
| 21 |
+
"8": "LABEL_8",
|
| 22 |
+
"9": "LABEL_9",
|
| 23 |
+
"10": "LABEL_10",
|
| 24 |
+
"11": "LABEL_11",
|
| 25 |
+
"12": "LABEL_12",
|
| 26 |
+
"13": "LABEL_13",
|
| 27 |
+
"14": "LABEL_14",
|
| 28 |
+
"15": "LABEL_15",
|
| 29 |
+
"16": "LABEL_16",
|
| 30 |
+
"17": "LABEL_17",
|
| 31 |
+
"18": "LABEL_18",
|
| 32 |
+
"19": "LABEL_19",
|
| 33 |
+
"20": "LABEL_20",
|
| 34 |
+
"21": "LABEL_21",
|
| 35 |
+
"22": "LABEL_22",
|
| 36 |
+
"23": "LABEL_23",
|
| 37 |
+
"24": "LABEL_24",
|
| 38 |
+
"25": "LABEL_25",
|
| 39 |
+
"26": "LABEL_26",
|
| 40 |
+
"27": "LABEL_27",
|
| 41 |
+
"28": "LABEL_28",
|
| 42 |
+
"29": "LABEL_29",
|
| 43 |
+
"30": "LABEL_30",
|
| 44 |
+
"31": "LABEL_31",
|
| 45 |
+
"32": "LABEL_32",
|
| 46 |
+
"33": "LABEL_33",
|
| 47 |
+
"34": "LABEL_34",
|
| 48 |
+
"35": "LABEL_35",
|
| 49 |
+
"36": "LABEL_36",
|
| 50 |
+
"37": "LABEL_37",
|
| 51 |
+
"38": "LABEL_38",
|
| 52 |
+
"39": "LABEL_39",
|
| 53 |
+
"40": "LABEL_40",
|
| 54 |
+
"41": "LABEL_41",
|
| 55 |
+
"42": "LABEL_42",
|
| 56 |
+
"43": "LABEL_43",
|
| 57 |
+
"44": "LABEL_44",
|
| 58 |
+
"45": "LABEL_45",
|
| 59 |
+
"46": "LABEL_46",
|
| 60 |
+
"47": "LABEL_47",
|
| 61 |
+
"48": "LABEL_48",
|
| 62 |
+
"49": "LABEL_49",
|
| 63 |
+
"50": "LABEL_50",
|
| 64 |
+
"51": "LABEL_51",
|
| 65 |
+
"52": "LABEL_52",
|
| 66 |
+
"53": "LABEL_53",
|
| 67 |
+
"54": "LABEL_54",
|
| 68 |
+
"55": "LABEL_55",
|
| 69 |
+
"56": "LABEL_56",
|
| 70 |
+
"57": "LABEL_57",
|
| 71 |
+
"58": "LABEL_58"
|
| 72 |
+
},
|
| 73 |
+
"initializer_range": 0.02,
|
| 74 |
+
"intermediate_size": 3072,
|
| 75 |
+
"label2id": {
|
| 76 |
+
"LABEL_0": 0,
|
| 77 |
+
"LABEL_1": 1,
|
| 78 |
+
"LABEL_10": 10,
|
| 79 |
+
"LABEL_11": 11,
|
| 80 |
+
"LABEL_12": 12,
|
| 81 |
+
"LABEL_13": 13,
|
| 82 |
+
"LABEL_14": 14,
|
| 83 |
+
"LABEL_15": 15,
|
| 84 |
+
"LABEL_16": 16,
|
| 85 |
+
"LABEL_17": 17,
|
| 86 |
+
"LABEL_18": 18,
|
| 87 |
+
"LABEL_19": 19,
|
| 88 |
+
"LABEL_2": 2,
|
| 89 |
+
"LABEL_20": 20,
|
| 90 |
+
"LABEL_21": 21,
|
| 91 |
+
"LABEL_22": 22,
|
| 92 |
+
"LABEL_23": 23,
|
| 93 |
+
"LABEL_24": 24,
|
| 94 |
+
"LABEL_25": 25,
|
| 95 |
+
"LABEL_26": 26,
|
| 96 |
+
"LABEL_27": 27,
|
| 97 |
+
"LABEL_28": 28,
|
| 98 |
+
"LABEL_29": 29,
|
| 99 |
+
"LABEL_3": 3,
|
| 100 |
+
"LABEL_30": 30,
|
| 101 |
+
"LABEL_31": 31,
|
| 102 |
+
"LABEL_32": 32,
|
| 103 |
+
"LABEL_33": 33,
|
| 104 |
+
"LABEL_34": 34,
|
| 105 |
+
"LABEL_35": 35,
|
| 106 |
+
"LABEL_36": 36,
|
| 107 |
+
"LABEL_37": 37,
|
| 108 |
+
"LABEL_38": 38,
|
| 109 |
+
"LABEL_39": 39,
|
| 110 |
+
"LABEL_4": 4,
|
| 111 |
+
"LABEL_40": 40,
|
| 112 |
+
"LABEL_41": 41,
|
| 113 |
+
"LABEL_42": 42,
|
| 114 |
+
"LABEL_43": 43,
|
| 115 |
+
"LABEL_44": 44,
|
| 116 |
+
"LABEL_45": 45,
|
| 117 |
+
"LABEL_46": 46,
|
| 118 |
+
"LABEL_47": 47,
|
| 119 |
+
"LABEL_48": 48,
|
| 120 |
+
"LABEL_49": 49,
|
| 121 |
+
"LABEL_5": 5,
|
| 122 |
+
"LABEL_50": 50,
|
| 123 |
+
"LABEL_51": 51,
|
| 124 |
+
"LABEL_52": 52,
|
| 125 |
+
"LABEL_53": 53,
|
| 126 |
+
"LABEL_54": 54,
|
| 127 |
+
"LABEL_55": 55,
|
| 128 |
+
"LABEL_56": 56,
|
| 129 |
+
"LABEL_57": 57,
|
| 130 |
+
"LABEL_58": 58,
|
| 131 |
+
"LABEL_6": 6,
|
| 132 |
+
"LABEL_7": 7,
|
| 133 |
+
"LABEL_8": 8,
|
| 134 |
+
"LABEL_9": 9
|
| 135 |
+
},
|
| 136 |
+
"layer_norm_eps": 1e-12,
|
| 137 |
+
"max_position_embeddings": 512,
|
| 138 |
+
"model_type": "bert",
|
| 139 |
+
"num_attention_heads": 12,
|
| 140 |
+
"num_hidden_layers": 12,
|
| 141 |
+
"pad_token_id": 0,
|
| 142 |
+
"position_embedding_type": "absolute",
|
| 143 |
+
"problem_type": "single_label_classification",
|
| 144 |
+
"torch_dtype": "float32",
|
| 145 |
+
"transformers_version": "4.47.1",
|
| 146 |
+
"type_vocab_size": 2,
|
| 147 |
+
"use_cache": true,
|
| 148 |
+
"vocab_size": 30522
|
| 149 |
+
}
|
checkpoint-1185/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d94de6b4844e865dd1d2145f06c4908aa3003830866a882cf1964f8cc8e0b1f4
|
| 3 |
+
size 438133980
|
checkpoint-1185/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:15b9a61dde14318aeaa62eb94a9116532fe1b83ea311e99a5d83c78b1019abb6
|
| 3 |
+
size 876383354
|
checkpoint-1185/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0a3a9be2c29365ae4d84a07bdadde42bda71e7bea22de1149fb5d09bef45fda7
|
| 3 |
+
size 13990
|
checkpoint-1185/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8402780fadfeb49285d847a5e1c7e9bd248c0e333ebfaf21b5bdf5cd6d44570e
|
| 3 |
+
size 1064
|
checkpoint-1185/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
checkpoint-1185/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-1185/tokenizer_config.json
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": false,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"model_max_length": 512,
|
| 50 |
+
"pad_token": "[PAD]",
|
| 51 |
+
"sep_token": "[SEP]",
|
| 52 |
+
"strip_accents": null,
|
| 53 |
+
"tokenize_chinese_chars": true,
|
| 54 |
+
"tokenizer_class": "BertTokenizer",
|
| 55 |
+
"unk_token": "[UNK]"
|
| 56 |
+
}
|
checkpoint-1185/trainer_state.json
ADDED
|
@@ -0,0 +1,234 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 5.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 1185,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.2109704641350211,
|
| 13 |
+
"grad_norm": 17.118331909179688,
|
| 14 |
+
"learning_rate": 1.9156118143459917e-05,
|
| 15 |
+
"loss": 4.0557,
|
| 16 |
+
"step": 50
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.4219409282700422,
|
| 20 |
+
"grad_norm": 17.078466415405273,
|
| 21 |
+
"learning_rate": 1.8312236286919833e-05,
|
| 22 |
+
"loss": 4.018,
|
| 23 |
+
"step": 100
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.6329113924050633,
|
| 27 |
+
"grad_norm": 18.140644073486328,
|
| 28 |
+
"learning_rate": 1.746835443037975e-05,
|
| 29 |
+
"loss": 3.9942,
|
| 30 |
+
"step": 150
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.8438818565400844,
|
| 34 |
+
"grad_norm": 14.235407829284668,
|
| 35 |
+
"learning_rate": 1.662447257383966e-05,
|
| 36 |
+
"loss": 3.9127,
|
| 37 |
+
"step": 200
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 1.0,
|
| 41 |
+
"eval_loss": 3.480090379714966,
|
| 42 |
+
"eval_runtime": 253.6051,
|
| 43 |
+
"eval_samples_per_second": 0.469,
|
| 44 |
+
"eval_steps_per_second": 0.237,
|
| 45 |
+
"step": 237
|
| 46 |
+
},
|
| 47 |
+
{
|
| 48 |
+
"epoch": 1.0548523206751055,
|
| 49 |
+
"grad_norm": 14.833487510681152,
|
| 50 |
+
"learning_rate": 1.578059071729958e-05,
|
| 51 |
+
"loss": 3.6803,
|
| 52 |
+
"step": 250
|
| 53 |
+
},
|
| 54 |
+
{
|
| 55 |
+
"epoch": 1.2658227848101267,
|
| 56 |
+
"grad_norm": 18.978660583496094,
|
| 57 |
+
"learning_rate": 1.4936708860759495e-05,
|
| 58 |
+
"loss": 3.4012,
|
| 59 |
+
"step": 300
|
| 60 |
+
},
|
| 61 |
+
{
|
| 62 |
+
"epoch": 1.4767932489451476,
|
| 63 |
+
"grad_norm": 18.297719955444336,
|
| 64 |
+
"learning_rate": 1.4092827004219412e-05,
|
| 65 |
+
"loss": 3.1245,
|
| 66 |
+
"step": 350
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 1.6877637130801688,
|
| 70 |
+
"grad_norm": 16.072620391845703,
|
| 71 |
+
"learning_rate": 1.3248945147679326e-05,
|
| 72 |
+
"loss": 3.122,
|
| 73 |
+
"step": 400
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"epoch": 1.8987341772151898,
|
| 77 |
+
"grad_norm": 14.235777854919434,
|
| 78 |
+
"learning_rate": 1.240506329113924e-05,
|
| 79 |
+
"loss": 2.7533,
|
| 80 |
+
"step": 450
|
| 81 |
+
},
|
| 82 |
+
{
|
| 83 |
+
"epoch": 2.0,
|
| 84 |
+
"eval_loss": 2.393742084503174,
|
| 85 |
+
"eval_runtime": 230.7613,
|
| 86 |
+
"eval_samples_per_second": 0.516,
|
| 87 |
+
"eval_steps_per_second": 0.26,
|
| 88 |
+
"step": 474
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"epoch": 2.109704641350211,
|
| 92 |
+
"grad_norm": 15.517921447753906,
|
| 93 |
+
"learning_rate": 1.1561181434599158e-05,
|
| 94 |
+
"loss": 2.5393,
|
| 95 |
+
"step": 500
|
| 96 |
+
},
|
| 97 |
+
{
|
| 98 |
+
"epoch": 2.320675105485232,
|
| 99 |
+
"grad_norm": 12.792672157287598,
|
| 100 |
+
"learning_rate": 1.0717299578059072e-05,
|
| 101 |
+
"loss": 2.4086,
|
| 102 |
+
"step": 550
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"epoch": 2.5316455696202533,
|
| 106 |
+
"grad_norm": 15.125005722045898,
|
| 107 |
+
"learning_rate": 9.87341772151899e-06,
|
| 108 |
+
"loss": 2.3738,
|
| 109 |
+
"step": 600
|
| 110 |
+
},
|
| 111 |
+
{
|
| 112 |
+
"epoch": 2.742616033755274,
|
| 113 |
+
"grad_norm": 14.103851318359375,
|
| 114 |
+
"learning_rate": 9.029535864978903e-06,
|
| 115 |
+
"loss": 2.3799,
|
| 116 |
+
"step": 650
|
| 117 |
+
},
|
| 118 |
+
{
|
| 119 |
+
"epoch": 2.9535864978902953,
|
| 120 |
+
"grad_norm": 13.69605541229248,
|
| 121 |
+
"learning_rate": 8.18565400843882e-06,
|
| 122 |
+
"loss": 2.0113,
|
| 123 |
+
"step": 700
|
| 124 |
+
},
|
| 125 |
+
{
|
| 126 |
+
"epoch": 3.0,
|
| 127 |
+
"eval_loss": 1.8042149543762207,
|
| 128 |
+
"eval_runtime": 229.2696,
|
| 129 |
+
"eval_samples_per_second": 0.519,
|
| 130 |
+
"eval_steps_per_second": 0.262,
|
| 131 |
+
"step": 711
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"epoch": 3.1645569620253164,
|
| 135 |
+
"grad_norm": 15.687886238098145,
|
| 136 |
+
"learning_rate": 7.341772151898735e-06,
|
| 137 |
+
"loss": 1.9995,
|
| 138 |
+
"step": 750
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"epoch": 3.3755274261603376,
|
| 142 |
+
"grad_norm": 13.975310325622559,
|
| 143 |
+
"learning_rate": 6.49789029535865e-06,
|
| 144 |
+
"loss": 1.9129,
|
| 145 |
+
"step": 800
|
| 146 |
+
},
|
| 147 |
+
{
|
| 148 |
+
"epoch": 3.586497890295359,
|
| 149 |
+
"grad_norm": 14.94083309173584,
|
| 150 |
+
"learning_rate": 5.654008438818566e-06,
|
| 151 |
+
"loss": 2.0081,
|
| 152 |
+
"step": 850
|
| 153 |
+
},
|
| 154 |
+
{
|
| 155 |
+
"epoch": 3.7974683544303796,
|
| 156 |
+
"grad_norm": 16.014385223388672,
|
| 157 |
+
"learning_rate": 4.8101265822784815e-06,
|
| 158 |
+
"loss": 1.7606,
|
| 159 |
+
"step": 900
|
| 160 |
+
},
|
| 161 |
+
{
|
| 162 |
+
"epoch": 4.0,
|
| 163 |
+
"eval_loss": 1.5033947229385376,
|
| 164 |
+
"eval_runtime": 229.0556,
|
| 165 |
+
"eval_samples_per_second": 0.52,
|
| 166 |
+
"eval_steps_per_second": 0.262,
|
| 167 |
+
"step": 948
|
| 168 |
+
},
|
| 169 |
+
{
|
| 170 |
+
"epoch": 4.008438818565401,
|
| 171 |
+
"grad_norm": 10.119342803955078,
|
| 172 |
+
"learning_rate": 3.9662447257383965e-06,
|
| 173 |
+
"loss": 1.7556,
|
| 174 |
+
"step": 950
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"epoch": 4.219409282700422,
|
| 178 |
+
"grad_norm": 16.55748748779297,
|
| 179 |
+
"learning_rate": 3.1223628691983127e-06,
|
| 180 |
+
"loss": 1.6124,
|
| 181 |
+
"step": 1000
|
| 182 |
+
},
|
| 183 |
+
{
|
| 184 |
+
"epoch": 4.430379746835443,
|
| 185 |
+
"grad_norm": 10.142475128173828,
|
| 186 |
+
"learning_rate": 2.278481012658228e-06,
|
| 187 |
+
"loss": 1.7171,
|
| 188 |
+
"step": 1050
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"epoch": 4.641350210970464,
|
| 192 |
+
"grad_norm": 15.821284294128418,
|
| 193 |
+
"learning_rate": 1.4345991561181436e-06,
|
| 194 |
+
"loss": 1.7856,
|
| 195 |
+
"step": 1100
|
| 196 |
+
},
|
| 197 |
+
{
|
| 198 |
+
"epoch": 4.852320675105485,
|
| 199 |
+
"grad_norm": 11.253273963928223,
|
| 200 |
+
"learning_rate": 5.907172995780591e-07,
|
| 201 |
+
"loss": 1.7111,
|
| 202 |
+
"step": 1150
|
| 203 |
+
},
|
| 204 |
+
{
|
| 205 |
+
"epoch": 5.0,
|
| 206 |
+
"eval_loss": 1.4161142110824585,
|
| 207 |
+
"eval_runtime": 234.4231,
|
| 208 |
+
"eval_samples_per_second": 0.508,
|
| 209 |
+
"eval_steps_per_second": 0.256,
|
| 210 |
+
"step": 1185
|
| 211 |
+
}
|
| 212 |
+
],
|
| 213 |
+
"logging_steps": 50,
|
| 214 |
+
"max_steps": 1185,
|
| 215 |
+
"num_input_tokens_seen": 0,
|
| 216 |
+
"num_train_epochs": 5,
|
| 217 |
+
"save_steps": 500,
|
| 218 |
+
"stateful_callbacks": {
|
| 219 |
+
"TrainerControl": {
|
| 220 |
+
"args": {
|
| 221 |
+
"should_epoch_stop": false,
|
| 222 |
+
"should_evaluate": false,
|
| 223 |
+
"should_log": false,
|
| 224 |
+
"should_save": true,
|
| 225 |
+
"should_training_stop": true
|
| 226 |
+
},
|
| 227 |
+
"attributes": {}
|
| 228 |
+
}
|
| 229 |
+
},
|
| 230 |
+
"total_flos": 622576104944640.0,
|
| 231 |
+
"train_batch_size": 2,
|
| 232 |
+
"trial_name": null,
|
| 233 |
+
"trial_params": null
|
| 234 |
+
}
|
checkpoint-1185/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c864801bcd65938a6635dc9ef295501c2002bc5cda8339db5ff3243768ab40f
|
| 3 |
+
size 5240
|
checkpoint-1185/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-237/config.json
ADDED
|
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "bert-base-uncased",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"BertForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"gradient_checkpointing": false,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"id2label": {
|
| 13 |
+
"0": "LABEL_0",
|
| 14 |
+
"1": "LABEL_1",
|
| 15 |
+
"2": "LABEL_2",
|
| 16 |
+
"3": "LABEL_3",
|
| 17 |
+
"4": "LABEL_4",
|
| 18 |
+
"5": "LABEL_5",
|
| 19 |
+
"6": "LABEL_6",
|
| 20 |
+
"7": "LABEL_7",
|
| 21 |
+
"8": "LABEL_8",
|
| 22 |
+
"9": "LABEL_9",
|
| 23 |
+
"10": "LABEL_10",
|
| 24 |
+
"11": "LABEL_11",
|
| 25 |
+
"12": "LABEL_12",
|
| 26 |
+
"13": "LABEL_13",
|
| 27 |
+
"14": "LABEL_14",
|
| 28 |
+
"15": "LABEL_15",
|
| 29 |
+
"16": "LABEL_16",
|
| 30 |
+
"17": "LABEL_17",
|
| 31 |
+
"18": "LABEL_18",
|
| 32 |
+
"19": "LABEL_19",
|
| 33 |
+
"20": "LABEL_20",
|
| 34 |
+
"21": "LABEL_21",
|
| 35 |
+
"22": "LABEL_22",
|
| 36 |
+
"23": "LABEL_23",
|
| 37 |
+
"24": "LABEL_24",
|
| 38 |
+
"25": "LABEL_25",
|
| 39 |
+
"26": "LABEL_26",
|
| 40 |
+
"27": "LABEL_27",
|
| 41 |
+
"28": "LABEL_28",
|
| 42 |
+
"29": "LABEL_29",
|
| 43 |
+
"30": "LABEL_30",
|
| 44 |
+
"31": "LABEL_31",
|
| 45 |
+
"32": "LABEL_32",
|
| 46 |
+
"33": "LABEL_33",
|
| 47 |
+
"34": "LABEL_34",
|
| 48 |
+
"35": "LABEL_35",
|
| 49 |
+
"36": "LABEL_36",
|
| 50 |
+
"37": "LABEL_37",
|
| 51 |
+
"38": "LABEL_38",
|
| 52 |
+
"39": "LABEL_39",
|
| 53 |
+
"40": "LABEL_40",
|
| 54 |
+
"41": "LABEL_41",
|
| 55 |
+
"42": "LABEL_42",
|
| 56 |
+
"43": "LABEL_43",
|
| 57 |
+
"44": "LABEL_44",
|
| 58 |
+
"45": "LABEL_45",
|
| 59 |
+
"46": "LABEL_46",
|
| 60 |
+
"47": "LABEL_47",
|
| 61 |
+
"48": "LABEL_48",
|
| 62 |
+
"49": "LABEL_49",
|
| 63 |
+
"50": "LABEL_50",
|
| 64 |
+
"51": "LABEL_51",
|
| 65 |
+
"52": "LABEL_52",
|
| 66 |
+
"53": "LABEL_53",
|
| 67 |
+
"54": "LABEL_54",
|
| 68 |
+
"55": "LABEL_55",
|
| 69 |
+
"56": "LABEL_56",
|
| 70 |
+
"57": "LABEL_57",
|
| 71 |
+
"58": "LABEL_58"
|
| 72 |
+
},
|
| 73 |
+
"initializer_range": 0.02,
|
| 74 |
+
"intermediate_size": 3072,
|
| 75 |
+
"label2id": {
|
| 76 |
+
"LABEL_0": 0,
|
| 77 |
+
"LABEL_1": 1,
|
| 78 |
+
"LABEL_10": 10,
|
| 79 |
+
"LABEL_11": 11,
|
| 80 |
+
"LABEL_12": 12,
|
| 81 |
+
"LABEL_13": 13,
|
| 82 |
+
"LABEL_14": 14,
|
| 83 |
+
"LABEL_15": 15,
|
| 84 |
+
"LABEL_16": 16,
|
| 85 |
+
"LABEL_17": 17,
|
| 86 |
+
"LABEL_18": 18,
|
| 87 |
+
"LABEL_19": 19,
|
| 88 |
+
"LABEL_2": 2,
|
| 89 |
+
"LABEL_20": 20,
|
| 90 |
+
"LABEL_21": 21,
|
| 91 |
+
"LABEL_22": 22,
|
| 92 |
+
"LABEL_23": 23,
|
| 93 |
+
"LABEL_24": 24,
|
| 94 |
+
"LABEL_25": 25,
|
| 95 |
+
"LABEL_26": 26,
|
| 96 |
+
"LABEL_27": 27,
|
| 97 |
+
"LABEL_28": 28,
|
| 98 |
+
"LABEL_29": 29,
|
| 99 |
+
"LABEL_3": 3,
|
| 100 |
+
"LABEL_30": 30,
|
| 101 |
+
"LABEL_31": 31,
|
| 102 |
+
"LABEL_32": 32,
|
| 103 |
+
"LABEL_33": 33,
|
| 104 |
+
"LABEL_34": 34,
|
| 105 |
+
"LABEL_35": 35,
|
| 106 |
+
"LABEL_36": 36,
|
| 107 |
+
"LABEL_37": 37,
|
| 108 |
+
"LABEL_38": 38,
|
| 109 |
+
"LABEL_39": 39,
|
| 110 |
+
"LABEL_4": 4,
|
| 111 |
+
"LABEL_40": 40,
|
| 112 |
+
"LABEL_41": 41,
|
| 113 |
+
"LABEL_42": 42,
|
| 114 |
+
"LABEL_43": 43,
|
| 115 |
+
"LABEL_44": 44,
|
| 116 |
+
"LABEL_45": 45,
|
| 117 |
+
"LABEL_46": 46,
|
| 118 |
+
"LABEL_47": 47,
|
| 119 |
+
"LABEL_48": 48,
|
| 120 |
+
"LABEL_49": 49,
|
| 121 |
+
"LABEL_5": 5,
|
| 122 |
+
"LABEL_50": 50,
|
| 123 |
+
"LABEL_51": 51,
|
| 124 |
+
"LABEL_52": 52,
|
| 125 |
+
"LABEL_53": 53,
|
| 126 |
+
"LABEL_54": 54,
|
| 127 |
+
"LABEL_55": 55,
|
| 128 |
+
"LABEL_56": 56,
|
| 129 |
+
"LABEL_57": 57,
|
| 130 |
+
"LABEL_58": 58,
|
| 131 |
+
"LABEL_6": 6,
|
| 132 |
+
"LABEL_7": 7,
|
| 133 |
+
"LABEL_8": 8,
|
| 134 |
+
"LABEL_9": 9
|
| 135 |
+
},
|
| 136 |
+
"layer_norm_eps": 1e-12,
|
| 137 |
+
"max_position_embeddings": 512,
|
| 138 |
+
"model_type": "bert",
|
| 139 |
+
"num_attention_heads": 12,
|
| 140 |
+
"num_hidden_layers": 12,
|
| 141 |
+
"pad_token_id": 0,
|
| 142 |
+
"position_embedding_type": "absolute",
|
| 143 |
+
"problem_type": "single_label_classification",
|
| 144 |
+
"torch_dtype": "float32",
|
| 145 |
+
"transformers_version": "4.47.1",
|
| 146 |
+
"type_vocab_size": 2,
|
| 147 |
+
"use_cache": true,
|
| 148 |
+
"vocab_size": 30522
|
| 149 |
+
}
|
checkpoint-237/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9059cbc7406daa2de5b5a90a264f0fea948f38cd85f2dfaf56efeb9aad659707
|
| 3 |
+
size 438133980
|
checkpoint-237/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:17fa4f33b492c8943a7ed225a6927f0f10ef578105c6645c2f19b4f8eb504a19
|
| 3 |
+
size 876383354
|
checkpoint-237/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:050f9824d04918fcf622694fba6928aa8c89accb27c09af953c866c57c937897
|
| 3 |
+
size 13990
|
checkpoint-237/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:679ef07df3089743e6bb549a1b1dfcdb043edc0dc3f913acf64397b23b940e7a
|
| 3 |
+
size 1064
|
checkpoint-237/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
checkpoint-237/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-237/tokenizer_config.json
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": false,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"model_max_length": 512,
|
| 50 |
+
"pad_token": "[PAD]",
|
| 51 |
+
"sep_token": "[SEP]",
|
| 52 |
+
"strip_accents": null,
|
| 53 |
+
"tokenize_chinese_chars": true,
|
| 54 |
+
"tokenizer_class": "BertTokenizer",
|
| 55 |
+
"unk_token": "[UNK]"
|
| 56 |
+
}
|
checkpoint-237/trainer_state.json
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 1.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 237,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.2109704641350211,
|
| 13 |
+
"grad_norm": 17.118331909179688,
|
| 14 |
+
"learning_rate": 1.9156118143459917e-05,
|
| 15 |
+
"loss": 4.0557,
|
| 16 |
+
"step": 50
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.4219409282700422,
|
| 20 |
+
"grad_norm": 17.078466415405273,
|
| 21 |
+
"learning_rate": 1.8312236286919833e-05,
|
| 22 |
+
"loss": 4.018,
|
| 23 |
+
"step": 100
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.6329113924050633,
|
| 27 |
+
"grad_norm": 18.140644073486328,
|
| 28 |
+
"learning_rate": 1.746835443037975e-05,
|
| 29 |
+
"loss": 3.9942,
|
| 30 |
+
"step": 150
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.8438818565400844,
|
| 34 |
+
"grad_norm": 14.235407829284668,
|
| 35 |
+
"learning_rate": 1.662447257383966e-05,
|
| 36 |
+
"loss": 3.9127,
|
| 37 |
+
"step": 200
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 1.0,
|
| 41 |
+
"eval_loss": 3.480090379714966,
|
| 42 |
+
"eval_runtime": 253.6051,
|
| 43 |
+
"eval_samples_per_second": 0.469,
|
| 44 |
+
"eval_steps_per_second": 0.237,
|
| 45 |
+
"step": 237
|
| 46 |
+
}
|
| 47 |
+
],
|
| 48 |
+
"logging_steps": 50,
|
| 49 |
+
"max_steps": 1185,
|
| 50 |
+
"num_input_tokens_seen": 0,
|
| 51 |
+
"num_train_epochs": 5,
|
| 52 |
+
"save_steps": 500,
|
| 53 |
+
"stateful_callbacks": {
|
| 54 |
+
"TrainerControl": {
|
| 55 |
+
"args": {
|
| 56 |
+
"should_epoch_stop": false,
|
| 57 |
+
"should_evaluate": false,
|
| 58 |
+
"should_log": false,
|
| 59 |
+
"should_save": true,
|
| 60 |
+
"should_training_stop": false
|
| 61 |
+
},
|
| 62 |
+
"attributes": {}
|
| 63 |
+
}
|
| 64 |
+
},
|
| 65 |
+
"total_flos": 124515220988928.0,
|
| 66 |
+
"train_batch_size": 2,
|
| 67 |
+
"trial_name": null,
|
| 68 |
+
"trial_params": null
|
| 69 |
+
}
|
checkpoint-237/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c864801bcd65938a6635dc9ef295501c2002bc5cda8339db5ff3243768ab40f
|
| 3 |
+
size 5240
|
checkpoint-237/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-474/config.json
ADDED
|
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "bert-base-uncased",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"BertForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"gradient_checkpointing": false,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"id2label": {
|
| 13 |
+
"0": "LABEL_0",
|
| 14 |
+
"1": "LABEL_1",
|
| 15 |
+
"2": "LABEL_2",
|
| 16 |
+
"3": "LABEL_3",
|
| 17 |
+
"4": "LABEL_4",
|
| 18 |
+
"5": "LABEL_5",
|
| 19 |
+
"6": "LABEL_6",
|
| 20 |
+
"7": "LABEL_7",
|
| 21 |
+
"8": "LABEL_8",
|
| 22 |
+
"9": "LABEL_9",
|
| 23 |
+
"10": "LABEL_10",
|
| 24 |
+
"11": "LABEL_11",
|
| 25 |
+
"12": "LABEL_12",
|
| 26 |
+
"13": "LABEL_13",
|
| 27 |
+
"14": "LABEL_14",
|
| 28 |
+
"15": "LABEL_15",
|
| 29 |
+
"16": "LABEL_16",
|
| 30 |
+
"17": "LABEL_17",
|
| 31 |
+
"18": "LABEL_18",
|
| 32 |
+
"19": "LABEL_19",
|
| 33 |
+
"20": "LABEL_20",
|
| 34 |
+
"21": "LABEL_21",
|
| 35 |
+
"22": "LABEL_22",
|
| 36 |
+
"23": "LABEL_23",
|
| 37 |
+
"24": "LABEL_24",
|
| 38 |
+
"25": "LABEL_25",
|
| 39 |
+
"26": "LABEL_26",
|
| 40 |
+
"27": "LABEL_27",
|
| 41 |
+
"28": "LABEL_28",
|
| 42 |
+
"29": "LABEL_29",
|
| 43 |
+
"30": "LABEL_30",
|
| 44 |
+
"31": "LABEL_31",
|
| 45 |
+
"32": "LABEL_32",
|
| 46 |
+
"33": "LABEL_33",
|
| 47 |
+
"34": "LABEL_34",
|
| 48 |
+
"35": "LABEL_35",
|
| 49 |
+
"36": "LABEL_36",
|
| 50 |
+
"37": "LABEL_37",
|
| 51 |
+
"38": "LABEL_38",
|
| 52 |
+
"39": "LABEL_39",
|
| 53 |
+
"40": "LABEL_40",
|
| 54 |
+
"41": "LABEL_41",
|
| 55 |
+
"42": "LABEL_42",
|
| 56 |
+
"43": "LABEL_43",
|
| 57 |
+
"44": "LABEL_44",
|
| 58 |
+
"45": "LABEL_45",
|
| 59 |
+
"46": "LABEL_46",
|
| 60 |
+
"47": "LABEL_47",
|
| 61 |
+
"48": "LABEL_48",
|
| 62 |
+
"49": "LABEL_49",
|
| 63 |
+
"50": "LABEL_50",
|
| 64 |
+
"51": "LABEL_51",
|
| 65 |
+
"52": "LABEL_52",
|
| 66 |
+
"53": "LABEL_53",
|
| 67 |
+
"54": "LABEL_54",
|
| 68 |
+
"55": "LABEL_55",
|
| 69 |
+
"56": "LABEL_56",
|
| 70 |
+
"57": "LABEL_57",
|
| 71 |
+
"58": "LABEL_58"
|
| 72 |
+
},
|
| 73 |
+
"initializer_range": 0.02,
|
| 74 |
+
"intermediate_size": 3072,
|
| 75 |
+
"label2id": {
|
| 76 |
+
"LABEL_0": 0,
|
| 77 |
+
"LABEL_1": 1,
|
| 78 |
+
"LABEL_10": 10,
|
| 79 |
+
"LABEL_11": 11,
|
| 80 |
+
"LABEL_12": 12,
|
| 81 |
+
"LABEL_13": 13,
|
| 82 |
+
"LABEL_14": 14,
|
| 83 |
+
"LABEL_15": 15,
|
| 84 |
+
"LABEL_16": 16,
|
| 85 |
+
"LABEL_17": 17,
|
| 86 |
+
"LABEL_18": 18,
|
| 87 |
+
"LABEL_19": 19,
|
| 88 |
+
"LABEL_2": 2,
|
| 89 |
+
"LABEL_20": 20,
|
| 90 |
+
"LABEL_21": 21,
|
| 91 |
+
"LABEL_22": 22,
|
| 92 |
+
"LABEL_23": 23,
|
| 93 |
+
"LABEL_24": 24,
|
| 94 |
+
"LABEL_25": 25,
|
| 95 |
+
"LABEL_26": 26,
|
| 96 |
+
"LABEL_27": 27,
|
| 97 |
+
"LABEL_28": 28,
|
| 98 |
+
"LABEL_29": 29,
|
| 99 |
+
"LABEL_3": 3,
|
| 100 |
+
"LABEL_30": 30,
|
| 101 |
+
"LABEL_31": 31,
|
| 102 |
+
"LABEL_32": 32,
|
| 103 |
+
"LABEL_33": 33,
|
| 104 |
+
"LABEL_34": 34,
|
| 105 |
+
"LABEL_35": 35,
|
| 106 |
+
"LABEL_36": 36,
|
| 107 |
+
"LABEL_37": 37,
|
| 108 |
+
"LABEL_38": 38,
|
| 109 |
+
"LABEL_39": 39,
|
| 110 |
+
"LABEL_4": 4,
|
| 111 |
+
"LABEL_40": 40,
|
| 112 |
+
"LABEL_41": 41,
|
| 113 |
+
"LABEL_42": 42,
|
| 114 |
+
"LABEL_43": 43,
|
| 115 |
+
"LABEL_44": 44,
|
| 116 |
+
"LABEL_45": 45,
|
| 117 |
+
"LABEL_46": 46,
|
| 118 |
+
"LABEL_47": 47,
|
| 119 |
+
"LABEL_48": 48,
|
| 120 |
+
"LABEL_49": 49,
|
| 121 |
+
"LABEL_5": 5,
|
| 122 |
+
"LABEL_50": 50,
|
| 123 |
+
"LABEL_51": 51,
|
| 124 |
+
"LABEL_52": 52,
|
| 125 |
+
"LABEL_53": 53,
|
| 126 |
+
"LABEL_54": 54,
|
| 127 |
+
"LABEL_55": 55,
|
| 128 |
+
"LABEL_56": 56,
|
| 129 |
+
"LABEL_57": 57,
|
| 130 |
+
"LABEL_58": 58,
|
| 131 |
+
"LABEL_6": 6,
|
| 132 |
+
"LABEL_7": 7,
|
| 133 |
+
"LABEL_8": 8,
|
| 134 |
+
"LABEL_9": 9
|
| 135 |
+
},
|
| 136 |
+
"layer_norm_eps": 1e-12,
|
| 137 |
+
"max_position_embeddings": 512,
|
| 138 |
+
"model_type": "bert",
|
| 139 |
+
"num_attention_heads": 12,
|
| 140 |
+
"num_hidden_layers": 12,
|
| 141 |
+
"pad_token_id": 0,
|
| 142 |
+
"position_embedding_type": "absolute",
|
| 143 |
+
"problem_type": "single_label_classification",
|
| 144 |
+
"torch_dtype": "float32",
|
| 145 |
+
"transformers_version": "4.47.1",
|
| 146 |
+
"type_vocab_size": 2,
|
| 147 |
+
"use_cache": true,
|
| 148 |
+
"vocab_size": 30522
|
| 149 |
+
}
|
checkpoint-474/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6e18d1156429d6faa40c45b848fe79005ddba349136166b5640f806b100daa5
|
| 3 |
+
size 438133980
|
checkpoint-474/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dcc70523e7c839c1cac5e389a25d089d093c61baf207f9df6fdddab828024c8f
|
| 3 |
+
size 876383354
|
checkpoint-474/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a0f1be354591bb6634b2c5cc73ef8264c11fe2bdea540509de8564a55ee228ca
|
| 3 |
+
size 13990
|
checkpoint-474/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f6cb5aeaa4e6ea76c07158fd5faf33d3ddee75b9887090ff37cc434a28c9211
|
| 3 |
+
size 1064
|
checkpoint-474/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
checkpoint-474/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-474/tokenizer_config.json
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": false,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"model_max_length": 512,
|
| 50 |
+
"pad_token": "[PAD]",
|
| 51 |
+
"sep_token": "[SEP]",
|
| 52 |
+
"strip_accents": null,
|
| 53 |
+
"tokenize_chinese_chars": true,
|
| 54 |
+
"tokenizer_class": "BertTokenizer",
|
| 55 |
+
"unk_token": "[UNK]"
|
| 56 |
+
}
|
checkpoint-474/trainer_state.json
ADDED
|
@@ -0,0 +1,112 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 2.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 474,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.2109704641350211,
|
| 13 |
+
"grad_norm": 17.118331909179688,
|
| 14 |
+
"learning_rate": 1.9156118143459917e-05,
|
| 15 |
+
"loss": 4.0557,
|
| 16 |
+
"step": 50
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.4219409282700422,
|
| 20 |
+
"grad_norm": 17.078466415405273,
|
| 21 |
+
"learning_rate": 1.8312236286919833e-05,
|
| 22 |
+
"loss": 4.018,
|
| 23 |
+
"step": 100
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.6329113924050633,
|
| 27 |
+
"grad_norm": 18.140644073486328,
|
| 28 |
+
"learning_rate": 1.746835443037975e-05,
|
| 29 |
+
"loss": 3.9942,
|
| 30 |
+
"step": 150
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.8438818565400844,
|
| 34 |
+
"grad_norm": 14.235407829284668,
|
| 35 |
+
"learning_rate": 1.662447257383966e-05,
|
| 36 |
+
"loss": 3.9127,
|
| 37 |
+
"step": 200
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 1.0,
|
| 41 |
+
"eval_loss": 3.480090379714966,
|
| 42 |
+
"eval_runtime": 253.6051,
|
| 43 |
+
"eval_samples_per_second": 0.469,
|
| 44 |
+
"eval_steps_per_second": 0.237,
|
| 45 |
+
"step": 237
|
| 46 |
+
},
|
| 47 |
+
{
|
| 48 |
+
"epoch": 1.0548523206751055,
|
| 49 |
+
"grad_norm": 14.833487510681152,
|
| 50 |
+
"learning_rate": 1.578059071729958e-05,
|
| 51 |
+
"loss": 3.6803,
|
| 52 |
+
"step": 250
|
| 53 |
+
},
|
| 54 |
+
{
|
| 55 |
+
"epoch": 1.2658227848101267,
|
| 56 |
+
"grad_norm": 18.978660583496094,
|
| 57 |
+
"learning_rate": 1.4936708860759495e-05,
|
| 58 |
+
"loss": 3.4012,
|
| 59 |
+
"step": 300
|
| 60 |
+
},
|
| 61 |
+
{
|
| 62 |
+
"epoch": 1.4767932489451476,
|
| 63 |
+
"grad_norm": 18.297719955444336,
|
| 64 |
+
"learning_rate": 1.4092827004219412e-05,
|
| 65 |
+
"loss": 3.1245,
|
| 66 |
+
"step": 350
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 1.6877637130801688,
|
| 70 |
+
"grad_norm": 16.072620391845703,
|
| 71 |
+
"learning_rate": 1.3248945147679326e-05,
|
| 72 |
+
"loss": 3.122,
|
| 73 |
+
"step": 400
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"epoch": 1.8987341772151898,
|
| 77 |
+
"grad_norm": 14.235777854919434,
|
| 78 |
+
"learning_rate": 1.240506329113924e-05,
|
| 79 |
+
"loss": 2.7533,
|
| 80 |
+
"step": 450
|
| 81 |
+
},
|
| 82 |
+
{
|
| 83 |
+
"epoch": 2.0,
|
| 84 |
+
"eval_loss": 2.393742084503174,
|
| 85 |
+
"eval_runtime": 230.7613,
|
| 86 |
+
"eval_samples_per_second": 0.516,
|
| 87 |
+
"eval_steps_per_second": 0.26,
|
| 88 |
+
"step": 474
|
| 89 |
+
}
|
| 90 |
+
],
|
| 91 |
+
"logging_steps": 50,
|
| 92 |
+
"max_steps": 1185,
|
| 93 |
+
"num_input_tokens_seen": 0,
|
| 94 |
+
"num_train_epochs": 5,
|
| 95 |
+
"save_steps": 500,
|
| 96 |
+
"stateful_callbacks": {
|
| 97 |
+
"TrainerControl": {
|
| 98 |
+
"args": {
|
| 99 |
+
"should_epoch_stop": false,
|
| 100 |
+
"should_evaluate": false,
|
| 101 |
+
"should_log": false,
|
| 102 |
+
"should_save": true,
|
| 103 |
+
"should_training_stop": false
|
| 104 |
+
},
|
| 105 |
+
"attributes": {}
|
| 106 |
+
}
|
| 107 |
+
},
|
| 108 |
+
"total_flos": 249030441977856.0,
|
| 109 |
+
"train_batch_size": 2,
|
| 110 |
+
"trial_name": null,
|
| 111 |
+
"trial_params": null
|
| 112 |
+
}
|
checkpoint-474/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c864801bcd65938a6635dc9ef295501c2002bc5cda8339db5ff3243768ab40f
|
| 3 |
+
size 5240
|
checkpoint-474/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-711/config.json
ADDED
|
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "bert-base-uncased",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"BertForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"gradient_checkpointing": false,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"id2label": {
|
| 13 |
+
"0": "LABEL_0",
|
| 14 |
+
"1": "LABEL_1",
|
| 15 |
+
"2": "LABEL_2",
|
| 16 |
+
"3": "LABEL_3",
|
| 17 |
+
"4": "LABEL_4",
|
| 18 |
+
"5": "LABEL_5",
|
| 19 |
+
"6": "LABEL_6",
|
| 20 |
+
"7": "LABEL_7",
|
| 21 |
+
"8": "LABEL_8",
|
| 22 |
+
"9": "LABEL_9",
|
| 23 |
+
"10": "LABEL_10",
|
| 24 |
+
"11": "LABEL_11",
|
| 25 |
+
"12": "LABEL_12",
|
| 26 |
+
"13": "LABEL_13",
|
| 27 |
+
"14": "LABEL_14",
|
| 28 |
+
"15": "LABEL_15",
|
| 29 |
+
"16": "LABEL_16",
|
| 30 |
+
"17": "LABEL_17",
|
| 31 |
+
"18": "LABEL_18",
|
| 32 |
+
"19": "LABEL_19",
|
| 33 |
+
"20": "LABEL_20",
|
| 34 |
+
"21": "LABEL_21",
|
| 35 |
+
"22": "LABEL_22",
|
| 36 |
+
"23": "LABEL_23",
|
| 37 |
+
"24": "LABEL_24",
|
| 38 |
+
"25": "LABEL_25",
|
| 39 |
+
"26": "LABEL_26",
|
| 40 |
+
"27": "LABEL_27",
|
| 41 |
+
"28": "LABEL_28",
|
| 42 |
+
"29": "LABEL_29",
|
| 43 |
+
"30": "LABEL_30",
|
| 44 |
+
"31": "LABEL_31",
|
| 45 |
+
"32": "LABEL_32",
|
| 46 |
+
"33": "LABEL_33",
|
| 47 |
+
"34": "LABEL_34",
|
| 48 |
+
"35": "LABEL_35",
|
| 49 |
+
"36": "LABEL_36",
|
| 50 |
+
"37": "LABEL_37",
|
| 51 |
+
"38": "LABEL_38",
|
| 52 |
+
"39": "LABEL_39",
|
| 53 |
+
"40": "LABEL_40",
|
| 54 |
+
"41": "LABEL_41",
|
| 55 |
+
"42": "LABEL_42",
|
| 56 |
+
"43": "LABEL_43",
|
| 57 |
+
"44": "LABEL_44",
|
| 58 |
+
"45": "LABEL_45",
|
| 59 |
+
"46": "LABEL_46",
|
| 60 |
+
"47": "LABEL_47",
|
| 61 |
+
"48": "LABEL_48",
|
| 62 |
+
"49": "LABEL_49",
|
| 63 |
+
"50": "LABEL_50",
|
| 64 |
+
"51": "LABEL_51",
|
| 65 |
+
"52": "LABEL_52",
|
| 66 |
+
"53": "LABEL_53",
|
| 67 |
+
"54": "LABEL_54",
|
| 68 |
+
"55": "LABEL_55",
|
| 69 |
+
"56": "LABEL_56",
|
| 70 |
+
"57": "LABEL_57",
|
| 71 |
+
"58": "LABEL_58"
|
| 72 |
+
},
|
| 73 |
+
"initializer_range": 0.02,
|
| 74 |
+
"intermediate_size": 3072,
|
| 75 |
+
"label2id": {
|
| 76 |
+
"LABEL_0": 0,
|
| 77 |
+
"LABEL_1": 1,
|
| 78 |
+
"LABEL_10": 10,
|
| 79 |
+
"LABEL_11": 11,
|
| 80 |
+
"LABEL_12": 12,
|
| 81 |
+
"LABEL_13": 13,
|
| 82 |
+
"LABEL_14": 14,
|
| 83 |
+
"LABEL_15": 15,
|
| 84 |
+
"LABEL_16": 16,
|
| 85 |
+
"LABEL_17": 17,
|
| 86 |
+
"LABEL_18": 18,
|
| 87 |
+
"LABEL_19": 19,
|
| 88 |
+
"LABEL_2": 2,
|
| 89 |
+
"LABEL_20": 20,
|
| 90 |
+
"LABEL_21": 21,
|
| 91 |
+
"LABEL_22": 22,
|
| 92 |
+
"LABEL_23": 23,
|
| 93 |
+
"LABEL_24": 24,
|
| 94 |
+
"LABEL_25": 25,
|
| 95 |
+
"LABEL_26": 26,
|
| 96 |
+
"LABEL_27": 27,
|
| 97 |
+
"LABEL_28": 28,
|
| 98 |
+
"LABEL_29": 29,
|
| 99 |
+
"LABEL_3": 3,
|
| 100 |
+
"LABEL_30": 30,
|
| 101 |
+
"LABEL_31": 31,
|
| 102 |
+
"LABEL_32": 32,
|
| 103 |
+
"LABEL_33": 33,
|
| 104 |
+
"LABEL_34": 34,
|
| 105 |
+
"LABEL_35": 35,
|
| 106 |
+
"LABEL_36": 36,
|
| 107 |
+
"LABEL_37": 37,
|
| 108 |
+
"LABEL_38": 38,
|
| 109 |
+
"LABEL_39": 39,
|
| 110 |
+
"LABEL_4": 4,
|
| 111 |
+
"LABEL_40": 40,
|
| 112 |
+
"LABEL_41": 41,
|
| 113 |
+
"LABEL_42": 42,
|
| 114 |
+
"LABEL_43": 43,
|
| 115 |
+
"LABEL_44": 44,
|
| 116 |
+
"LABEL_45": 45,
|
| 117 |
+
"LABEL_46": 46,
|
| 118 |
+
"LABEL_47": 47,
|
| 119 |
+
"LABEL_48": 48,
|
| 120 |
+
"LABEL_49": 49,
|
| 121 |
+
"LABEL_5": 5,
|
| 122 |
+
"LABEL_50": 50,
|
| 123 |
+
"LABEL_51": 51,
|
| 124 |
+
"LABEL_52": 52,
|
| 125 |
+
"LABEL_53": 53,
|
| 126 |
+
"LABEL_54": 54,
|
| 127 |
+
"LABEL_55": 55,
|
| 128 |
+
"LABEL_56": 56,
|
| 129 |
+
"LABEL_57": 57,
|
| 130 |
+
"LABEL_58": 58,
|
| 131 |
+
"LABEL_6": 6,
|
| 132 |
+
"LABEL_7": 7,
|
| 133 |
+
"LABEL_8": 8,
|
| 134 |
+
"LABEL_9": 9
|
| 135 |
+
},
|
| 136 |
+
"layer_norm_eps": 1e-12,
|
| 137 |
+
"max_position_embeddings": 512,
|
| 138 |
+
"model_type": "bert",
|
| 139 |
+
"num_attention_heads": 12,
|
| 140 |
+
"num_hidden_layers": 12,
|
| 141 |
+
"pad_token_id": 0,
|
| 142 |
+
"position_embedding_type": "absolute",
|
| 143 |
+
"problem_type": "single_label_classification",
|
| 144 |
+
"torch_dtype": "float32",
|
| 145 |
+
"transformers_version": "4.47.1",
|
| 146 |
+
"type_vocab_size": 2,
|
| 147 |
+
"use_cache": true,
|
| 148 |
+
"vocab_size": 30522
|
| 149 |
+
}
|
checkpoint-711/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2402e33af9eefa597c98b6ea6e14f017676ee3c07d5bde19b6236a637c988790
|
| 3 |
+
size 438133980
|
checkpoint-711/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9244dba1a359bcdf74db68e199776e53c26a8c6be740932f0e588533fa62d800
|
| 3 |
+
size 876383354
|
checkpoint-711/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:45cdde48674967d19dc253fa32d18dc2598d974a3692de61daf0d594c2e52b61
|
| 3 |
+
size 13990
|
checkpoint-711/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c2b003a63e288fca8333868ee25acf375876bf835114b79379555b12de6dd948
|
| 3 |
+
size 1064
|
checkpoint-711/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
checkpoint-711/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-711/tokenizer_config.json
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": false,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"model_max_length": 512,
|
| 50 |
+
"pad_token": "[PAD]",
|
| 51 |
+
"sep_token": "[SEP]",
|
| 52 |
+
"strip_accents": null,
|
| 53 |
+
"tokenize_chinese_chars": true,
|
| 54 |
+
"tokenizer_class": "BertTokenizer",
|
| 55 |
+
"unk_token": "[UNK]"
|
| 56 |
+
}
|
checkpoint-711/trainer_state.json
ADDED
|
@@ -0,0 +1,155 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 3.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 711,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.2109704641350211,
|
| 13 |
+
"grad_norm": 17.118331909179688,
|
| 14 |
+
"learning_rate": 1.9156118143459917e-05,
|
| 15 |
+
"loss": 4.0557,
|
| 16 |
+
"step": 50
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.4219409282700422,
|
| 20 |
+
"grad_norm": 17.078466415405273,
|
| 21 |
+
"learning_rate": 1.8312236286919833e-05,
|
| 22 |
+
"loss": 4.018,
|
| 23 |
+
"step": 100
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.6329113924050633,
|
| 27 |
+
"grad_norm": 18.140644073486328,
|
| 28 |
+
"learning_rate": 1.746835443037975e-05,
|
| 29 |
+
"loss": 3.9942,
|
| 30 |
+
"step": 150
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.8438818565400844,
|
| 34 |
+
"grad_norm": 14.235407829284668,
|
| 35 |
+
"learning_rate": 1.662447257383966e-05,
|
| 36 |
+
"loss": 3.9127,
|
| 37 |
+
"step": 200
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 1.0,
|
| 41 |
+
"eval_loss": 3.480090379714966,
|
| 42 |
+
"eval_runtime": 253.6051,
|
| 43 |
+
"eval_samples_per_second": 0.469,
|
| 44 |
+
"eval_steps_per_second": 0.237,
|
| 45 |
+
"step": 237
|
| 46 |
+
},
|
| 47 |
+
{
|
| 48 |
+
"epoch": 1.0548523206751055,
|
| 49 |
+
"grad_norm": 14.833487510681152,
|
| 50 |
+
"learning_rate": 1.578059071729958e-05,
|
| 51 |
+
"loss": 3.6803,
|
| 52 |
+
"step": 250
|
| 53 |
+
},
|
| 54 |
+
{
|
| 55 |
+
"epoch": 1.2658227848101267,
|
| 56 |
+
"grad_norm": 18.978660583496094,
|
| 57 |
+
"learning_rate": 1.4936708860759495e-05,
|
| 58 |
+
"loss": 3.4012,
|
| 59 |
+
"step": 300
|
| 60 |
+
},
|
| 61 |
+
{
|
| 62 |
+
"epoch": 1.4767932489451476,
|
| 63 |
+
"grad_norm": 18.297719955444336,
|
| 64 |
+
"learning_rate": 1.4092827004219412e-05,
|
| 65 |
+
"loss": 3.1245,
|
| 66 |
+
"step": 350
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 1.6877637130801688,
|
| 70 |
+
"grad_norm": 16.072620391845703,
|
| 71 |
+
"learning_rate": 1.3248945147679326e-05,
|
| 72 |
+
"loss": 3.122,
|
| 73 |
+
"step": 400
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"epoch": 1.8987341772151898,
|
| 77 |
+
"grad_norm": 14.235777854919434,
|
| 78 |
+
"learning_rate": 1.240506329113924e-05,
|
| 79 |
+
"loss": 2.7533,
|
| 80 |
+
"step": 450
|
| 81 |
+
},
|
| 82 |
+
{
|
| 83 |
+
"epoch": 2.0,
|
| 84 |
+
"eval_loss": 2.393742084503174,
|
| 85 |
+
"eval_runtime": 230.7613,
|
| 86 |
+
"eval_samples_per_second": 0.516,
|
| 87 |
+
"eval_steps_per_second": 0.26,
|
| 88 |
+
"step": 474
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"epoch": 2.109704641350211,
|
| 92 |
+
"grad_norm": 15.517921447753906,
|
| 93 |
+
"learning_rate": 1.1561181434599158e-05,
|
| 94 |
+
"loss": 2.5393,
|
| 95 |
+
"step": 500
|
| 96 |
+
},
|
| 97 |
+
{
|
| 98 |
+
"epoch": 2.320675105485232,
|
| 99 |
+
"grad_norm": 12.792672157287598,
|
| 100 |
+
"learning_rate": 1.0717299578059072e-05,
|
| 101 |
+
"loss": 2.4086,
|
| 102 |
+
"step": 550
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"epoch": 2.5316455696202533,
|
| 106 |
+
"grad_norm": 15.125005722045898,
|
| 107 |
+
"learning_rate": 9.87341772151899e-06,
|
| 108 |
+
"loss": 2.3738,
|
| 109 |
+
"step": 600
|
| 110 |
+
},
|
| 111 |
+
{
|
| 112 |
+
"epoch": 2.742616033755274,
|
| 113 |
+
"grad_norm": 14.103851318359375,
|
| 114 |
+
"learning_rate": 9.029535864978903e-06,
|
| 115 |
+
"loss": 2.3799,
|
| 116 |
+
"step": 650
|
| 117 |
+
},
|
| 118 |
+
{
|
| 119 |
+
"epoch": 2.9535864978902953,
|
| 120 |
+
"grad_norm": 13.69605541229248,
|
| 121 |
+
"learning_rate": 8.18565400843882e-06,
|
| 122 |
+
"loss": 2.0113,
|
| 123 |
+
"step": 700
|
| 124 |
+
},
|
| 125 |
+
{
|
| 126 |
+
"epoch": 3.0,
|
| 127 |
+
"eval_loss": 1.8042149543762207,
|
| 128 |
+
"eval_runtime": 229.2696,
|
| 129 |
+
"eval_samples_per_second": 0.519,
|
| 130 |
+
"eval_steps_per_second": 0.262,
|
| 131 |
+
"step": 711
|
| 132 |
+
}
|
| 133 |
+
],
|
| 134 |
+
"logging_steps": 50,
|
| 135 |
+
"max_steps": 1185,
|
| 136 |
+
"num_input_tokens_seen": 0,
|
| 137 |
+
"num_train_epochs": 5,
|
| 138 |
+
"save_steps": 500,
|
| 139 |
+
"stateful_callbacks": {
|
| 140 |
+
"TrainerControl": {
|
| 141 |
+
"args": {
|
| 142 |
+
"should_epoch_stop": false,
|
| 143 |
+
"should_evaluate": false,
|
| 144 |
+
"should_log": false,
|
| 145 |
+
"should_save": true,
|
| 146 |
+
"should_training_stop": false
|
| 147 |
+
},
|
| 148 |
+
"attributes": {}
|
| 149 |
+
}
|
| 150 |
+
},
|
| 151 |
+
"total_flos": 373545662966784.0,
|
| 152 |
+
"train_batch_size": 2,
|
| 153 |
+
"trial_name": null,
|
| 154 |
+
"trial_params": null
|
| 155 |
+
}
|
checkpoint-711/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c864801bcd65938a6635dc9ef295501c2002bc5cda8339db5ff3243768ab40f
|
| 3 |
+
size 5240
|
checkpoint-711/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-948/config.json
ADDED
|
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "bert-base-uncased",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"BertForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"gradient_checkpointing": false,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"id2label": {
|
| 13 |
+
"0": "LABEL_0",
|
| 14 |
+
"1": "LABEL_1",
|
| 15 |
+
"2": "LABEL_2",
|
| 16 |
+
"3": "LABEL_3",
|
| 17 |
+
"4": "LABEL_4",
|
| 18 |
+
"5": "LABEL_5",
|
| 19 |
+
"6": "LABEL_6",
|
| 20 |
+
"7": "LABEL_7",
|
| 21 |
+
"8": "LABEL_8",
|
| 22 |
+
"9": "LABEL_9",
|
| 23 |
+
"10": "LABEL_10",
|
| 24 |
+
"11": "LABEL_11",
|
| 25 |
+
"12": "LABEL_12",
|
| 26 |
+
"13": "LABEL_13",
|
| 27 |
+
"14": "LABEL_14",
|
| 28 |
+
"15": "LABEL_15",
|
| 29 |
+
"16": "LABEL_16",
|
| 30 |
+
"17": "LABEL_17",
|
| 31 |
+
"18": "LABEL_18",
|
| 32 |
+
"19": "LABEL_19",
|
| 33 |
+
"20": "LABEL_20",
|
| 34 |
+
"21": "LABEL_21",
|
| 35 |
+
"22": "LABEL_22",
|
| 36 |
+
"23": "LABEL_23",
|
| 37 |
+
"24": "LABEL_24",
|
| 38 |
+
"25": "LABEL_25",
|
| 39 |
+
"26": "LABEL_26",
|
| 40 |
+
"27": "LABEL_27",
|
| 41 |
+
"28": "LABEL_28",
|
| 42 |
+
"29": "LABEL_29",
|
| 43 |
+
"30": "LABEL_30",
|
| 44 |
+
"31": "LABEL_31",
|
| 45 |
+
"32": "LABEL_32",
|
| 46 |
+
"33": "LABEL_33",
|
| 47 |
+
"34": "LABEL_34",
|
| 48 |
+
"35": "LABEL_35",
|
| 49 |
+
"36": "LABEL_36",
|
| 50 |
+
"37": "LABEL_37",
|
| 51 |
+
"38": "LABEL_38",
|
| 52 |
+
"39": "LABEL_39",
|
| 53 |
+
"40": "LABEL_40",
|
| 54 |
+
"41": "LABEL_41",
|
| 55 |
+
"42": "LABEL_42",
|
| 56 |
+
"43": "LABEL_43",
|
| 57 |
+
"44": "LABEL_44",
|
| 58 |
+
"45": "LABEL_45",
|
| 59 |
+
"46": "LABEL_46",
|
| 60 |
+
"47": "LABEL_47",
|
| 61 |
+
"48": "LABEL_48",
|
| 62 |
+
"49": "LABEL_49",
|
| 63 |
+
"50": "LABEL_50",
|
| 64 |
+
"51": "LABEL_51",
|
| 65 |
+
"52": "LABEL_52",
|
| 66 |
+
"53": "LABEL_53",
|
| 67 |
+
"54": "LABEL_54",
|
| 68 |
+
"55": "LABEL_55",
|
| 69 |
+
"56": "LABEL_56",
|
| 70 |
+
"57": "LABEL_57",
|
| 71 |
+
"58": "LABEL_58"
|
| 72 |
+
},
|
| 73 |
+
"initializer_range": 0.02,
|
| 74 |
+
"intermediate_size": 3072,
|
| 75 |
+
"label2id": {
|
| 76 |
+
"LABEL_0": 0,
|
| 77 |
+
"LABEL_1": 1,
|
| 78 |
+
"LABEL_10": 10,
|
| 79 |
+
"LABEL_11": 11,
|
| 80 |
+
"LABEL_12": 12,
|
| 81 |
+
"LABEL_13": 13,
|
| 82 |
+
"LABEL_14": 14,
|
| 83 |
+
"LABEL_15": 15,
|
| 84 |
+
"LABEL_16": 16,
|
| 85 |
+
"LABEL_17": 17,
|
| 86 |
+
"LABEL_18": 18,
|
| 87 |
+
"LABEL_19": 19,
|
| 88 |
+
"LABEL_2": 2,
|
| 89 |
+
"LABEL_20": 20,
|
| 90 |
+
"LABEL_21": 21,
|
| 91 |
+
"LABEL_22": 22,
|
| 92 |
+
"LABEL_23": 23,
|
| 93 |
+
"LABEL_24": 24,
|
| 94 |
+
"LABEL_25": 25,
|
| 95 |
+
"LABEL_26": 26,
|
| 96 |
+
"LABEL_27": 27,
|
| 97 |
+
"LABEL_28": 28,
|
| 98 |
+
"LABEL_29": 29,
|
| 99 |
+
"LABEL_3": 3,
|
| 100 |
+
"LABEL_30": 30,
|
| 101 |
+
"LABEL_31": 31,
|
| 102 |
+
"LABEL_32": 32,
|
| 103 |
+
"LABEL_33": 33,
|
| 104 |
+
"LABEL_34": 34,
|
| 105 |
+
"LABEL_35": 35,
|
| 106 |
+
"LABEL_36": 36,
|
| 107 |
+
"LABEL_37": 37,
|
| 108 |
+
"LABEL_38": 38,
|
| 109 |
+
"LABEL_39": 39,
|
| 110 |
+
"LABEL_4": 4,
|
| 111 |
+
"LABEL_40": 40,
|
| 112 |
+
"LABEL_41": 41,
|
| 113 |
+
"LABEL_42": 42,
|
| 114 |
+
"LABEL_43": 43,
|
| 115 |
+
"LABEL_44": 44,
|
| 116 |
+
"LABEL_45": 45,
|
| 117 |
+
"LABEL_46": 46,
|
| 118 |
+
"LABEL_47": 47,
|
| 119 |
+
"LABEL_48": 48,
|
| 120 |
+
"LABEL_49": 49,
|
| 121 |
+
"LABEL_5": 5,
|
| 122 |
+
"LABEL_50": 50,
|
| 123 |
+
"LABEL_51": 51,
|
| 124 |
+
"LABEL_52": 52,
|
| 125 |
+
"LABEL_53": 53,
|
| 126 |
+
"LABEL_54": 54,
|
| 127 |
+
"LABEL_55": 55,
|
| 128 |
+
"LABEL_56": 56,
|
| 129 |
+
"LABEL_57": 57,
|
| 130 |
+
"LABEL_58": 58,
|
| 131 |
+
"LABEL_6": 6,
|
| 132 |
+
"LABEL_7": 7,
|
| 133 |
+
"LABEL_8": 8,
|
| 134 |
+
"LABEL_9": 9
|
| 135 |
+
},
|
| 136 |
+
"layer_norm_eps": 1e-12,
|
| 137 |
+
"max_position_embeddings": 512,
|
| 138 |
+
"model_type": "bert",
|
| 139 |
+
"num_attention_heads": 12,
|
| 140 |
+
"num_hidden_layers": 12,
|
| 141 |
+
"pad_token_id": 0,
|
| 142 |
+
"position_embedding_type": "absolute",
|
| 143 |
+
"problem_type": "single_label_classification",
|
| 144 |
+
"torch_dtype": "float32",
|
| 145 |
+
"transformers_version": "4.47.1",
|
| 146 |
+
"type_vocab_size": 2,
|
| 147 |
+
"use_cache": true,
|
| 148 |
+
"vocab_size": 30522
|
| 149 |
+
}
|
checkpoint-948/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f2d6af8c7c51f1752fc7cf9179ffe579f68c9822fe9c0b4cdfeae91b98a55c5
|
| 3 |
+
size 438133980
|
checkpoint-948/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f04f91cbe88376dd48c769222d9e827667fb9e17ddf9a7d471b3149bf487d09b
|
| 3 |
+
size 876383354
|
checkpoint-948/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ace1ae70ed3086e737c887e839b0768fa97104ac3fcc4aaaf0b060c61d1cba45
|
| 3 |
+
size 13990
|
checkpoint-948/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0783dd16df486c722a3ebe892adbfc984d3c78d06305f35ef88f1a6417c97a03
|
| 3 |
+
size 1064
|
checkpoint-948/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|