davidfred commited on
Commit
8954431
·
verified ·
1 Parent(s): e473856

Upload folder using huggingface_hub

Browse files
checkpoint-700/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 512,
22
+ "model_type": "bert",
23
+ "num_attention_heads": 12,
24
+ "num_hidden_layers": 12,
25
+ "pad_token_id": 0,
26
+ "position_embedding_type": "absolute",
27
+ "problem_type": "regression",
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.46.2",
30
+ "type_vocab_size": 2,
31
+ "use_cache": true,
32
+ "vocab_size": 30522
33
+ }
checkpoint-700/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6360c6a9127342f0bf83a6b2b8b9742c6a1d60eebb0de956649de38d5fd5d666
3
+ size 437955572
checkpoint-700/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37a1b16efb8234d643ef0984060cdb3aa1da309fcf2ea774720cc39735f9278e
3
+ size 876032250
checkpoint-700/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8142e7505788be9614fece6d9e7494673947881136ae0f4b3862a3093331437
3
+ size 14244
checkpoint-700/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21403d5a03b1fcf9048c6478c3efaf09d7b8ea4d82cd726bd9f2e78d5628ac37
3
+ size 1064
checkpoint-700/trainer_state.json ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 45.52845528455285,
5
+ "eval_steps": 100,
6
+ "global_step": 700,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 3.252032520325203,
13
+ "grad_norm": 146.2987823486328,
14
+ "learning_rate": 2.25e-05,
15
+ "loss": 3.7155,
16
+ "step": 50
17
+ },
18
+ {
19
+ "epoch": 6.504065040650406,
20
+ "grad_norm": 25.606613159179688,
21
+ "learning_rate": 4.75e-05,
22
+ "loss": 0.3122,
23
+ "step": 100
24
+ },
25
+ {
26
+ "epoch": 6.504065040650406,
27
+ "eval_loss": 0.06210774928331375,
28
+ "eval_runtime": 0.4632,
29
+ "eval_samples_per_second": 118.733,
30
+ "eval_steps_per_second": 30.223,
31
+ "step": 100
32
+ },
33
+ {
34
+ "epoch": 9.75609756097561,
35
+ "grad_norm": 70.79466247558594,
36
+ "learning_rate": 4.653846153846154e-05,
37
+ "loss": 0.0857,
38
+ "step": 150
39
+ },
40
+ {
41
+ "epoch": 13.008130081300813,
42
+ "grad_norm": 24.777908325195312,
43
+ "learning_rate": 4.269230769230769e-05,
44
+ "loss": 0.0344,
45
+ "step": 200
46
+ },
47
+ {
48
+ "epoch": 13.008130081300813,
49
+ "eval_loss": 0.02328495681285858,
50
+ "eval_runtime": 0.4673,
51
+ "eval_samples_per_second": 117.696,
52
+ "eval_steps_per_second": 29.959,
53
+ "step": 200
54
+ },
55
+ {
56
+ "epoch": 16.260162601626018,
57
+ "grad_norm": 8.731575012207031,
58
+ "learning_rate": 3.884615384615385e-05,
59
+ "loss": 0.0301,
60
+ "step": 250
61
+ },
62
+ {
63
+ "epoch": 19.51219512195122,
64
+ "grad_norm": 4.41317081451416,
65
+ "learning_rate": 3.5e-05,
66
+ "loss": 0.0213,
67
+ "step": 300
68
+ },
69
+ {
70
+ "epoch": 19.51219512195122,
71
+ "eval_loss": 0.021754104644060135,
72
+ "eval_runtime": 0.4593,
73
+ "eval_samples_per_second": 119.74,
74
+ "eval_steps_per_second": 30.479,
75
+ "step": 300
76
+ },
77
+ {
78
+ "epoch": 22.764227642276424,
79
+ "grad_norm": 4.309276580810547,
80
+ "learning_rate": 3.115384615384615e-05,
81
+ "loss": 0.0173,
82
+ "step": 350
83
+ },
84
+ {
85
+ "epoch": 26.016260162601625,
86
+ "grad_norm": 4.0787177085876465,
87
+ "learning_rate": 2.7307692307692305e-05,
88
+ "loss": 0.0133,
89
+ "step": 400
90
+ },
91
+ {
92
+ "epoch": 26.016260162601625,
93
+ "eval_loss": 0.007923820987343788,
94
+ "eval_runtime": 0.466,
95
+ "eval_samples_per_second": 118.03,
96
+ "eval_steps_per_second": 30.044,
97
+ "step": 400
98
+ },
99
+ {
100
+ "epoch": 29.26829268292683,
101
+ "grad_norm": 2.8265419006347656,
102
+ "learning_rate": 2.3461538461538464e-05,
103
+ "loss": 0.0128,
104
+ "step": 450
105
+ },
106
+ {
107
+ "epoch": 32.520325203252035,
108
+ "grad_norm": 3.378178358078003,
109
+ "learning_rate": 1.9615384615384617e-05,
110
+ "loss": 0.0106,
111
+ "step": 500
112
+ },
113
+ {
114
+ "epoch": 32.520325203252035,
115
+ "eval_loss": 0.009708667173981667,
116
+ "eval_runtime": 0.4796,
117
+ "eval_samples_per_second": 114.679,
118
+ "eval_steps_per_second": 29.191,
119
+ "step": 500
120
+ },
121
+ {
122
+ "epoch": 35.77235772357724,
123
+ "grad_norm": 8.75239086151123,
124
+ "learning_rate": 1.576923076923077e-05,
125
+ "loss": 0.0084,
126
+ "step": 550
127
+ },
128
+ {
129
+ "epoch": 39.02439024390244,
130
+ "grad_norm": 6.199377059936523,
131
+ "learning_rate": 1.1923076923076925e-05,
132
+ "loss": 0.0077,
133
+ "step": 600
134
+ },
135
+ {
136
+ "epoch": 39.02439024390244,
137
+ "eval_loss": 0.010460535995662212,
138
+ "eval_runtime": 0.456,
139
+ "eval_samples_per_second": 120.603,
140
+ "eval_steps_per_second": 30.699,
141
+ "step": 600
142
+ },
143
+ {
144
+ "epoch": 42.27642276422764,
145
+ "grad_norm": 4.703422546386719,
146
+ "learning_rate": 8.076923076923077e-06,
147
+ "loss": 0.0065,
148
+ "step": 650
149
+ },
150
+ {
151
+ "epoch": 45.52845528455285,
152
+ "grad_norm": 3.5736358165740967,
153
+ "learning_rate": 4.230769230769231e-06,
154
+ "loss": 0.0054,
155
+ "step": 700
156
+ },
157
+ {
158
+ "epoch": 45.52845528455285,
159
+ "eval_loss": 0.00792544987052679,
160
+ "eval_runtime": 0.4594,
161
+ "eval_samples_per_second": 119.722,
162
+ "eval_steps_per_second": 30.475,
163
+ "step": 700
164
+ }
165
+ ],
166
+ "logging_steps": 50,
167
+ "max_steps": 750,
168
+ "num_input_tokens_seen": 0,
169
+ "num_train_epochs": 50,
170
+ "save_steps": 100,
171
+ "stateful_callbacks": {
172
+ "TrainerControl": {
173
+ "args": {
174
+ "should_epoch_stop": false,
175
+ "should_evaluate": false,
176
+ "should_log": false,
177
+ "should_save": true,
178
+ "should_training_stop": false
179
+ },
180
+ "attributes": {}
181
+ }
182
+ },
183
+ "total_flos": 5881794831836160.0,
184
+ "train_batch_size": 4,
185
+ "trial_name": null,
186
+ "trial_params": null
187
+ }
checkpoint-700/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5d456a9b82954adbf560ec2a0a497c07cc203010e4d45e32fef528674631e04
3
+ size 5240
checkpoint-750/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 512,
22
+ "model_type": "bert",
23
+ "num_attention_heads": 12,
24
+ "num_hidden_layers": 12,
25
+ "pad_token_id": 0,
26
+ "position_embedding_type": "absolute",
27
+ "problem_type": "regression",
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.46.2",
30
+ "type_vocab_size": 2,
31
+ "use_cache": true,
32
+ "vocab_size": 30522
33
+ }
checkpoint-750/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c6e5ccc9edbaa8f1faa4afa4f758aa85112fb842c4155822267ed720f1139ce
3
+ size 437955572
checkpoint-750/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8ae4e2e485c256840d46661719ca0ecab3a1d96b6ec7058baf0a3205731a577
3
+ size 876032250
checkpoint-750/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b45a236ad6cc86e8d7a8a2b4c917dc53a8af640c7e1770cf245a9de3869566d6
3
+ size 14244
checkpoint-750/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f29d69aec3ca2fcf9f341b05f5651afb3339d2916402e7b67843eda3443defa
3
+ size 1064
checkpoint-750/trainer_state.json ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 48.78048780487805,
5
+ "eval_steps": 100,
6
+ "global_step": 750,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 3.252032520325203,
13
+ "grad_norm": 146.2987823486328,
14
+ "learning_rate": 2.25e-05,
15
+ "loss": 3.7155,
16
+ "step": 50
17
+ },
18
+ {
19
+ "epoch": 6.504065040650406,
20
+ "grad_norm": 25.606613159179688,
21
+ "learning_rate": 4.75e-05,
22
+ "loss": 0.3122,
23
+ "step": 100
24
+ },
25
+ {
26
+ "epoch": 6.504065040650406,
27
+ "eval_loss": 0.06210774928331375,
28
+ "eval_runtime": 0.4632,
29
+ "eval_samples_per_second": 118.733,
30
+ "eval_steps_per_second": 30.223,
31
+ "step": 100
32
+ },
33
+ {
34
+ "epoch": 9.75609756097561,
35
+ "grad_norm": 70.79466247558594,
36
+ "learning_rate": 4.653846153846154e-05,
37
+ "loss": 0.0857,
38
+ "step": 150
39
+ },
40
+ {
41
+ "epoch": 13.008130081300813,
42
+ "grad_norm": 24.777908325195312,
43
+ "learning_rate": 4.269230769230769e-05,
44
+ "loss": 0.0344,
45
+ "step": 200
46
+ },
47
+ {
48
+ "epoch": 13.008130081300813,
49
+ "eval_loss": 0.02328495681285858,
50
+ "eval_runtime": 0.4673,
51
+ "eval_samples_per_second": 117.696,
52
+ "eval_steps_per_second": 29.959,
53
+ "step": 200
54
+ },
55
+ {
56
+ "epoch": 16.260162601626018,
57
+ "grad_norm": 8.731575012207031,
58
+ "learning_rate": 3.884615384615385e-05,
59
+ "loss": 0.0301,
60
+ "step": 250
61
+ },
62
+ {
63
+ "epoch": 19.51219512195122,
64
+ "grad_norm": 4.41317081451416,
65
+ "learning_rate": 3.5e-05,
66
+ "loss": 0.0213,
67
+ "step": 300
68
+ },
69
+ {
70
+ "epoch": 19.51219512195122,
71
+ "eval_loss": 0.021754104644060135,
72
+ "eval_runtime": 0.4593,
73
+ "eval_samples_per_second": 119.74,
74
+ "eval_steps_per_second": 30.479,
75
+ "step": 300
76
+ },
77
+ {
78
+ "epoch": 22.764227642276424,
79
+ "grad_norm": 4.309276580810547,
80
+ "learning_rate": 3.115384615384615e-05,
81
+ "loss": 0.0173,
82
+ "step": 350
83
+ },
84
+ {
85
+ "epoch": 26.016260162601625,
86
+ "grad_norm": 4.0787177085876465,
87
+ "learning_rate": 2.7307692307692305e-05,
88
+ "loss": 0.0133,
89
+ "step": 400
90
+ },
91
+ {
92
+ "epoch": 26.016260162601625,
93
+ "eval_loss": 0.007923820987343788,
94
+ "eval_runtime": 0.466,
95
+ "eval_samples_per_second": 118.03,
96
+ "eval_steps_per_second": 30.044,
97
+ "step": 400
98
+ },
99
+ {
100
+ "epoch": 29.26829268292683,
101
+ "grad_norm": 2.8265419006347656,
102
+ "learning_rate": 2.3461538461538464e-05,
103
+ "loss": 0.0128,
104
+ "step": 450
105
+ },
106
+ {
107
+ "epoch": 32.520325203252035,
108
+ "grad_norm": 3.378178358078003,
109
+ "learning_rate": 1.9615384615384617e-05,
110
+ "loss": 0.0106,
111
+ "step": 500
112
+ },
113
+ {
114
+ "epoch": 32.520325203252035,
115
+ "eval_loss": 0.009708667173981667,
116
+ "eval_runtime": 0.4796,
117
+ "eval_samples_per_second": 114.679,
118
+ "eval_steps_per_second": 29.191,
119
+ "step": 500
120
+ },
121
+ {
122
+ "epoch": 35.77235772357724,
123
+ "grad_norm": 8.75239086151123,
124
+ "learning_rate": 1.576923076923077e-05,
125
+ "loss": 0.0084,
126
+ "step": 550
127
+ },
128
+ {
129
+ "epoch": 39.02439024390244,
130
+ "grad_norm": 6.199377059936523,
131
+ "learning_rate": 1.1923076923076925e-05,
132
+ "loss": 0.0077,
133
+ "step": 600
134
+ },
135
+ {
136
+ "epoch": 39.02439024390244,
137
+ "eval_loss": 0.010460535995662212,
138
+ "eval_runtime": 0.456,
139
+ "eval_samples_per_second": 120.603,
140
+ "eval_steps_per_second": 30.699,
141
+ "step": 600
142
+ },
143
+ {
144
+ "epoch": 42.27642276422764,
145
+ "grad_norm": 4.703422546386719,
146
+ "learning_rate": 8.076923076923077e-06,
147
+ "loss": 0.0065,
148
+ "step": 650
149
+ },
150
+ {
151
+ "epoch": 45.52845528455285,
152
+ "grad_norm": 3.5736358165740967,
153
+ "learning_rate": 4.230769230769231e-06,
154
+ "loss": 0.0054,
155
+ "step": 700
156
+ },
157
+ {
158
+ "epoch": 45.52845528455285,
159
+ "eval_loss": 0.00792544987052679,
160
+ "eval_runtime": 0.4594,
161
+ "eval_samples_per_second": 119.722,
162
+ "eval_steps_per_second": 30.475,
163
+ "step": 700
164
+ },
165
+ {
166
+ "epoch": 48.78048780487805,
167
+ "grad_norm": 3.0123989582061768,
168
+ "learning_rate": 3.846153846153847e-07,
169
+ "loss": 0.0049,
170
+ "step": 750
171
+ }
172
+ ],
173
+ "logging_steps": 50,
174
+ "max_steps": 750,
175
+ "num_input_tokens_seen": 0,
176
+ "num_train_epochs": 50,
177
+ "save_steps": 100,
178
+ "stateful_callbacks": {
179
+ "TrainerControl": {
180
+ "args": {
181
+ "should_epoch_stop": false,
182
+ "should_evaluate": false,
183
+ "should_log": false,
184
+ "should_save": true,
185
+ "should_training_stop": true
186
+ },
187
+ "attributes": {}
188
+ }
189
+ },
190
+ "total_flos": 6301979414544384.0,
191
+ "train_batch_size": 4,
192
+ "trial_name": null,
193
+ "trial_params": null
194
+ }
checkpoint-750/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5d456a9b82954adbf560ec2a0a497c07cc203010e4d45e32fef528674631e04
3
+ size 5240
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 512,
22
+ "model_type": "bert",
23
+ "num_attention_heads": 12,
24
+ "num_hidden_layers": 12,
25
+ "pad_token_id": 0,
26
+ "position_embedding_type": "absolute",
27
+ "problem_type": "regression",
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.46.2",
30
+ "type_vocab_size": 2,
31
+ "use_cache": true,
32
+ "vocab_size": 30522
33
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c6e5ccc9edbaa8f1faa4afa4f758aa85112fb842c4155822267ed720f1139ce
3
+ size 437955572
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "BertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff