Hank1 commited on
Commit
83583de
·
verified ·
1 Parent(s): a2c80d3

Upload 12 files

Browse files
deberta-mnli/checkpoint-24544/config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/ibex/user/anb/Model_merging/EMR_Merging-main/merge_lm/ckpts/deberta/deberta-base",
3
+ "architectures": [
4
+ "DebertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "id2label": {
11
+ "0": "LABEL_0",
12
+ "1": "LABEL_1",
13
+ "2": "LABEL_2"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0,
19
+ "LABEL_1": 1,
20
+ "LABEL_2": 2
21
+ },
22
+ "layer_norm_eps": 1e-07,
23
+ "max_position_embeddings": 512,
24
+ "max_relative_positions": -1,
25
+ "model_type": "deberta",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "pad_token_id": 0,
29
+ "pooler_dropout": 0,
30
+ "pooler_hidden_act": "gelu",
31
+ "pooler_hidden_size": 768,
32
+ "pos_att_type": [
33
+ "c2p",
34
+ "p2c"
35
+ ],
36
+ "position_biased_input": false,
37
+ "relative_attention": true,
38
+ "torch_dtype": "float32",
39
+ "transformers_version": "4.45.1",
40
+ "type_vocab_size": 0,
41
+ "vocab_size": 50265
42
+ }
deberta-mnli/checkpoint-24544/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
deberta-mnli/checkpoint-24544/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:298d07f2f91fc467c0d00f92328adebef199c2f8b33b2e25f858d8dc0239edc5
3
+ size 556802644
deberta-mnli/checkpoint-24544/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b17778356ce7a420024428acc45bda191ba15abe5d8e2469445e95f9d1d6170b
3
+ size 1113724282
deberta-mnli/checkpoint-24544/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36a2571db718d633322af805463b0b2faf2e475ff1ca9e040162e39bf99ac6dc
3
+ size 14244
deberta-mnli/checkpoint-24544/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60e55f3aecf8b8786448fdb441e26658f605e89258d669fdd92bd1383ac19ec2
3
+ size 1064
deberta-mnli/checkpoint-24544/special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": true,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
deberta-mnli/checkpoint-24544/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
deberta-mnli/checkpoint-24544/tokenizer_config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "[PAD]",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "[CLS]",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "[SEP]",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "[UNK]",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "50264": {
38
+ "content": "[MASK]",
39
+ "lstrip": true,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ }
45
+ },
46
+ "bos_token": "[CLS]",
47
+ "clean_up_tokenization_spaces": false,
48
+ "cls_token": "[CLS]",
49
+ "do_lower_case": false,
50
+ "eos_token": "[SEP]",
51
+ "errors": "replace",
52
+ "mask_token": "[MASK]",
53
+ "model_max_length": 1000000000000000019884624838656,
54
+ "pad_token": "[PAD]",
55
+ "sep_token": "[SEP]",
56
+ "tokenizer_class": "DebertaTokenizer",
57
+ "unk_token": "[UNK]",
58
+ "vocab_type": "gpt2"
59
+ }
deberta-mnli/checkpoint-24544/trainer_state.json ADDED
@@ -0,0 +1,412 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8852776362710137,
3
+ "best_model_checkpoint": "./deberta/deberta-mnli/checkpoint-24544",
4
+ "epoch": 4.0,
5
+ "eval_steps": 500,
6
+ "global_step": 24544,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.08148631029986962,
13
+ "grad_norm": 9.7207612991333,
14
+ "learning_rate": 1.9738391105297582e-05,
15
+ "loss": 0.7226,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.16297262059973924,
20
+ "grad_norm": 6.806437015533447,
21
+ "learning_rate": 1.9411379986919555e-05,
22
+ "loss": 0.4742,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.24445893089960888,
27
+ "grad_norm": 6.399604320526123,
28
+ "learning_rate": 1.9084368868541532e-05,
29
+ "loss": 0.4375,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 0.3259452411994785,
34
+ "grad_norm": 6.315491676330566,
35
+ "learning_rate": 1.8757357750163508e-05,
36
+ "loss": 0.4139,
37
+ "step": 2000
38
+ },
39
+ {
40
+ "epoch": 0.4074315514993481,
41
+ "grad_norm": 5.844660758972168,
42
+ "learning_rate": 1.843034663178548e-05,
43
+ "loss": 0.4046,
44
+ "step": 2500
45
+ },
46
+ {
47
+ "epoch": 0.48891786179921776,
48
+ "grad_norm": 3.927537679672241,
49
+ "learning_rate": 1.8103335513407458e-05,
50
+ "loss": 0.3975,
51
+ "step": 3000
52
+ },
53
+ {
54
+ "epoch": 0.5704041720990873,
55
+ "grad_norm": 3.775937557220459,
56
+ "learning_rate": 1.777632439502943e-05,
57
+ "loss": 0.3827,
58
+ "step": 3500
59
+ },
60
+ {
61
+ "epoch": 0.651890482398957,
62
+ "grad_norm": 3.8927953243255615,
63
+ "learning_rate": 1.7449313276651408e-05,
64
+ "loss": 0.3781,
65
+ "step": 4000
66
+ },
67
+ {
68
+ "epoch": 0.7333767926988266,
69
+ "grad_norm": 6.637417316436768,
70
+ "learning_rate": 1.7122302158273384e-05,
71
+ "loss": 0.3783,
72
+ "step": 4500
73
+ },
74
+ {
75
+ "epoch": 0.8148631029986962,
76
+ "grad_norm": 6.947172164916992,
77
+ "learning_rate": 1.6795291039895357e-05,
78
+ "loss": 0.3649,
79
+ "step": 5000
80
+ },
81
+ {
82
+ "epoch": 0.8963494132985659,
83
+ "grad_norm": 7.5690531730651855,
84
+ "learning_rate": 1.6468279921517334e-05,
85
+ "loss": 0.3669,
86
+ "step": 5500
87
+ },
88
+ {
89
+ "epoch": 0.9778357235984355,
90
+ "grad_norm": 6.342957496643066,
91
+ "learning_rate": 1.6141268803139307e-05,
92
+ "loss": 0.3627,
93
+ "step": 6000
94
+ },
95
+ {
96
+ "epoch": 1.0,
97
+ "eval_accuracy": 0.8798777381558839,
98
+ "eval_loss": 0.32735681533813477,
99
+ "eval_runtime": 11.0908,
100
+ "eval_samples_per_second": 884.971,
101
+ "eval_steps_per_second": 13.885,
102
+ "step": 6136
103
+ },
104
+ {
105
+ "epoch": 1.0593220338983051,
106
+ "grad_norm": 4.847009658813477,
107
+ "learning_rate": 1.5814257684761283e-05,
108
+ "loss": 0.2928,
109
+ "step": 6500
110
+ },
111
+ {
112
+ "epoch": 1.1408083441981747,
113
+ "grad_norm": 7.34955358505249,
114
+ "learning_rate": 1.548724656638326e-05,
115
+ "loss": 0.2716,
116
+ "step": 7000
117
+ },
118
+ {
119
+ "epoch": 1.2222946544980444,
120
+ "grad_norm": 3.077202796936035,
121
+ "learning_rate": 1.5160235448005235e-05,
122
+ "loss": 0.2784,
123
+ "step": 7500
124
+ },
125
+ {
126
+ "epoch": 1.303780964797914,
127
+ "grad_norm": 4.981752872467041,
128
+ "learning_rate": 1.483322432962721e-05,
129
+ "loss": 0.2747,
130
+ "step": 8000
131
+ },
132
+ {
133
+ "epoch": 1.3852672750977835,
134
+ "grad_norm": 4.702407360076904,
135
+ "learning_rate": 1.4506213211249183e-05,
136
+ "loss": 0.2714,
137
+ "step": 8500
138
+ },
139
+ {
140
+ "epoch": 1.4667535853976532,
141
+ "grad_norm": 5.884511947631836,
142
+ "learning_rate": 1.4179202092871159e-05,
143
+ "loss": 0.2741,
144
+ "step": 9000
145
+ },
146
+ {
147
+ "epoch": 1.548239895697523,
148
+ "grad_norm": 7.298000335693359,
149
+ "learning_rate": 1.3852190974493134e-05,
150
+ "loss": 0.2727,
151
+ "step": 9500
152
+ },
153
+ {
154
+ "epoch": 1.6297262059973925,
155
+ "grad_norm": 5.3628339767456055,
156
+ "learning_rate": 1.3525179856115109e-05,
157
+ "loss": 0.2711,
158
+ "step": 10000
159
+ },
160
+ {
161
+ "epoch": 1.711212516297262,
162
+ "grad_norm": 5.2779765129089355,
163
+ "learning_rate": 1.3198168737737085e-05,
164
+ "loss": 0.2709,
165
+ "step": 10500
166
+ },
167
+ {
168
+ "epoch": 1.7926988265971318,
169
+ "grad_norm": 8.934818267822266,
170
+ "learning_rate": 1.2871157619359058e-05,
171
+ "loss": 0.2713,
172
+ "step": 11000
173
+ },
174
+ {
175
+ "epoch": 1.8741851368970013,
176
+ "grad_norm": 4.539167881011963,
177
+ "learning_rate": 1.2544146500981033e-05,
178
+ "loss": 0.2696,
179
+ "step": 11500
180
+ },
181
+ {
182
+ "epoch": 1.9556714471968708,
183
+ "grad_norm": 4.068707466125488,
184
+ "learning_rate": 1.221713538260301e-05,
185
+ "loss": 0.2698,
186
+ "step": 12000
187
+ },
188
+ {
189
+ "epoch": 2.0,
190
+ "eval_accuracy": 0.8799796230259807,
191
+ "eval_loss": 0.34052515029907227,
192
+ "eval_runtime": 11.0931,
193
+ "eval_samples_per_second": 884.781,
194
+ "eval_steps_per_second": 13.882,
195
+ "step": 12272
196
+ },
197
+ {
198
+ "epoch": 2.0371577574967406,
199
+ "grad_norm": 5.666093826293945,
200
+ "learning_rate": 1.1890124264224984e-05,
201
+ "loss": 0.2327,
202
+ "step": 12500
203
+ },
204
+ {
205
+ "epoch": 2.1186440677966103,
206
+ "grad_norm": 5.032351970672607,
207
+ "learning_rate": 1.1563113145846961e-05,
208
+ "loss": 0.1798,
209
+ "step": 13000
210
+ },
211
+ {
212
+ "epoch": 2.2001303780964796,
213
+ "grad_norm": 4.264206886291504,
214
+ "learning_rate": 1.1236102027468936e-05,
215
+ "loss": 0.1905,
216
+ "step": 13500
217
+ },
218
+ {
219
+ "epoch": 2.2816166883963493,
220
+ "grad_norm": 7.420534133911133,
221
+ "learning_rate": 1.0909090909090909e-05,
222
+ "loss": 0.1908,
223
+ "step": 14000
224
+ },
225
+ {
226
+ "epoch": 2.363102998696219,
227
+ "grad_norm": 7.359429836273193,
228
+ "learning_rate": 1.0582079790712885e-05,
229
+ "loss": 0.1843,
230
+ "step": 14500
231
+ },
232
+ {
233
+ "epoch": 2.444589308996089,
234
+ "grad_norm": 7.396298408508301,
235
+ "learning_rate": 1.025506867233486e-05,
236
+ "loss": 0.1848,
237
+ "step": 15000
238
+ },
239
+ {
240
+ "epoch": 2.526075619295958,
241
+ "grad_norm": 8.213353157043457,
242
+ "learning_rate": 9.928057553956835e-06,
243
+ "loss": 0.1885,
244
+ "step": 15500
245
+ },
246
+ {
247
+ "epoch": 2.607561929595828,
248
+ "grad_norm": 3.7509984970092773,
249
+ "learning_rate": 9.60104643557881e-06,
250
+ "loss": 0.1869,
251
+ "step": 16000
252
+ },
253
+ {
254
+ "epoch": 2.6890482398956976,
255
+ "grad_norm": 5.412872791290283,
256
+ "learning_rate": 9.274035317200786e-06,
257
+ "loss": 0.1834,
258
+ "step": 16500
259
+ },
260
+ {
261
+ "epoch": 2.770534550195567,
262
+ "grad_norm": 6.171872615814209,
263
+ "learning_rate": 8.94702419882276e-06,
264
+ "loss": 0.1932,
265
+ "step": 17000
266
+ },
267
+ {
268
+ "epoch": 2.8520208604954367,
269
+ "grad_norm": 4.4419050216674805,
270
+ "learning_rate": 8.620013080444736e-06,
271
+ "loss": 0.1796,
272
+ "step": 17500
273
+ },
274
+ {
275
+ "epoch": 2.9335071707953064,
276
+ "grad_norm": 4.847419738769531,
277
+ "learning_rate": 8.29300196206671e-06,
278
+ "loss": 0.1857,
279
+ "step": 18000
280
+ },
281
+ {
282
+ "epoch": 3.0,
283
+ "eval_accuracy": 0.8828323993886907,
284
+ "eval_loss": 0.3725295662879944,
285
+ "eval_runtime": 11.0431,
286
+ "eval_samples_per_second": 888.792,
287
+ "eval_steps_per_second": 13.945,
288
+ "step": 18408
289
+ },
290
+ {
291
+ "epoch": 3.014993481095176,
292
+ "grad_norm": 5.3139729499816895,
293
+ "learning_rate": 7.965990843688686e-06,
294
+ "loss": 0.1749,
295
+ "step": 18500
296
+ },
297
+ {
298
+ "epoch": 3.0964797913950455,
299
+ "grad_norm": 8.973730087280273,
300
+ "learning_rate": 7.63897972531066e-06,
301
+ "loss": 0.1251,
302
+ "step": 19000
303
+ },
304
+ {
305
+ "epoch": 3.1779661016949152,
306
+ "grad_norm": 5.263895511627197,
307
+ "learning_rate": 7.311968606932637e-06,
308
+ "loss": 0.1311,
309
+ "step": 19500
310
+ },
311
+ {
312
+ "epoch": 3.259452411994785,
313
+ "grad_norm": 7.2389397621154785,
314
+ "learning_rate": 6.984957488554611e-06,
315
+ "loss": 0.1254,
316
+ "step": 20000
317
+ },
318
+ {
319
+ "epoch": 3.3409387222946547,
320
+ "grad_norm": 5.466920375823975,
321
+ "learning_rate": 6.6579463701765866e-06,
322
+ "loss": 0.131,
323
+ "step": 20500
324
+ },
325
+ {
326
+ "epoch": 3.422425032594524,
327
+ "grad_norm": 3.5215988159179688,
328
+ "learning_rate": 6.330935251798561e-06,
329
+ "loss": 0.1289,
330
+ "step": 21000
331
+ },
332
+ {
333
+ "epoch": 3.5039113428943938,
334
+ "grad_norm": 5.995067119598389,
335
+ "learning_rate": 6.003924133420537e-06,
336
+ "loss": 0.1348,
337
+ "step": 21500
338
+ },
339
+ {
340
+ "epoch": 3.5853976531942635,
341
+ "grad_norm": 6.122405529022217,
342
+ "learning_rate": 5.676913015042512e-06,
343
+ "loss": 0.135,
344
+ "step": 22000
345
+ },
346
+ {
347
+ "epoch": 3.666883963494133,
348
+ "grad_norm": 7.12537145614624,
349
+ "learning_rate": 5.349901896664487e-06,
350
+ "loss": 0.1351,
351
+ "step": 22500
352
+ },
353
+ {
354
+ "epoch": 3.7483702737940026,
355
+ "grad_norm": 2.369924783706665,
356
+ "learning_rate": 5.022890778286462e-06,
357
+ "loss": 0.1261,
358
+ "step": 23000
359
+ },
360
+ {
361
+ "epoch": 3.8298565840938723,
362
+ "grad_norm": 9.514494895935059,
363
+ "learning_rate": 4.695879659908437e-06,
364
+ "loss": 0.1342,
365
+ "step": 23500
366
+ },
367
+ {
368
+ "epoch": 3.9113428943937416,
369
+ "grad_norm": 3.0919198989868164,
370
+ "learning_rate": 4.368868541530412e-06,
371
+ "loss": 0.1362,
372
+ "step": 24000
373
+ },
374
+ {
375
+ "epoch": 3.9928292046936114,
376
+ "grad_norm": 4.291722774505615,
377
+ "learning_rate": 4.041857423152388e-06,
378
+ "loss": 0.126,
379
+ "step": 24500
380
+ },
381
+ {
382
+ "epoch": 4.0,
383
+ "eval_accuracy": 0.8852776362710137,
384
+ "eval_loss": 0.43032392859458923,
385
+ "eval_runtime": 10.8684,
386
+ "eval_samples_per_second": 903.081,
387
+ "eval_steps_per_second": 14.17,
388
+ "step": 24544
389
+ }
390
+ ],
391
+ "logging_steps": 500,
392
+ "max_steps": 30680,
393
+ "num_input_tokens_seen": 0,
394
+ "num_train_epochs": 5,
395
+ "save_steps": 500,
396
+ "stateful_callbacks": {
397
+ "TrainerControl": {
398
+ "args": {
399
+ "should_epoch_stop": false,
400
+ "should_evaluate": false,
401
+ "should_log": false,
402
+ "should_save": true,
403
+ "should_training_stop": false
404
+ },
405
+ "attributes": {}
406
+ }
407
+ },
408
+ "total_flos": 9.890772120374488e+16,
409
+ "train_batch_size": 64,
410
+ "trial_name": null,
411
+ "trial_params": null
412
+ }
deberta-mnli/checkpoint-24544/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c99385eb30c86c40ef5ece4146e57520c621b89437abcdf5139cfd693d43451
3
+ size 5176
deberta-mnli/checkpoint-24544/vocab.json ADDED
The diff for this file is too large to render. See raw diff