Upload folder using huggingface_hub

#1
by animaai - opened
config.json ADDED
@@ -0,0 +1,423 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./roberta_cleen_all_exist_new_0/model_thulje/epoch_1",
3
+ "architectures": [
4
+ "RobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "crf_state_dict": {
10
+ "_constraint_mask": [
11
+ [
12
+ 1.0,
13
+ 1.0,
14
+ 1.0,
15
+ 1.0,
16
+ 1.0,
17
+ 1.0,
18
+ 0.0,
19
+ 0.0,
20
+ 0.0,
21
+ 0.0,
22
+ 1.0,
23
+ 0.0,
24
+ 1.0
25
+ ],
26
+ [
27
+ 1.0,
28
+ 1.0,
29
+ 1.0,
30
+ 1.0,
31
+ 1.0,
32
+ 0.0,
33
+ 1.0,
34
+ 0.0,
35
+ 0.0,
36
+ 0.0,
37
+ 1.0,
38
+ 0.0,
39
+ 1.0
40
+ ],
41
+ [
42
+ 1.0,
43
+ 1.0,
44
+ 1.0,
45
+ 1.0,
46
+ 1.0,
47
+ 0.0,
48
+ 0.0,
49
+ 1.0,
50
+ 0.0,
51
+ 0.0,
52
+ 1.0,
53
+ 0.0,
54
+ 1.0
55
+ ],
56
+ [
57
+ 1.0,
58
+ 1.0,
59
+ 1.0,
60
+ 1.0,
61
+ 1.0,
62
+ 0.0,
63
+ 0.0,
64
+ 0.0,
65
+ 1.0,
66
+ 0.0,
67
+ 1.0,
68
+ 0.0,
69
+ 1.0
70
+ ],
71
+ [
72
+ 1.0,
73
+ 1.0,
74
+ 1.0,
75
+ 1.0,
76
+ 1.0,
77
+ 0.0,
78
+ 0.0,
79
+ 0.0,
80
+ 0.0,
81
+ 1.0,
82
+ 1.0,
83
+ 0.0,
84
+ 1.0
85
+ ],
86
+ [
87
+ 1.0,
88
+ 1.0,
89
+ 1.0,
90
+ 1.0,
91
+ 1.0,
92
+ 1.0,
93
+ 0.0,
94
+ 0.0,
95
+ 0.0,
96
+ 0.0,
97
+ 1.0,
98
+ 0.0,
99
+ 1.0
100
+ ],
101
+ [
102
+ 1.0,
103
+ 1.0,
104
+ 1.0,
105
+ 1.0,
106
+ 1.0,
107
+ 0.0,
108
+ 1.0,
109
+ 0.0,
110
+ 0.0,
111
+ 0.0,
112
+ 1.0,
113
+ 0.0,
114
+ 1.0
115
+ ],
116
+ [
117
+ 1.0,
118
+ 1.0,
119
+ 1.0,
120
+ 1.0,
121
+ 1.0,
122
+ 0.0,
123
+ 0.0,
124
+ 1.0,
125
+ 0.0,
126
+ 0.0,
127
+ 1.0,
128
+ 0.0,
129
+ 1.0
130
+ ],
131
+ [
132
+ 1.0,
133
+ 1.0,
134
+ 1.0,
135
+ 1.0,
136
+ 1.0,
137
+ 0.0,
138
+ 0.0,
139
+ 0.0,
140
+ 1.0,
141
+ 0.0,
142
+ 1.0,
143
+ 0.0,
144
+ 1.0
145
+ ],
146
+ [
147
+ 1.0,
148
+ 1.0,
149
+ 1.0,
150
+ 1.0,
151
+ 1.0,
152
+ 0.0,
153
+ 0.0,
154
+ 0.0,
155
+ 0.0,
156
+ 1.0,
157
+ 1.0,
158
+ 0.0,
159
+ 1.0
160
+ ],
161
+ [
162
+ 1.0,
163
+ 1.0,
164
+ 1.0,
165
+ 1.0,
166
+ 1.0,
167
+ 0.0,
168
+ 0.0,
169
+ 0.0,
170
+ 0.0,
171
+ 0.0,
172
+ 1.0,
173
+ 0.0,
174
+ 1.0
175
+ ],
176
+ [
177
+ 1.0,
178
+ 1.0,
179
+ 1.0,
180
+ 1.0,
181
+ 1.0,
182
+ 0.0,
183
+ 0.0,
184
+ 0.0,
185
+ 0.0,
186
+ 0.0,
187
+ 1.0,
188
+ 0.0,
189
+ 0.0
190
+ ],
191
+ [
192
+ 0.0,
193
+ 0.0,
194
+ 0.0,
195
+ 0.0,
196
+ 0.0,
197
+ 0.0,
198
+ 0.0,
199
+ 0.0,
200
+ 0.0,
201
+ 0.0,
202
+ 0.0,
203
+ 0.0,
204
+ 0.0
205
+ ]
206
+ ],
207
+ "end_transitions": [
208
+ 2.240643262863159,
209
+ -0.2742827236652374,
210
+ 0.6393970251083374,
211
+ 0.005434449762105942,
212
+ -0.1513204723596573,
213
+ -1.2698227167129517,
214
+ 0.19887013733386993,
215
+ 0.28996115922927856,
216
+ 0.17393897473812103,
217
+ -0.33823758363723755,
218
+ 0.8555233478546143
219
+ ],
220
+ "start_transitions": [
221
+ -0.4646048843860626,
222
+ -0.06933217495679855,
223
+ 1.976535677909851,
224
+ -0.01753147505223751,
225
+ -0.7673749327659607,
226
+ -0.05074487626552582,
227
+ -2.2183399200439453,
228
+ 0.82484370470047,
229
+ 1.2874979972839355,
230
+ 1.3923321962356567,
231
+ 0.43642526865005493
232
+ ],
233
+ "transitions": [
234
+ [
235
+ 0.2501518726348877,
236
+ 0.0009667582344263792,
237
+ 0.30999264121055603,
238
+ 0.39742279052734375,
239
+ 0.31025633215904236,
240
+ 0.09058089554309845,
241
+ -0.019848795607686043,
242
+ -0.22607830166816711,
243
+ -0.37826070189476013,
244
+ -0.3088703155517578,
245
+ -0.40294963121414185
246
+ ],
247
+ [
248
+ 0.1511581689119339,
249
+ 0.06581507623195648,
250
+ 0.04197699949145317,
251
+ 0.13711129128932953,
252
+ -0.1339503526687622,
253
+ 0.2818056643009186,
254
+ -0.5844879150390625,
255
+ 0.2564857006072998,
256
+ -0.17954248189926147,
257
+ -0.022843224927783012,
258
+ 0.3660793900489807
259
+ ],
260
+ [
261
+ -0.22967128455638885,
262
+ 0.24774327874183655,
263
+ 0.2004176527261734,
264
+ 0.1407482624053955,
265
+ 0.5036953687667847,
266
+ -0.26214247941970825,
267
+ -0.06645846366882324,
268
+ -0.3897196352481842,
269
+ 0.3363957405090332,
270
+ 0.3760513961315155,
271
+ 0.08766741305589676
272
+ ],
273
+ [
274
+ 0.08801333606243134,
275
+ 0.559565544128418,
276
+ -0.1347968876361847,
277
+ -0.04922352358698845,
278
+ -0.27180802822113037,
279
+ 0.12606672942638397,
280
+ 0.3549225330352783,
281
+ -0.3206363320350647,
282
+ 0.0662713572382927,
283
+ 0.10210344195365906,
284
+ -0.4600335657596588
285
+ ],
286
+ [
287
+ -0.25136229395866394,
288
+ 0.11263541132211685,
289
+ 0.30574309825897217,
290
+ -0.12007039040327072,
291
+ -0.14439710974693298,
292
+ 0.02785634994506836,
293
+ -0.03327173739671707,
294
+ -0.4631956219673157,
295
+ 0.16959922015666962,
296
+ -0.8500743508338928,
297
+ -0.3696420192718506
298
+ ],
299
+ [
300
+ -0.14124469459056854,
301
+ 0.2320529967546463,
302
+ 0.3118607699871063,
303
+ 0.3109912872314453,
304
+ 0.0008499743416905403,
305
+ 0.30968427658081055,
306
+ -0.05566711351275444,
307
+ 0.4032119810581207,
308
+ -0.022859327495098114,
309
+ 0.15043704211711884,
310
+ 0.5076503157615662
311
+ ],
312
+ [
313
+ 0.12845280766487122,
314
+ 0.28300824761390686,
315
+ 0.13735704123973846,
316
+ 0.007540121674537659,
317
+ -0.2670679986476898,
318
+ 0.312574565410614,
319
+ -0.41549786925315857,
320
+ -0.5133036375045776,
321
+ 0.1582707017660141,
322
+ 0.19924618303775787,
323
+ 0.030154455453157425
324
+ ],
325
+ [
326
+ -0.3130203187465668,
327
+ 0.25306785106658936,
328
+ -0.27456873655319214,
329
+ -0.43190330266952515,
330
+ 0.2754882574081421,
331
+ 0.08095683157444,
332
+ -0.1389872282743454,
333
+ -0.17365975677967072,
334
+ -0.07709870487451553,
335
+ -0.42576122283935547,
336
+ 0.08613783866167068
337
+ ],
338
+ [
339
+ 0.3981761336326599,
340
+ 0.14935140311717987,
341
+ 0.23046627640724182,
342
+ -0.2853967547416687,
343
+ 0.07103794068098068,
344
+ -0.11574462801218033,
345
+ -0.07706587016582489,
346
+ -0.1586235910654068,
347
+ 0.4237719178199768,
348
+ -0.10810907930135727,
349
+ -0.6446168422698975
350
+ ],
351
+ [
352
+ 0.0846155658364296,
353
+ 0.03686073422431946,
354
+ -0.5546830296516418,
355
+ -0.6503274440765381,
356
+ 0.12535876035690308,
357
+ -0.3845370411872864,
358
+ 0.021639790385961533,
359
+ -0.5573288202285767,
360
+ 0.2939161956310272,
361
+ -0.37080487608909607,
362
+ -0.05652831494808197
363
+ ],
364
+ [
365
+ 0.11467353254556656,
366
+ 0.20622697472572327,
367
+ -0.14056970179080963,
368
+ 0.4235305190086365,
369
+ 0.33264821767807007,
370
+ 0.198853999376297,
371
+ 0.09324410557746887,
372
+ 0.15920676290988922,
373
+ -0.19516964256763458,
374
+ 0.5432201027870178,
375
+ -0.014160040766000748
376
+ ]
377
+ ]
378
+ },
379
+ "eos_token_id": 2,
380
+ "hidden_act": "gelu",
381
+ "hidden_dropout_prob": 0.1,
382
+ "hidden_size": 768,
383
+ "id2label": {
384
+ "0": "B-Dislikes",
385
+ "1": "B-Family",
386
+ "2": "B-Job",
387
+ "3": "B-Likes",
388
+ "4": "B-Pet",
389
+ "5": "I-Dislikes",
390
+ "6": "I-Family",
391
+ "7": "I-Job",
392
+ "8": "I-Likes",
393
+ "9": "I-Pet",
394
+ "10": "O"
395
+ },
396
+ "initializer_range": 0.02,
397
+ "intermediate_size": 3072,
398
+ "label2id": {
399
+ "B-Dislikes": 0,
400
+ "B-Family": 1,
401
+ "B-Job": 2,
402
+ "B-Likes": 3,
403
+ "B-Pet": 4,
404
+ "I-Dislikes": 5,
405
+ "I-Family": 6,
406
+ "I-Job": 7,
407
+ "I-Likes": 8,
408
+ "I-Pet": 9,
409
+ "O": 10
410
+ },
411
+ "layer_norm_eps": 1e-05,
412
+ "max_position_embeddings": 514,
413
+ "model_type": "roberta",
414
+ "num_attention_heads": 12,
415
+ "num_hidden_layers": 12,
416
+ "pad_token_id": 1,
417
+ "position_embedding_type": "absolute",
418
+ "torch_dtype": "float32",
419
+ "transformers_version": "4.20.1",
420
+ "type_vocab_size": 1,
421
+ "use_cache": true,
422
+ "vocab_size": 50265
423
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7bfa12f8aa6fa91ba70cf16836fc2c5ffd37fc5a4aafde0e8f2c849e2890e63
3
+ size 496326701
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<s>",
4
+ "cls_token": "<s>",
5
+ "eos_token": "</s>",
6
+ "errors": "replace",
7
+ "mask_token": "<mask>",
8
+ "model_max_length": 512,
9
+ "name_or_path": "./roberta_cleen_all_exist_new_0/model_thulje/epoch_1",
10
+ "pad_token": "<pad>",
11
+ "sep_token": "</s>",
12
+ "special_tokens_map_file": null,
13
+ "tokenizer_class": "RobertaTokenizer",
14
+ "trim_offsets": true,
15
+ "unk_token": "<unk>"
16
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff