niobures commited on
Commit
47f6225
·
verified ·
1 Parent(s): 433895e

bert-large-uncased-whole-word-masking-finetuned-squad, roberta-base-discourse-marker-prediction, xlm-roberta-base

Browse files
BERT/bert-large-uncased-whole-word-masking-finetuned-squad/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d25f8217b8877a253a602a22b94b6e5fff7a3bcbc8ee61fe91b368d0fc8a0ce4
3
+ size 1340995544
BERT/bert-large-uncased-whole-word-masking-finetuned-squad/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e2349f21ec01f7c90729cd54452e28d1a3aa7f2d7f8dfd91b437fb4fef022da
3
+ size 670783496
BERT/bert-large-uncased-whole-word-masking-finetuned-squad/model_int8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe79991f4bf5aa1dd07ca36edf6deb4fc86a09dac9682f0a751943d78d7bfa4d
3
+ size 336791930
BERT/bert-large-uncased-whole-word-masking-finetuned-squad/model_opt.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4747d76b753d28e50bfee5d96de9fd7408fcf7e0340a96bffa9a260a90c06d2
3
+ size 1340944121
BERT/bert-large-uncased-whole-word-masking-finetuned-squad/model_uint8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0723656cd81b24345b08ad041a71c096445df62893e80c7a147dba1cb5f62e39
3
+ size 336792017
RoBERTa/roberta-base-discourse-marker-prediction/.gitattributes ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ftz filter=lfs diff=lfs merge=lfs -text
6
+ *.gz filter=lfs diff=lfs merge=lfs -text
7
+ *.h5 filter=lfs diff=lfs merge=lfs -text
8
+ *.joblib filter=lfs diff=lfs merge=lfs -text
9
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
+ *.model filter=lfs diff=lfs merge=lfs -text
11
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
12
+ *.onnx filter=lfs diff=lfs merge=lfs -text
13
+ *.ot filter=lfs diff=lfs merge=lfs -text
14
+ *.parquet filter=lfs diff=lfs merge=lfs -text
15
+ *.pb filter=lfs diff=lfs merge=lfs -text
16
+ *.pt filter=lfs diff=lfs merge=lfs -text
17
+ *.pth filter=lfs diff=lfs merge=lfs -text
18
+ *.rar filter=lfs diff=lfs merge=lfs -text
19
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
20
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
21
+ *.tflite filter=lfs diff=lfs merge=lfs -text
22
+ *.tgz filter=lfs diff=lfs merge=lfs -text
23
+ *.wasm filter=lfs diff=lfs merge=lfs -text
24
+ *.xz filter=lfs diff=lfs merge=lfs -text
25
+ *.zip filter=lfs diff=lfs merge=lfs -text
26
+ *.zstandard filter=lfs diff=lfs merge=lfs -text
27
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ model.safetensors filter=lfs diff=lfs merge=lfs -text
RoBERTa/roberta-base-discourse-marker-prediction/README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - discourse-marker-prediction
6
+ - discourse-connective-prediction
7
+ - discourse-connective
8
+ - discourse-marker
9
+ - discourse-relation-prediction
10
+ - pragmatics
11
+ - discourse
12
+ license: apache-2.0
13
+ datasets:
14
+ - discovery
15
+ metrics:
16
+ - accuracy
17
+ widget:
18
+ - text: "But no, Amazon selling 3D printers is not new.</s></s>The promise of 3D printing is very great."
19
+ ---
20
+
21
+ # Discourse marker prediction / discourse connective prediction pretrained model
22
+
23
+ `roberta-base` pretrained on discourse marker prediction on the Discovery dataset with a validation accuracy of 30.93% (majority class is 0.57%)
24
+
25
+ https://github.com/sileod/discovery
26
+
27
+ https://huggingface.co/datasets/discovery
28
+
29
+ This model can also be used as a pretrained model for NLU, pragmatics and discourse tasks
30
+
31
+ ## Citing & Authors
32
+
33
+ ```bibtex
34
+ @inproceedings{sileo-etal-2019-mining,
35
+ title = "Mining Discourse Markers for Unsupervised Sentence Representation Learning",
36
+ author = "Sileo, Damien and
37
+ Van De Cruys, Tim and
38
+ Pradel, Camille and
39
+ Muller, Philippe",
40
+ booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
41
+ month = jun,
42
+ year = "2019",
43
+ address = "Minneapolis, Minnesota",
44
+ publisher = "Association for Computational Linguistics",
45
+ url = "https://aclanthology.org/N19-1351",
46
+ doi = "10.18653/v1/N19-1351",
47
+ pages = "3477--3486",
48
+ }
49
+ ```
RoBERTa/roberta-base-discourse-marker-prediction/config.json ADDED
@@ -0,0 +1,380 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "roberta-base-discourse-marker-prediction/",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "[no-conn]",
15
+ "1": "absolutely,",
16
+ "10": "alternately,",
17
+ "100": "notably,",
18
+ "101": "now,",
19
+ "102": "obviously,",
20
+ "103": "occasionally,",
21
+ "104": "oddly,",
22
+ "105": "often,",
23
+ "106": "on_the_contrary,",
24
+ "107": "on_the_other_hand",
25
+ "108": "once,",
26
+ "109": "only,",
27
+ "11": "alternatively",
28
+ "110": "optionally,",
29
+ "111": "or,",
30
+ "112": "originally,",
31
+ "113": "otherwise,",
32
+ "114": "overall,",
33
+ "115": "particularly,",
34
+ "116": "perhaps,",
35
+ "117": "personally,",
36
+ "118": "plus,",
37
+ "119": "preferably,",
38
+ "12": "although,",
39
+ "120": "presently,",
40
+ "121": "presumably,",
41
+ "122": "previously,",
42
+ "123": "probably,",
43
+ "124": "rather,",
44
+ "125": "realistically,",
45
+ "126": "really,",
46
+ "127": "recently,",
47
+ "128": "regardless,",
48
+ "129": "remarkably,",
49
+ "13": "altogether,",
50
+ "130": "sadly,",
51
+ "131": "second,",
52
+ "132": "secondly,",
53
+ "133": "separately,",
54
+ "134": "seriously,",
55
+ "135": "significantly,",
56
+ "136": "similarly,",
57
+ "137": "simultaneously",
58
+ "138": "slowly,",
59
+ "139": "so,",
60
+ "14": "amazingly,",
61
+ "140": "sometimes,",
62
+ "141": "soon,",
63
+ "142": "specifically,",
64
+ "143": "still,",
65
+ "144": "strangely,",
66
+ "145": "subsequently,",
67
+ "146": "suddenly,",
68
+ "147": "supposedly,",
69
+ "148": "surely,",
70
+ "149": "surprisingly,",
71
+ "15": "and",
72
+ "150": "technically,",
73
+ "151": "thankfully,",
74
+ "152": "then,",
75
+ "153": "theoretically,",
76
+ "154": "thereafter,",
77
+ "155": "thereby,",
78
+ "156": "therefore",
79
+ "157": "third,",
80
+ "158": "thirdly,",
81
+ "159": "this,",
82
+ "16": "anyway,",
83
+ "160": "though,",
84
+ "161": "thus,",
85
+ "162": "together,",
86
+ "163": "traditionally,",
87
+ "164": "truly,",
88
+ "165": "truthfully,",
89
+ "166": "typically,",
90
+ "167": "ultimately,",
91
+ "168": "undoubtedly,",
92
+ "169": "unfortunately,",
93
+ "17": "apparently,",
94
+ "170": "unsurprisingly,",
95
+ "171": "usually,",
96
+ "172": "well,",
97
+ "173": "yet,",
98
+ "18": "arguably,",
99
+ "19": "as_a_result,",
100
+ "2": "accordingly",
101
+ "20": "basically,",
102
+ "21": "because_of_that",
103
+ "22": "because_of_this",
104
+ "23": "besides,",
105
+ "24": "but",
106
+ "25": "by_comparison,",
107
+ "26": "by_contrast,",
108
+ "27": "by_doing_this,",
109
+ "28": "by_then",
110
+ "29": "certainly,",
111
+ "3": "actually,",
112
+ "30": "clearly,",
113
+ "31": "coincidentally,",
114
+ "32": "collectively,",
115
+ "33": "consequently",
116
+ "34": "conversely",
117
+ "35": "curiously,",
118
+ "36": "currently,",
119
+ "37": "elsewhere,",
120
+ "38": "especially,",
121
+ "39": "essentially,",
122
+ "4": "additionally",
123
+ "40": "eventually,",
124
+ "41": "evidently,",
125
+ "42": "finally,",
126
+ "43": "first,",
127
+ "44": "firstly,",
128
+ "45": "for_example",
129
+ "46": "for_instance",
130
+ "47": "fortunately,",
131
+ "48": "frankly,",
132
+ "49": "frequently,",
133
+ "5": "admittedly,",
134
+ "50": "further,",
135
+ "51": "furthermore",
136
+ "52": "generally,",
137
+ "53": "gradually,",
138
+ "54": "happily,",
139
+ "55": "hence,",
140
+ "56": "here,",
141
+ "57": "historically,",
142
+ "58": "honestly,",
143
+ "59": "hopefully,",
144
+ "6": "afterward",
145
+ "60": "however",
146
+ "61": "ideally,",
147
+ "62": "immediately,",
148
+ "63": "importantly,",
149
+ "64": "in_contrast,",
150
+ "65": "in_fact,",
151
+ "66": "in_other_words",
152
+ "67": "in_particular,",
153
+ "68": "in_short,",
154
+ "69": "in_sum,",
155
+ "7": "again,",
156
+ "70": "in_the_end,",
157
+ "71": "in_the_meantime,",
158
+ "72": "in_turn,",
159
+ "73": "incidentally,",
160
+ "74": "increasingly,",
161
+ "75": "indeed,",
162
+ "76": "inevitably,",
163
+ "77": "initially,",
164
+ "78": "instead,",
165
+ "79": "interestingly,",
166
+ "8": "already,",
167
+ "80": "ironically,",
168
+ "81": "lastly,",
169
+ "82": "lately,",
170
+ "83": "later,",
171
+ "84": "likewise,",
172
+ "85": "locally,",
173
+ "86": "luckily,",
174
+ "87": "maybe,",
175
+ "88": "meaning,",
176
+ "89": "meantime,",
177
+ "9": "also,",
178
+ "90": "meanwhile,",
179
+ "91": "moreover",
180
+ "92": "mostly,",
181
+ "93": "namely,",
182
+ "94": "nationally,",
183
+ "95": "naturally,",
184
+ "96": "nevertheless",
185
+ "97": "next,",
186
+ "98": "nonetheless",
187
+ "99": "normally,"
188
+ },
189
+ "initializer_range": 0.02,
190
+ "intermediate_size": 3072,
191
+ "label2id": {
192
+ "[no-conn]": 0,
193
+ "absolutely,": 1,
194
+ "accordingly": 2,
195
+ "actually,": 3,
196
+ "additionally": 4,
197
+ "admittedly,": 5,
198
+ "afterward": 6,
199
+ "again,": 7,
200
+ "already,": 8,
201
+ "also,": 9,
202
+ "alternately,": 10,
203
+ "alternatively": 11,
204
+ "although,": 12,
205
+ "altogether,": 13,
206
+ "amazingly,": 14,
207
+ "and": 15,
208
+ "anyway,": 16,
209
+ "apparently,": 17,
210
+ "arguably,": 18,
211
+ "as_a_result,": 19,
212
+ "basically,": 20,
213
+ "because_of_that": 21,
214
+ "because_of_this": 22,
215
+ "besides,": 23,
216
+ "but": 24,
217
+ "by_comparison,": 25,
218
+ "by_contrast,": 26,
219
+ "by_doing_this,": 27,
220
+ "by_then": 28,
221
+ "certainly,": 29,
222
+ "clearly,": 30,
223
+ "coincidentally,": 31,
224
+ "collectively,": 32,
225
+ "consequently": 33,
226
+ "conversely": 34,
227
+ "curiously,": 35,
228
+ "currently,": 36,
229
+ "elsewhere,": 37,
230
+ "especially,": 38,
231
+ "essentially,": 39,
232
+ "eventually,": 40,
233
+ "evidently,": 41,
234
+ "finally,": 42,
235
+ "first,": 43,
236
+ "firstly,": 44,
237
+ "for_example": 45,
238
+ "for_instance": 46,
239
+ "fortunately,": 47,
240
+ "frankly,": 48,
241
+ "frequently,": 49,
242
+ "further,": 50,
243
+ "furthermore": 51,
244
+ "generally,": 52,
245
+ "gradually,": 53,
246
+ "happily,": 54,
247
+ "hence,": 55,
248
+ "here,": 56,
249
+ "historically,": 57,
250
+ "honestly,": 58,
251
+ "hopefully,": 59,
252
+ "however": 60,
253
+ "ideally,": 61,
254
+ "immediately,": 62,
255
+ "importantly,": 63,
256
+ "in_contrast,": 64,
257
+ "in_fact,": 65,
258
+ "in_other_words": 66,
259
+ "in_particular,": 67,
260
+ "in_short,": 68,
261
+ "in_sum,": 69,
262
+ "in_the_end,": 70,
263
+ "in_the_meantime,": 71,
264
+ "in_turn,": 72,
265
+ "incidentally,": 73,
266
+ "increasingly,": 74,
267
+ "indeed,": 75,
268
+ "inevitably,": 76,
269
+ "initially,": 77,
270
+ "instead,": 78,
271
+ "interestingly,": 79,
272
+ "ironically,": 80,
273
+ "lastly,": 81,
274
+ "lately,": 82,
275
+ "later,": 83,
276
+ "likewise,": 84,
277
+ "locally,": 85,
278
+ "luckily,": 86,
279
+ "maybe,": 87,
280
+ "meaning,": 88,
281
+ "meantime,": 89,
282
+ "meanwhile,": 90,
283
+ "moreover": 91,
284
+ "mostly,": 92,
285
+ "namely,": 93,
286
+ "nationally,": 94,
287
+ "naturally,": 95,
288
+ "nevertheless": 96,
289
+ "next,": 97,
290
+ "nonetheless": 98,
291
+ "normally,": 99,
292
+ "notably,": 100,
293
+ "now,": 101,
294
+ "obviously,": 102,
295
+ "occasionally,": 103,
296
+ "oddly,": 104,
297
+ "often,": 105,
298
+ "on_the_contrary,": 106,
299
+ "on_the_other_hand": 107,
300
+ "once,": 108,
301
+ "only,": 109,
302
+ "optionally,": 110,
303
+ "or,": 111,
304
+ "originally,": 112,
305
+ "otherwise,": 113,
306
+ "overall,": 114,
307
+ "particularly,": 115,
308
+ "perhaps,": 116,
309
+ "personally,": 117,
310
+ "plus,": 118,
311
+ "preferably,": 119,
312
+ "presently,": 120,
313
+ "presumably,": 121,
314
+ "previously,": 122,
315
+ "probably,": 123,
316
+ "rather,": 124,
317
+ "realistically,": 125,
318
+ "really,": 126,
319
+ "recently,": 127,
320
+ "regardless,": 128,
321
+ "remarkably,": 129,
322
+ "sadly,": 130,
323
+ "second,": 131,
324
+ "secondly,": 132,
325
+ "separately,": 133,
326
+ "seriously,": 134,
327
+ "significantly,": 135,
328
+ "similarly,": 136,
329
+ "simultaneously": 137,
330
+ "slowly,": 138,
331
+ "so,": 139,
332
+ "sometimes,": 140,
333
+ "soon,": 141,
334
+ "specifically,": 142,
335
+ "still,": 143,
336
+ "strangely,": 144,
337
+ "subsequently,": 145,
338
+ "suddenly,": 146,
339
+ "supposedly,": 147,
340
+ "surely,": 148,
341
+ "surprisingly,": 149,
342
+ "technically,": 150,
343
+ "thankfully,": 151,
344
+ "then,": 152,
345
+ "theoretically,": 153,
346
+ "thereafter,": 154,
347
+ "thereby,": 155,
348
+ "therefore": 156,
349
+ "third,": 157,
350
+ "thirdly,": 158,
351
+ "this,": 159,
352
+ "though,": 160,
353
+ "thus,": 161,
354
+ "together,": 162,
355
+ "traditionally,": 163,
356
+ "truly,": 164,
357
+ "truthfully,": 165,
358
+ "typically,": 166,
359
+ "ultimately,": 167,
360
+ "undoubtedly,": 168,
361
+ "unfortunately,": 169,
362
+ "unsurprisingly,": 170,
363
+ "usually,": 171,
364
+ "well,": 172,
365
+ "yet,": 173
366
+ },
367
+ "layer_norm_eps": 1e-05,
368
+ "max_position_embeddings": 514,
369
+ "model_type": "roberta",
370
+ "num_attention_heads": 12,
371
+ "num_hidden_layers": 12,
372
+ "pad_token_id": 1,
373
+ "position_embedding_type": "absolute",
374
+ "problem_type": "single_label_classification",
375
+ "torch_dtype": "float32",
376
+ "transformers_version": "4.17.0",
377
+ "type_vocab_size": 1,
378
+ "use_cache": true,
379
+ "vocab_size": 50265
380
+ }
RoBERTa/roberta-base-discourse-marker-prediction/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
RoBERTa/roberta-base-discourse-marker-prediction/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b028bf3cf656959677292b56aa99f74479f6b88d15857e27f0c1fec21c12fe48
3
+ size 499146104
RoBERTa/roberta-base-discourse-marker-prediction/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57cc527274cdcfaf49f1366594eb30512c45ecaf2a3a5b800eb82676868408ba
3
+ size 499201261
RoBERTa/roberta-base-discourse-marker-prediction/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
RoBERTa/roberta-base-discourse-marker-prediction/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
RoBERTa/roberta-base-discourse-marker-prediction/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>", "add_prefix_space": false, "trim_offsets": true, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "roberta-base", "tokenizer_class": "RobertaTokenizer"}
RoBERTa/roberta-base-discourse-marker-prediction/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
XLM-RoBERTa/xlm-roberta-base/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "XLMRobertaForMaskedLM"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-05,
14
+ "max_position_embeddings": 514,
15
+ "model_type": "xlm-roberta",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "output_past": true,
19
+ "pad_token_id": 1,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.17.0.dev0",
22
+ "type_vocab_size": 1,
23
+ "use_cache": true,
24
+ "vocab_size": 250002
25
+ }
XLM-RoBERTa/xlm-roberta-base/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a76bfe6a405f1a9ace42b2dbd8fbd284dd8127a732ddcf2145b0fc9413b30d40
3
+ size 1881470773
XLM-RoBERTa/xlm-roberta-base/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
XLM-RoBERTa/xlm-roberta-base/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_max_length": 512}