veryfansome commited on
Commit
b249cec
·
verified ·
1 Parent(s): 8e63bf6

Upload 9 files

Browse files
models/ud_ewt_gum_pud_20250610/README.md ADDED
The diff for this file is too large to render. See raw diff
 
models/ud_ewt_gum_pud_20250610/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
models/ud_ewt_gum_pud_20250610/config.json ADDED
@@ -0,0 +1,262 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-base",
3
+ "architectures": [
4
+ "MultiHeadModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 3072,
12
+ "label_maps": {
13
+ "Abbr": [
14
+ "X",
15
+ "Yes"
16
+ ],
17
+ "Case": [
18
+ "X",
19
+ "Nom",
20
+ "Acc"
21
+ ],
22
+ "Definite": [
23
+ "X",
24
+ "Def",
25
+ "Ind"
26
+ ],
27
+ "Degree": [
28
+ "X",
29
+ "Sup",
30
+ "Pos",
31
+ "Cmp"
32
+ ],
33
+ "Foreign": [
34
+ "X",
35
+ "Yes"
36
+ ],
37
+ "Gender": [
38
+ "Fem",
39
+ "X",
40
+ "Neut",
41
+ "Masc"
42
+ ],
43
+ "Mood": [
44
+ "X",
45
+ "Ind",
46
+ "Imp"
47
+ ],
48
+ "NumType": [
49
+ "Card",
50
+ "X",
51
+ "Ord",
52
+ "Mult"
53
+ ],
54
+ "Number": [
55
+ "X",
56
+ "Plur",
57
+ "Sing"
58
+ ],
59
+ "Person": [
60
+ "X",
61
+ "1",
62
+ "2",
63
+ "3"
64
+ ],
65
+ "Polarity": [
66
+ "X",
67
+ "Neg"
68
+ ],
69
+ "Poss": [
70
+ "X",
71
+ "Yes"
72
+ ],
73
+ "PronType": [
74
+ "Dem",
75
+ "Rel",
76
+ "Prs",
77
+ "X",
78
+ "Art",
79
+ "Int"
80
+ ],
81
+ "Reflex": [
82
+ "X",
83
+ "Yes"
84
+ ],
85
+ "Tense": [
86
+ "Pres",
87
+ "X",
88
+ "Past"
89
+ ],
90
+ "VerbForm": [
91
+ "Ger",
92
+ "Part",
93
+ "Fin",
94
+ "X",
95
+ "Inf"
96
+ ],
97
+ "Voice": [
98
+ "X",
99
+ "Pass"
100
+ ],
101
+ "deprel": [
102
+ "cop",
103
+ "appos",
104
+ "nmod:tmod",
105
+ "nmod:poss",
106
+ "ccomp",
107
+ "compound",
108
+ "csubj",
109
+ "case",
110
+ "obl:tmod",
111
+ "cc:preconj",
112
+ "flat",
113
+ "compound:prt",
114
+ "xcomp",
115
+ "nmod:npmod",
116
+ "nummod",
117
+ "aux",
118
+ "root",
119
+ "expl",
120
+ "vocative",
121
+ "cc",
122
+ "discourse",
123
+ "punct",
124
+ "list",
125
+ "mark",
126
+ "nmod",
127
+ "advmod",
128
+ "nsubj:pass",
129
+ "acl:relcl",
130
+ "acl",
131
+ "det:predet",
132
+ "obj",
133
+ "amod",
134
+ "advcl",
135
+ "conj",
136
+ "aux:pass",
137
+ "parataxis",
138
+ "nsubj",
139
+ "csubj:pass",
140
+ "det",
141
+ "iobj",
142
+ "obl:npmod",
143
+ "obl",
144
+ "fixed"
145
+ ],
146
+ "pos": [
147
+ "SYM",
148
+ "ADV",
149
+ "NUM",
150
+ "NOUN",
151
+ "INTJ",
152
+ "SCONJ",
153
+ "PROPN",
154
+ "ADJ",
155
+ "PART",
156
+ "CCONJ",
157
+ "X",
158
+ "VERB",
159
+ "PRON",
160
+ "PUNCT",
161
+ "ADP",
162
+ "AUX",
163
+ "DET"
164
+ ],
165
+ "xpos": [
166
+ "SYM",
167
+ "``",
168
+ "AFX",
169
+ "ADD",
170
+ "POS",
171
+ "PRP$",
172
+ "NFP",
173
+ "VBZ",
174
+ "RBR",
175
+ "JJ",
176
+ "JJR",
177
+ "WRB",
178
+ "VB",
179
+ ".",
180
+ "WP",
181
+ "RP",
182
+ "-RRB-",
183
+ "MD",
184
+ "PRP",
185
+ "VBP",
186
+ "IN",
187
+ "PDT",
188
+ "$",
189
+ "RBS",
190
+ "FW",
191
+ "NNS",
192
+ "-LRB-",
193
+ ",",
194
+ "HYPH",
195
+ "RB",
196
+ "WP$",
197
+ "CC",
198
+ "VBD",
199
+ "EX",
200
+ "LS",
201
+ "CD",
202
+ "NNP",
203
+ "NN",
204
+ "JJS",
205
+ "VBG",
206
+ ":",
207
+ "UH",
208
+ "''",
209
+ "DT",
210
+ "VBN",
211
+ "WDT",
212
+ "NNPS",
213
+ "TO"
214
+ ]
215
+ },
216
+ "layer_norm_eps": 1e-07,
217
+ "legacy": true,
218
+ "max_position_embeddings": 512,
219
+ "max_relative_positions": -1,
220
+ "model_type": "deberta-v2",
221
+ "norm_rel_ebd": "layer_norm",
222
+ "num_attention_heads": 12,
223
+ "num_hidden_layers": 12,
224
+ "num_labels_dict": {
225
+ "Abbr": 2,
226
+ "Case": 3,
227
+ "Definite": 3,
228
+ "Degree": 4,
229
+ "Foreign": 2,
230
+ "Gender": 4,
231
+ "Mood": 3,
232
+ "NumType": 4,
233
+ "Number": 3,
234
+ "Person": 4,
235
+ "Polarity": 2,
236
+ "Poss": 2,
237
+ "PronType": 6,
238
+ "Reflex": 2,
239
+ "Tense": 3,
240
+ "VerbForm": 5,
241
+ "Voice": 2,
242
+ "deprel": 43,
243
+ "pos": 17,
244
+ "xpos": 48
245
+ },
246
+ "pad_token_id": 0,
247
+ "pooler_dropout": 0,
248
+ "pooler_hidden_act": "gelu",
249
+ "pooler_hidden_size": 768,
250
+ "pos_att_type": [
251
+ "p2c",
252
+ "c2p"
253
+ ],
254
+ "position_biased_input": false,
255
+ "position_buckets": 256,
256
+ "relative_attention": true,
257
+ "share_att_key": true,
258
+ "torch_dtype": "float32",
259
+ "transformers_version": "4.49.0",
260
+ "type_vocab_size": 0,
261
+ "vocab_size": 128100
262
+ }
models/ud_ewt_gum_pud_20250610/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbb82ce82b366fcac7a2e4cf25b68a4c5685b3f104a661767c64087d1b6b8de2
3
+ size 783103664
models/ud_ewt_gum_pud_20250610/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
models/ud_ewt_gum_pud_20250610/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
models/ud_ewt_gum_pud_20250610/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
models/ud_ewt_gum_pud_20250610/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "[PAD]",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "[CLS]",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "[SEP]",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "[UNK]",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "128000": {
37
+ "content": "[MASK]",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "[CLS]",
46
+ "clean_up_tokenization_spaces": false,
47
+ "cls_token": "[CLS]",
48
+ "do_lower_case": false,
49
+ "eos_token": "[SEP]",
50
+ "extra_special_tokens": {},
51
+ "mask_token": "[MASK]",
52
+ "model_max_length": 1000000000000000019884624838656,
53
+ "pad_token": "[PAD]",
54
+ "sep_token": "[SEP]",
55
+ "sp_model_kwargs": {},
56
+ "split_by_punct": false,
57
+ "tokenizer_class": "DebertaV2Tokenizer",
58
+ "unk_token": "[UNK]",
59
+ "vocab_type": "spm"
60
+ }
models/ud_ewt_gum_pud_20250610/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ecbcfdb8463d3a83ca811cbc76f741693919cd4cb81f75e176cc7719038f517
3
+ size 5304