liljacharlotte commited on
Commit
17f0232
·
verified ·
1 Parent(s): 85bd50f

Upload NorbertForSequenceClassification

Browse files
Files changed (3) hide show
  1. config.json +167 -0
  2. configuration_norbert.py +34 -0
  3. model.safetensors +3 -0
config.json ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/cluster/work/projects/ec30/liljacs_/models/joint_full/norpac_46",
3
+ "architectures": [
4
+ "NorbertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "auto_map": {
8
+ "AutoConfig": "configuration_norbert.NorbertConfig",
9
+ "AutoModel": "modeling_norbert.NorbertModel",
10
+ "AutoModelForMaskedLM": "modeling_norbert.NorbertForMaskedLM",
11
+ "AutoModelForMultipleChoice": "modeling_norbert.NorbertForMultipleChoice",
12
+ "AutoModelForQuestionAnswering": "modeling_norbert.NorbertForQuestionAnswering",
13
+ "AutoModelForSequenceClassification": "modeling_norbert.NorbertForSequenceClassification",
14
+ "AutoModelForTokenClassification": "modeling_norbert.NorbertForTokenClassification"
15
+ },
16
+ "hidden_dropout_prob": 0.1,
17
+ "hidden_size": 1024,
18
+ "id2label": {
19
+ "0": "act_mixed",
20
+ "1": "act_neg",
21
+ "2": "act_pos",
22
+ "3": "cp_mixed",
23
+ "4": "cp_neg",
24
+ "5": "cp_pos",
25
+ "6": "dur_neg",
26
+ "7": "dur_pos",
27
+ "8": "excos_neg",
28
+ "9": "excos_pos",
29
+ "10": "ftc_neg",
30
+ "11": "ftc_pos",
31
+ "12": "gd_neg",
32
+ "13": "gd_pos",
33
+ "14": "incc_neg",
34
+ "15": "incc_pos",
35
+ "16": "iop_neg",
36
+ "17": "iop_pos",
37
+ "18": "irr_mixed",
38
+ "19": "irr_neg",
39
+ "20": "irr_pos",
40
+ "21": "isp_neg",
41
+ "22": "isp_pos",
42
+ "23": "lang_neg",
43
+ "24": "lang_pos",
44
+ "25": "med_mixed",
45
+ "26": "med_neg",
46
+ "27": "med_pos",
47
+ "28": "no-asp",
48
+ "29": "oits_mixed",
49
+ "30": "oits_neg",
50
+ "31": "oits_pos",
51
+ "32": "pip_mixed",
52
+ "33": "pip_neg",
53
+ "34": "pip_pos",
54
+ "35": "ppe_mixed",
55
+ "36": "ppe_neg",
56
+ "37": "ppe_pos",
57
+ "38": "ppr_mixed",
58
+ "39": "ppr_neg",
59
+ "40": "ppr_pos",
60
+ "41": "qfm_mixed",
61
+ "42": "qfm_neg",
62
+ "43": "qfm_pos",
63
+ "44": "sct_mixed",
64
+ "45": "sct_neg",
65
+ "46": "sct_pos",
66
+ "47": "slohs_mixed",
67
+ "48": "slohs_neg",
68
+ "49": "slohs_pos",
69
+ "50": "sr_mixed",
70
+ "51": "sr_neg",
71
+ "52": "sr_pos",
72
+ "53": "td_mixed",
73
+ "54": "td_neg",
74
+ "55": "td_pos",
75
+ "56": "tshp_mixed",
76
+ "57": "tshp_neg",
77
+ "58": "tshp_pos",
78
+ "59": "wol_neg",
79
+ "60": "wol_pos",
80
+ "61": "wtc_mixed",
81
+ "62": "wtc_neg",
82
+ "63": "wtc_pos",
83
+ "64": "wtp_mixed",
84
+ "65": "wtp_neg",
85
+ "66": "wtp_pos"
86
+ },
87
+ "intermediate_size": 2730,
88
+ "label2id": {
89
+ "act_mixed": 0,
90
+ "act_neg": 1,
91
+ "act_pos": 2,
92
+ "cp_mixed": 3,
93
+ "cp_neg": 4,
94
+ "cp_pos": 5,
95
+ "dur_neg": 6,
96
+ "dur_pos": 7,
97
+ "excos_neg": 8,
98
+ "excos_pos": 9,
99
+ "ftc_neg": 10,
100
+ "ftc_pos": 11,
101
+ "gd_neg": 12,
102
+ "gd_pos": 13,
103
+ "incc_neg": 14,
104
+ "incc_pos": 15,
105
+ "iop_neg": 16,
106
+ "iop_pos": 17,
107
+ "irr_mixed": 18,
108
+ "irr_neg": 19,
109
+ "irr_pos": 20,
110
+ "isp_neg": 21,
111
+ "isp_pos": 22,
112
+ "lang_neg": 23,
113
+ "lang_pos": 24,
114
+ "med_mixed": 25,
115
+ "med_neg": 26,
116
+ "med_pos": 27,
117
+ "no-asp": 28,
118
+ "oits_mixed": 29,
119
+ "oits_neg": 30,
120
+ "oits_pos": 31,
121
+ "pip_mixed": 32,
122
+ "pip_neg": 33,
123
+ "pip_pos": 34,
124
+ "ppe_mixed": 35,
125
+ "ppe_neg": 36,
126
+ "ppe_pos": 37,
127
+ "ppr_mixed": 38,
128
+ "ppr_neg": 39,
129
+ "ppr_pos": 40,
130
+ "qfm_mixed": 41,
131
+ "qfm_neg": 42,
132
+ "qfm_pos": 43,
133
+ "sct_mixed": 44,
134
+ "sct_neg": 45,
135
+ "sct_pos": 46,
136
+ "slohs_mixed": 47,
137
+ "slohs_neg": 48,
138
+ "slohs_pos": 49,
139
+ "sr_mixed": 50,
140
+ "sr_neg": 51,
141
+ "sr_pos": 52,
142
+ "td_mixed": 53,
143
+ "td_neg": 54,
144
+ "td_pos": 55,
145
+ "tshp_mixed": 56,
146
+ "tshp_neg": 57,
147
+ "tshp_pos": 58,
148
+ "wol_neg": 59,
149
+ "wol_pos": 60,
150
+ "wtc_mixed": 61,
151
+ "wtc_neg": 62,
152
+ "wtc_pos": 63,
153
+ "wtp_mixed": 64,
154
+ "wtp_neg": 65,
155
+ "wtp_pos": 66
156
+ },
157
+ "layer_norm_eps": 1e-07,
158
+ "max_position_embeddings": 512,
159
+ "num_attention_heads": 16,
160
+ "num_hidden_layers": 24,
161
+ "output_all_encoded_layers": true,
162
+ "position_bucket_size": 32,
163
+ "problem_type": "multi_label_classification",
164
+ "torch_dtype": "float32",
165
+ "transformers_version": "4.47.1",
166
+ "vocab_size": 50000
167
+ }
configuration_norbert.py ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from transformers.configuration_utils import PretrainedConfig
2
+
3
+
4
+ class NorbertConfig(PretrainedConfig):
5
+ """Configuration class to store the configuration of a `NorbertModel`.
6
+ """
7
+ def __init__(
8
+ self,
9
+ vocab_size=50000,
10
+ attention_probs_dropout_prob=0.1,
11
+ hidden_dropout_prob=0.1,
12
+ hidden_size=768,
13
+ intermediate_size=2048,
14
+ max_position_embeddings=512,
15
+ position_bucket_size=32,
16
+ num_attention_heads=12,
17
+ num_hidden_layers=12,
18
+ layer_norm_eps=1.0e-7,
19
+ output_all_encoded_layers=True,
20
+ **kwargs,
21
+ ):
22
+ super().__init__(**kwargs)
23
+
24
+ self.vocab_size = vocab_size
25
+ self.hidden_size = hidden_size
26
+ self.num_hidden_layers = num_hidden_layers
27
+ self.num_attention_heads = num_attention_heads
28
+ self.intermediate_size = intermediate_size
29
+ self.hidden_dropout_prob = hidden_dropout_prob
30
+ self.attention_probs_dropout_prob = attention_probs_dropout_prob
31
+ self.max_position_embeddings = max_position_embeddings
32
+ self.output_all_encoded_layers = output_all_encoded_layers
33
+ self.position_bucket_size = position_bucket_size
34
+ self.layer_norm_eps = layer_norm_eps
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7520b60d429ad82e43aeb05cde7a05cf2a95a2025d73319eb215fc1d382637c
3
+ size 1417921780