Upload NorbertForSequenceClassification
Browse files- config.json +167 -0
- configuration_norbert.py +34 -0
- model.safetensors +3 -0
config.json
ADDED
|
@@ -0,0 +1,167 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "/cluster/work/projects/ec30/liljacs_/models/joint_full/norpac_46",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"NorbertForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"auto_map": {
|
| 8 |
+
"AutoConfig": "configuration_norbert.NorbertConfig",
|
| 9 |
+
"AutoModel": "modeling_norbert.NorbertModel",
|
| 10 |
+
"AutoModelForMaskedLM": "modeling_norbert.NorbertForMaskedLM",
|
| 11 |
+
"AutoModelForMultipleChoice": "modeling_norbert.NorbertForMultipleChoice",
|
| 12 |
+
"AutoModelForQuestionAnswering": "modeling_norbert.NorbertForQuestionAnswering",
|
| 13 |
+
"AutoModelForSequenceClassification": "modeling_norbert.NorbertForSequenceClassification",
|
| 14 |
+
"AutoModelForTokenClassification": "modeling_norbert.NorbertForTokenClassification"
|
| 15 |
+
},
|
| 16 |
+
"hidden_dropout_prob": 0.1,
|
| 17 |
+
"hidden_size": 1024,
|
| 18 |
+
"id2label": {
|
| 19 |
+
"0": "act_mixed",
|
| 20 |
+
"1": "act_neg",
|
| 21 |
+
"2": "act_pos",
|
| 22 |
+
"3": "cp_mixed",
|
| 23 |
+
"4": "cp_neg",
|
| 24 |
+
"5": "cp_pos",
|
| 25 |
+
"6": "dur_neg",
|
| 26 |
+
"7": "dur_pos",
|
| 27 |
+
"8": "excos_neg",
|
| 28 |
+
"9": "excos_pos",
|
| 29 |
+
"10": "ftc_neg",
|
| 30 |
+
"11": "ftc_pos",
|
| 31 |
+
"12": "gd_neg",
|
| 32 |
+
"13": "gd_pos",
|
| 33 |
+
"14": "incc_neg",
|
| 34 |
+
"15": "incc_pos",
|
| 35 |
+
"16": "iop_neg",
|
| 36 |
+
"17": "iop_pos",
|
| 37 |
+
"18": "irr_mixed",
|
| 38 |
+
"19": "irr_neg",
|
| 39 |
+
"20": "irr_pos",
|
| 40 |
+
"21": "isp_neg",
|
| 41 |
+
"22": "isp_pos",
|
| 42 |
+
"23": "lang_neg",
|
| 43 |
+
"24": "lang_pos",
|
| 44 |
+
"25": "med_mixed",
|
| 45 |
+
"26": "med_neg",
|
| 46 |
+
"27": "med_pos",
|
| 47 |
+
"28": "no-asp",
|
| 48 |
+
"29": "oits_mixed",
|
| 49 |
+
"30": "oits_neg",
|
| 50 |
+
"31": "oits_pos",
|
| 51 |
+
"32": "pip_mixed",
|
| 52 |
+
"33": "pip_neg",
|
| 53 |
+
"34": "pip_pos",
|
| 54 |
+
"35": "ppe_mixed",
|
| 55 |
+
"36": "ppe_neg",
|
| 56 |
+
"37": "ppe_pos",
|
| 57 |
+
"38": "ppr_mixed",
|
| 58 |
+
"39": "ppr_neg",
|
| 59 |
+
"40": "ppr_pos",
|
| 60 |
+
"41": "qfm_mixed",
|
| 61 |
+
"42": "qfm_neg",
|
| 62 |
+
"43": "qfm_pos",
|
| 63 |
+
"44": "sct_mixed",
|
| 64 |
+
"45": "sct_neg",
|
| 65 |
+
"46": "sct_pos",
|
| 66 |
+
"47": "slohs_mixed",
|
| 67 |
+
"48": "slohs_neg",
|
| 68 |
+
"49": "slohs_pos",
|
| 69 |
+
"50": "sr_mixed",
|
| 70 |
+
"51": "sr_neg",
|
| 71 |
+
"52": "sr_pos",
|
| 72 |
+
"53": "td_mixed",
|
| 73 |
+
"54": "td_neg",
|
| 74 |
+
"55": "td_pos",
|
| 75 |
+
"56": "tshp_mixed",
|
| 76 |
+
"57": "tshp_neg",
|
| 77 |
+
"58": "tshp_pos",
|
| 78 |
+
"59": "wol_neg",
|
| 79 |
+
"60": "wol_pos",
|
| 80 |
+
"61": "wtc_mixed",
|
| 81 |
+
"62": "wtc_neg",
|
| 82 |
+
"63": "wtc_pos",
|
| 83 |
+
"64": "wtp_mixed",
|
| 84 |
+
"65": "wtp_neg",
|
| 85 |
+
"66": "wtp_pos"
|
| 86 |
+
},
|
| 87 |
+
"intermediate_size": 2730,
|
| 88 |
+
"label2id": {
|
| 89 |
+
"act_mixed": 0,
|
| 90 |
+
"act_neg": 1,
|
| 91 |
+
"act_pos": 2,
|
| 92 |
+
"cp_mixed": 3,
|
| 93 |
+
"cp_neg": 4,
|
| 94 |
+
"cp_pos": 5,
|
| 95 |
+
"dur_neg": 6,
|
| 96 |
+
"dur_pos": 7,
|
| 97 |
+
"excos_neg": 8,
|
| 98 |
+
"excos_pos": 9,
|
| 99 |
+
"ftc_neg": 10,
|
| 100 |
+
"ftc_pos": 11,
|
| 101 |
+
"gd_neg": 12,
|
| 102 |
+
"gd_pos": 13,
|
| 103 |
+
"incc_neg": 14,
|
| 104 |
+
"incc_pos": 15,
|
| 105 |
+
"iop_neg": 16,
|
| 106 |
+
"iop_pos": 17,
|
| 107 |
+
"irr_mixed": 18,
|
| 108 |
+
"irr_neg": 19,
|
| 109 |
+
"irr_pos": 20,
|
| 110 |
+
"isp_neg": 21,
|
| 111 |
+
"isp_pos": 22,
|
| 112 |
+
"lang_neg": 23,
|
| 113 |
+
"lang_pos": 24,
|
| 114 |
+
"med_mixed": 25,
|
| 115 |
+
"med_neg": 26,
|
| 116 |
+
"med_pos": 27,
|
| 117 |
+
"no-asp": 28,
|
| 118 |
+
"oits_mixed": 29,
|
| 119 |
+
"oits_neg": 30,
|
| 120 |
+
"oits_pos": 31,
|
| 121 |
+
"pip_mixed": 32,
|
| 122 |
+
"pip_neg": 33,
|
| 123 |
+
"pip_pos": 34,
|
| 124 |
+
"ppe_mixed": 35,
|
| 125 |
+
"ppe_neg": 36,
|
| 126 |
+
"ppe_pos": 37,
|
| 127 |
+
"ppr_mixed": 38,
|
| 128 |
+
"ppr_neg": 39,
|
| 129 |
+
"ppr_pos": 40,
|
| 130 |
+
"qfm_mixed": 41,
|
| 131 |
+
"qfm_neg": 42,
|
| 132 |
+
"qfm_pos": 43,
|
| 133 |
+
"sct_mixed": 44,
|
| 134 |
+
"sct_neg": 45,
|
| 135 |
+
"sct_pos": 46,
|
| 136 |
+
"slohs_mixed": 47,
|
| 137 |
+
"slohs_neg": 48,
|
| 138 |
+
"slohs_pos": 49,
|
| 139 |
+
"sr_mixed": 50,
|
| 140 |
+
"sr_neg": 51,
|
| 141 |
+
"sr_pos": 52,
|
| 142 |
+
"td_mixed": 53,
|
| 143 |
+
"td_neg": 54,
|
| 144 |
+
"td_pos": 55,
|
| 145 |
+
"tshp_mixed": 56,
|
| 146 |
+
"tshp_neg": 57,
|
| 147 |
+
"tshp_pos": 58,
|
| 148 |
+
"wol_neg": 59,
|
| 149 |
+
"wol_pos": 60,
|
| 150 |
+
"wtc_mixed": 61,
|
| 151 |
+
"wtc_neg": 62,
|
| 152 |
+
"wtc_pos": 63,
|
| 153 |
+
"wtp_mixed": 64,
|
| 154 |
+
"wtp_neg": 65,
|
| 155 |
+
"wtp_pos": 66
|
| 156 |
+
},
|
| 157 |
+
"layer_norm_eps": 1e-07,
|
| 158 |
+
"max_position_embeddings": 512,
|
| 159 |
+
"num_attention_heads": 16,
|
| 160 |
+
"num_hidden_layers": 24,
|
| 161 |
+
"output_all_encoded_layers": true,
|
| 162 |
+
"position_bucket_size": 32,
|
| 163 |
+
"problem_type": "multi_label_classification",
|
| 164 |
+
"torch_dtype": "float32",
|
| 165 |
+
"transformers_version": "4.47.1",
|
| 166 |
+
"vocab_size": 50000
|
| 167 |
+
}
|
configuration_norbert.py
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from transformers.configuration_utils import PretrainedConfig
|
| 2 |
+
|
| 3 |
+
|
| 4 |
+
class NorbertConfig(PretrainedConfig):
|
| 5 |
+
"""Configuration class to store the configuration of a `NorbertModel`.
|
| 6 |
+
"""
|
| 7 |
+
def __init__(
|
| 8 |
+
self,
|
| 9 |
+
vocab_size=50000,
|
| 10 |
+
attention_probs_dropout_prob=0.1,
|
| 11 |
+
hidden_dropout_prob=0.1,
|
| 12 |
+
hidden_size=768,
|
| 13 |
+
intermediate_size=2048,
|
| 14 |
+
max_position_embeddings=512,
|
| 15 |
+
position_bucket_size=32,
|
| 16 |
+
num_attention_heads=12,
|
| 17 |
+
num_hidden_layers=12,
|
| 18 |
+
layer_norm_eps=1.0e-7,
|
| 19 |
+
output_all_encoded_layers=True,
|
| 20 |
+
**kwargs,
|
| 21 |
+
):
|
| 22 |
+
super().__init__(**kwargs)
|
| 23 |
+
|
| 24 |
+
self.vocab_size = vocab_size
|
| 25 |
+
self.hidden_size = hidden_size
|
| 26 |
+
self.num_hidden_layers = num_hidden_layers
|
| 27 |
+
self.num_attention_heads = num_attention_heads
|
| 28 |
+
self.intermediate_size = intermediate_size
|
| 29 |
+
self.hidden_dropout_prob = hidden_dropout_prob
|
| 30 |
+
self.attention_probs_dropout_prob = attention_probs_dropout_prob
|
| 31 |
+
self.max_position_embeddings = max_position_embeddings
|
| 32 |
+
self.output_all_encoded_layers = output_all_encoded_layers
|
| 33 |
+
self.position_bucket_size = position_bucket_size
|
| 34 |
+
self.layer_norm_eps = layer_norm_eps
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f7520b60d429ad82e43aeb05cde7a05cf2a95a2025d73319eb215fc1d382637c
|
| 3 |
+
size 1417921780
|