MBuechel's picture
Upload folder using huggingface_hub
0eb8bc6 verified
{
"_name_or_path": "FacebookAI/roberta-large",
"architectures": [
"RobertaForSequenceClassification"
],
"attention_probs_dropout_prob": 0.1,
"bos_token_id": 0,
"classifier_dropout": null,
"eos_token_id": 2,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 1024,
"id2label": {
"0": "T1003.001",
"1": "T1005",
"2": "T1012",
"3": "T1016",
"4": "T1021.001",
"5": "T1027",
"6": "T1033",
"7": "T1036.005",
"8": "T1041",
"9": "T1047",
"10": "T1053.005",
"11": "T1055",
"12": "T1056.001",
"13": "T1057",
"14": "T1059.003",
"15": "T1068",
"16": "T1070.004",
"17": "T1071.001",
"18": "T1072",
"19": "T1074.001",
"20": "T1078",
"21": "T1082",
"22": "T1083",
"23": "T1090",
"24": "T1095",
"25": "T1105",
"26": "T1106",
"27": "T1110",
"28": "T1112",
"29": "T1113",
"30": "T1140",
"31": "T1190",
"32": "T1204.002",
"33": "T1210",
"34": "T1218.011",
"35": "T1219",
"36": "T1484.001",
"37": "T1518.001",
"38": "T1543.003",
"39": "T1547.001",
"40": "T1548.002",
"41": "T1552.001",
"42": "T1557.001",
"43": "T1562.001",
"44": "T1564.001",
"45": "T1566.001",
"46": "T1569.002",
"47": "T1570",
"48": "T1573.001",
"49": "T1574.002"
},
"initializer_range": 0.02,
"intermediate_size": 4096,
"label2id": {
"T1003.001": 0,
"T1005": 1,
"T1012": 2,
"T1016": 3,
"T1021.001": 4,
"T1027": 5,
"T1033": 6,
"T1036.005": 7,
"T1041": 8,
"T1047": 9,
"T1053.005": 10,
"T1055": 11,
"T1056.001": 12,
"T1057": 13,
"T1059.003": 14,
"T1068": 15,
"T1070.004": 16,
"T1071.001": 17,
"T1072": 18,
"T1074.001": 19,
"T1078": 20,
"T1082": 21,
"T1083": 22,
"T1090": 23,
"T1095": 24,
"T1105": 25,
"T1106": 26,
"T1110": 27,
"T1112": 28,
"T1113": 29,
"T1140": 30,
"T1190": 31,
"T1204.002": 32,
"T1210": 33,
"T1218.011": 34,
"T1219": 35,
"T1484.001": 36,
"T1518.001": 37,
"T1543.003": 38,
"T1547.001": 39,
"T1548.002": 40,
"T1552.001": 41,
"T1557.001": 42,
"T1562.001": 43,
"T1564.001": 44,
"T1566.001": 45,
"T1569.002": 46,
"T1570": 47,
"T1573.001": 48,
"T1574.002": 49
},
"layer_norm_eps": 1e-05,
"max_position_embeddings": 514,
"model_type": "roberta",
"num_attention_heads": 16,
"num_hidden_layers": 24,
"pad_token_id": 1,
"position_embedding_type": "absolute",
"problem_type": "multi_label_classification",
"torch_dtype": "float32",
"transformers_version": "4.45.2",
"type_vocab_size": 1,
"use_cache": true,
"vocab_size": 50265
}