{ "_name_or_path": "FacebookAI/roberta-large", "architectures": [ "RobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "T1003.001", "1": "T1005", "2": "T1012", "3": "T1016", "4": "T1021.001", "5": "T1027", "6": "T1033", "7": "T1036.005", "8": "T1041", "9": "T1047", "10": "T1053.005", "11": "T1055", "12": "T1056.001", "13": "T1057", "14": "T1059.003", "15": "T1068", "16": "T1070.004", "17": "T1071.001", "18": "T1072", "19": "T1074.001", "20": "T1078", "21": "T1082", "22": "T1083", "23": "T1090", "24": "T1095", "25": "T1105", "26": "T1106", "27": "T1110", "28": "T1112", "29": "T1113", "30": "T1140", "31": "T1190", "32": "T1204.002", "33": "T1210", "34": "T1218.011", "35": "T1219", "36": "T1484.001", "37": "T1518.001", "38": "T1543.003", "39": "T1547.001", "40": "T1548.002", "41": "T1552.001", "42": "T1557.001", "43": "T1562.001", "44": "T1564.001", "45": "T1566.001", "46": "T1569.002", "47": "T1570", "48": "T1573.001", "49": "T1574.002" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "T1003.001": 0, "T1005": 1, "T1012": 2, "T1016": 3, "T1021.001": 4, "T1027": 5, "T1033": 6, "T1036.005": 7, "T1041": 8, "T1047": 9, "T1053.005": 10, "T1055": 11, "T1056.001": 12, "T1057": 13, "T1059.003": 14, "T1068": 15, "T1070.004": 16, "T1071.001": 17, "T1072": 18, "T1074.001": 19, "T1078": 20, "T1082": 21, "T1083": 22, "T1090": 23, "T1095": 24, "T1105": 25, "T1106": 26, "T1110": 27, "T1112": 28, "T1113": 29, "T1140": 30, "T1190": 31, "T1204.002": 32, "T1210": 33, "T1218.011": 34, "T1219": 35, "T1484.001": 36, "T1518.001": 37, "T1543.003": 38, "T1547.001": 39, "T1548.002": 40, "T1552.001": 41, "T1557.001": 42, "T1562.001": 43, "T1564.001": 44, "T1566.001": 45, "T1569.002": 46, "T1570": 47, "T1573.001": 48, "T1574.002": 49 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "multi_label_classification", "torch_dtype": "float32", "transformers_version": "4.45.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }