{ "_name_or_path": "roberta-base", "architectures": [ "BertForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "NCT03361748", "1": "NCT01660633", "2": "NCT00103662", "3": "NCT00430365", "4": "NCT02874742", "5": "NCT04181827", "6": "NCT01311687", "7": "NCT03651128", "8": "NCT02336815", "9": "NCT00114101", "10": "NCT01985126", "11": "NCT03158688", "12": "NCT02252172", "13": "NCT05066646", "14": "NCT01568866", "15": "NCT01712789", "16": "NCT03525678", "17": "NCT03180736", "18": "NCT05083169", "19": "NCT01564537", "20": "NCT02136134", "21": "NCT03548207", "22": "NCT01554852", "23": "NCT03275285", "24": "NCT01080391", "25": "NCT03412565", "26": "NCT02963493", "27": "NCT03151811", "28": "NCT04162210", "29": "NCT03277105", "30": "NCT02412878", "31": "NCT03399799", "32": "NCT04557098", "33": "NCT02076009", "34": "NCT02541383", "35": "NCT01023308", "36": "ISRCTN17354232", "37": "NCT01239797", "38": "NCT02990338", "39": "NCT03110562", "40": "NCT00551928", "41": "NCT02195479", "42": "NCT01998971", "43": "NCT02654132", "44": "NCT01734928" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "ISRCTN17354232": 36, "NCT00103662": 2, "NCT00114101": 9, "NCT00430365": 3, "NCT00551928": 40, "NCT01023308": 35, "NCT01080391": 24, "NCT01239797": 37, "NCT01311687": 6, "NCT01554852": 22, "NCT01564537": 19, "NCT01568866": 14, "NCT01660633": 1, "NCT01712789": 15, "NCT01734928": 44, "NCT01985126": 10, "NCT01998971": 42, "NCT02076009": 33, "NCT02136134": 20, "NCT02195479": 41, "NCT02252172": 12, "NCT02336815": 8, "NCT02412878": 30, "NCT02541383": 34, "NCT02654132": 43, "NCT02874742": 4, "NCT02963493": 26, "NCT02990338": 38, "NCT03110562": 39, "NCT03151811": 27, "NCT03158688": 11, "NCT03180736": 17, "NCT03275285": 23, "NCT03277105": 29, "NCT03361748": 0, "NCT03399799": 31, "NCT03412565": 25, "NCT03525678": 16, "NCT03548207": 21, "NCT03651128": 7, "NCT04162210": 28, "NCT04181827": 5, "NCT04557098": 32, "NCT05066646": 13, "NCT05083169": 18 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "multi_label_classification", "torch_dtype": "float32", "transformers_version": "4.33.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }