{ "architectures": [ "RobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "dtype": "float32", "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "OU", "1": "OO", "2": ".O", "3": "!O", "4": ",O", "5": ".U", "6": "!U", "7": ",U", "8": ":O", "9": ";O", "10": ":U", "11": "'O", "12": "-O", "13": "?O", "14": "?U" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "!O": 3, "!U": 6, "'O": 11, ",O": 4, ",U": 7, "-O": 12, ".O": 2, ".U": 5, ":O": 8, ":U": 10, ";O": 9, "?O": 13, "?U": 14, "OO": 1, "OU": 0 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "transformers_version": "4.57.6", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }