{ "_name_or_path": "allegro/herbert-base-cased", "architectures": [ "BertForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "O", "1": "B-NAME", "2": "I-NAME", "3": "B-SURNAME", "4": "I-SURNAME", "5": "B-AGE", "6": "I-AGE", "7": "B-DATE_BIRTH", "8": "I-DATE_BIRTH", "9": "B-DATE", "10": "I-DATE", "11": "B-SEX", "12": "I-SEX", "13": "B-RELIGION", "14": "I-RELIGION", "15": "B-POLITICAL", "16": "I-POLITICAL", "17": "B-ETHNICITY", "18": "I-ETHNICITY", "19": "B-ORIENTATION", "20": "I-ORIENTATION", "21": "B-HEALTH", "22": "I-HEALTH", "23": "B-RELATIVE", "24": "I-RELATIVE", "25": "B-CITY", "26": "I-CITY", "27": "B-ADDRESS", "28": "I-ADDRESS", "29": "B-EMAIL", "30": "I-EMAIL", "31": "B-PHONE", "32": "I-PHONE", "33": "B-PESEL", "34": "I-PESEL", "35": "B-DOCUMENT", "36": "I-DOCUMENT", "37": "B-COMPANY", "38": "I-COMPANY", "39": "B-SCHOOL", "40": "I-SCHOOL", "41": "B-JOB", "42": "I-JOB", "43": "B-BANK_ACCOUNT", "44": "I-BANK_ACCOUNT", "45": "B-CREDIT_CARD", "46": "I-CREDIT_CARD", "47": "B-USERNAME", "48": "I-USERNAME", "49": "B-SECRET", "50": "I-SECRET" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "B-ADDRESS": 27, "B-AGE": 5, "B-BANK_ACCOUNT": 43, "B-CITY": 25, "B-COMPANY": 37, "B-CREDIT_CARD": 45, "B-DATE": 9, "B-DATE_BIRTH": 7, "B-DOCUMENT": 35, "B-EMAIL": 29, "B-ETHNICITY": 17, "B-HEALTH": 21, "B-JOB": 41, "B-NAME": 1, "B-ORIENTATION": 19, "B-PESEL": 33, "B-PHONE": 31, "B-POLITICAL": 15, "B-RELATIVE": 23, "B-RELIGION": 13, "B-SCHOOL": 39, "B-SECRET": 49, "B-SEX": 11, "B-SURNAME": 3, "B-USERNAME": 47, "I-ADDRESS": 28, "I-AGE": 6, "I-BANK_ACCOUNT": 44, "I-CITY": 26, "I-COMPANY": 38, "I-CREDIT_CARD": 46, "I-DATE": 10, "I-DATE_BIRTH": 8, "I-DOCUMENT": 36, "I-EMAIL": 30, "I-ETHNICITY": 18, "I-HEALTH": 22, "I-JOB": 42, "I-NAME": 2, "I-ORIENTATION": 20, "I-PESEL": 34, "I-PHONE": 32, "I-POLITICAL": 16, "I-RELATIVE": 24, "I-RELIGION": 14, "I-SCHOOL": 40, "I-SECRET": 50, "I-SEX": 12, "I-SURNAME": 4, "I-USERNAME": 48, "O": 0 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 514, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "tokenizer_class": "HerbertTokenizerFast", "torch_dtype": "float32", "transformers_version": "4.47.0", "type_vocab_size": 2, "use_cache": true, "vocab_size": 50000 }