{ "_name_or_path": "./roberta-base-ainu-pos/", "architectures": [ "RobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "PPX", "1": "PSX", "2": "PF", "3": "N", "4": "NL", "5": "PRN", "6": "NMLZ", "7": "COMP", "8": "PRP.N", "9": "VI", "10": "VT", "11": "VC", "12": "VD", "13": "AUX", "14": "ADV", "15": "ADV.PP", "16": "DEM", "17": "PP", "18": "ADV.PRT", "19": "ADV.CONJ", "20": "CONJ", "21": "FIN.PRT", "22": "NUM", "23": "N.INTERR", "24": "INTJ", "25": "PUNCT" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "ADV": 14, "ADV.CONJ": 19, "ADV.PP": 15, "ADV.PRT": 18, "AUX": 13, "COMP": 7, "CONJ": 20, "DEM": 16, "FIN.PRT": 21, "INTJ": 24, "N": 3, "N.INTERR": 23, "NL": 4, "NMLZ": 6, "NUM": 22, "PF": 2, "PP": 17, "PPX": 0, "PRN": 5, "PRP.N": 8, "PSX": 1, "PUNCT": 25, "VC": 11, "VD": 12, "VI": 9, "VT": 10 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.38.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }