{ "_name_or_path": "Lexemo/roberta_large_legal_act_extraction", "architectures": [ "RobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "O", "1": "abbreviation_relevant_following_act", "2": "abbreviation_relevant_previous_act", "3": "another_act", "4": "another_act_abbreviation", "5": "another_act_equal_previous_act", "6": "another_act_sequence_end", "7": "another_act_sequence_start", "8": "another_article_equal_previous_article", "9": "article_current", "10": "article_relevant_current_act", "11": "article_relevant_current_act_range_end", "12": "article_relevant_current_act_range_start", "13": "article_relevant_following_act", "14": "article_relevant_following_act_range_end", "15": "article_relevant_following_act_range_start", "16": "article_relevant_previous_act", "17": "article_relevant_previous_act_range_end", "18": "article_relevant_previous_act_range_start", "19": "current_act", "20": "treaty_abbreviation", "21": "treaty_name", "22": "service_label" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "O": 0, "abbreviation_relevant_following_act": 1, "abbreviation_relevant_previous_act": 2, "another_act": 3, "another_act_abbreviation": 4, "another_act_equal_previous_act": 5, "another_act_sequence_end": 6, "another_act_sequence_start": 7, "another_article_equal_previous_article": 8, "article_current": 9, "article_relevant_current_act": 10, "article_relevant_current_act_range_end": 11, "article_relevant_current_act_range_start": 12, "article_relevant_following_act": 13, "article_relevant_following_act_range_end": 14, "article_relevant_following_act_range_start": 15, "article_relevant_previous_act": 16, "article_relevant_previous_act_range_end": 17, "article_relevant_previous_act_range_start": 18, "current_act": 19, "service_label": 22, "treaty_abbreviation": 20, "treaty_name": 21 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.17.0", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }