| { | |
| "_name_or_path": "roberta-large-btc", | |
| "adapters": { | |
| "adapters": {}, | |
| "config_map": {}, | |
| "fusion_config_map": {}, | |
| "fusions": {} | |
| }, | |
| "architectures": [ | |
| "RobertaForTokenClassification" | |
| ], | |
| "attention_probs_dropout_prob": 0.1, | |
| "bos_token_id": 0, | |
| "classifier_dropout": null, | |
| "crf_state_dict": { | |
| "_constraint_mask": [ | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 0.0 | |
| ], | |
| [ | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0 | |
| ] | |
| ], | |
| "end_transitions": [ | |
| -1.6918301582336426, | |
| 0.83443683385849, | |
| -1.369484543800354, | |
| 0.9246144890785217, | |
| 1.069233775138855, | |
| -0.16902348399162292, | |
| -1.1555670499801636 | |
| ], | |
| "start_transitions": [ | |
| 0.508945107460022, | |
| 0.8599593043327332, | |
| -0.7935769557952881, | |
| 2.3224387168884277, | |
| 1.7772325277328491, | |
| -0.14362987875938416, | |
| 1.1923292875289917 | |
| ], | |
| "transitions": [ | |
| [ | |
| -0.16531139612197876, | |
| -0.44062453508377075, | |
| -0.17101362347602844, | |
| 0.25618258118629456, | |
| 0.8218539953231812, | |
| -0.6791337132453918, | |
| -0.18991276621818542 | |
| ], | |
| [ | |
| 0.049584291875362396, | |
| -0.029006613418459892, | |
| 0.3484196066856384, | |
| 0.10896176099777222, | |
| 0.252639502286911, | |
| -0.29233261942863464, | |
| -0.09638770669698715 | |
| ], | |
| [ | |
| -0.38768860697746277, | |
| 0.08189011365175247, | |
| -0.310952365398407, | |
| -0.09067512303590775, | |
| -0.34503042697906494, | |
| 0.36938661336898804, | |
| 0.2102288007736206 | |
| ], | |
| [ | |
| 0.25750938057899475, | |
| 0.8332298398017883, | |
| -0.7139731049537659, | |
| -0.2698363959789276, | |
| -0.2478555291891098, | |
| -0.15140868723392487, | |
| -0.09971318393945694 | |
| ], | |
| [ | |
| -0.08811905235052109, | |
| 0.08423032611608505, | |
| -0.09703858941793442, | |
| 0.8387375473976135, | |
| 0.055684782564640045, | |
| 0.14125464856624603, | |
| -0.33670008182525635 | |
| ], | |
| [ | |
| 0.2373524159193039, | |
| 0.07985689491033554, | |
| 0.02303706295788288, | |
| 0.3609643578529358, | |
| 0.3878211975097656, | |
| -0.557222306728363, | |
| -0.34269899129867554 | |
| ], | |
| [ | |
| -0.26117709279060364, | |
| 0.45397937297821045, | |
| 1.2661653757095337, | |
| 0.4104919135570526, | |
| 0.13623365759849548, | |
| -0.05334918946027756, | |
| -0.017508944496512413 | |
| ] | |
| ] | |
| }, | |
| "eos_token_id": 2, | |
| "hidden_act": "gelu", | |
| "hidden_dropout_prob": 0.1, | |
| "hidden_size": 1024, | |
| "id2label": { | |
| "0": "B-LOC", | |
| "1": "B-ORG", | |
| "2": "B-PER", | |
| "3": "I-LOC", | |
| "4": "I-ORG", | |
| "5": "I-PER", | |
| "6": "O" | |
| }, | |
| "initializer_range": 0.02, | |
| "intermediate_size": 4096, | |
| "label2id": { | |
| "B-LOC": 0, | |
| "B-ORG": 1, | |
| "B-PER": 2, | |
| "I-LOC": 3, | |
| "I-ORG": 4, | |
| "I-PER": 5, | |
| "O": 6 | |
| }, | |
| "layer_norm_eps": 1e-05, | |
| "max_position_embeddings": 514, | |
| "model_type": "roberta", | |
| "num_attention_heads": 16, | |
| "num_hidden_layers": 24, | |
| "pad_token_id": 1, | |
| "position_embedding_type": "absolute", | |
| "torch_dtype": "float32", | |
| "transformers_version": "4.11.3", | |
| "type_vocab_size": 1, | |
| "use_cache": true, | |
| "vocab_size": 50265 | |
| } | |