| { | |
| "_name_or_path": "roberta-large-ttc", | |
| "adapters": { | |
| "adapters": {}, | |
| "config_map": {}, | |
| "fusion_config_map": {}, | |
| "fusions": {} | |
| }, | |
| "architectures": [ | |
| "RobertaForTokenClassification" | |
| ], | |
| "attention_probs_dropout_prob": 0.1, | |
| "bos_token_id": 0, | |
| "classifier_dropout": null, | |
| "crf_state_dict": { | |
| "_constraint_mask": [ | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 1.0 | |
| ], | |
| [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 1.0, | |
| 0.0, | |
| 0.0 | |
| ], | |
| [ | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0 | |
| ] | |
| ], | |
| "end_transitions": [ | |
| 0.8981873393058777, | |
| -1.6446324586868286, | |
| 0.15959225594997406, | |
| -1.065422773361206, | |
| 0.08138537406921387, | |
| 0.7564309239387512, | |
| 0.1831926703453064 | |
| ], | |
| "start_transitions": [ | |
| -1.4693340063095093, | |
| -0.9572128653526306, | |
| 0.21432557702064514, | |
| 0.6372387409210205, | |
| 0.9176182746887207, | |
| 0.3472508192062378, | |
| 0.4026823341846466 | |
| ], | |
| "transitions": [ | |
| [ | |
| -0.2170916646718979, | |
| -0.3218635320663452, | |
| 0.18208211660385132, | |
| -0.12818297743797302, | |
| 0.34018412232398987, | |
| -0.48195144534111023, | |
| -0.04889260604977608 | |
| ], | |
| [ | |
| 0.7136948108673096, | |
| -0.2505597174167633, | |
| -0.38005825877189636, | |
| -0.40438419580459595, | |
| 0.41300636529922485, | |
| -1.0484856367111206, | |
| 0.42295384407043457 | |
| ], | |
| [ | |
| 0.25146323442459106, | |
| 0.1469486951828003, | |
| -0.8152630925178528, | |
| -0.38353827595710754, | |
| 0.04554833471775055, | |
| -0.15171411633491516, | |
| -0.4340072572231293 | |
| ], | |
| [ | |
| 0.6300148963928223, | |
| -0.15564066171646118, | |
| -0.01947428472340107, | |
| -0.06949518620967865, | |
| -0.6198987364768982, | |
| -0.059607941657304764, | |
| 0.1473998874425888 | |
| ], | |
| [ | |
| 0.39721569418907166, | |
| 0.3447458744049072, | |
| -0.2558501064777374, | |
| 0.19512270390987396, | |
| 0.1556796282529831, | |
| -0.32401445508003235, | |
| -0.2883574068546295 | |
| ], | |
| [ | |
| 0.36640703678131104, | |
| 0.24679042398929596, | |
| 0.08182600140571594, | |
| -0.23832643032073975, | |
| -0.03733787685632706, | |
| -0.12589092552661896, | |
| -0.1488221138715744 | |
| ], | |
| [ | |
| 0.21845334768295288, | |
| -0.14681760966777802, | |
| 0.059233345091342926, | |
| -0.251701682806015, | |
| 0.04643300548195839, | |
| -0.06884881108999252, | |
| 0.1407143473625183 | |
| ] | |
| ] | |
| }, | |
| "eos_token_id": 2, | |
| "hidden_act": "gelu", | |
| "hidden_dropout_prob": 0.1, | |
| "hidden_size": 1024, | |
| "id2label": { | |
| "0": "B-LOC", | |
| "1": "B-ORG", | |
| "2": "B-PER", | |
| "3": "I-LOC", | |
| "4": "I-ORG", | |
| "5": "I-PER", | |
| "6": "O" | |
| }, | |
| "initializer_range": 0.02, | |
| "intermediate_size": 4096, | |
| "label2id": { | |
| "B-LOC": 0, | |
| "B-ORG": 1, | |
| "B-PER": 2, | |
| "I-LOC": 3, | |
| "I-ORG": 4, | |
| "I-PER": 5, | |
| "O": 6 | |
| }, | |
| "layer_norm_eps": 1e-05, | |
| "max_position_embeddings": 514, | |
| "model_type": "roberta", | |
| "num_attention_heads": 16, | |
| "num_hidden_layers": 24, | |
| "pad_token_id": 1, | |
| "position_embedding_type": "absolute", | |
| "torch_dtype": "float32", | |
| "transformers_version": "4.11.3", | |
| "type_vocab_size": 1, | |
| "use_cache": true, | |
| "vocab_size": 50265 | |
| } | |