| { |
| "_name_or_path": "roberta-large-ttc", |
| "adapters": { |
| "adapters": {}, |
| "config_map": {}, |
| "fusion_config_map": {}, |
| "fusions": {} |
| }, |
| "architectures": [ |
| "RobertaForTokenClassification" |
| ], |
| "attention_probs_dropout_prob": 0.1, |
| "bos_token_id": 0, |
| "classifier_dropout": null, |
| "crf_state_dict": { |
| "_constraint_mask": [ |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 0.0 |
| ], |
| [ |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0 |
| ] |
| ], |
| "end_transitions": [ |
| 0.8981873393058777, |
| -1.6446324586868286, |
| 0.15959225594997406, |
| -1.065422773361206, |
| 0.08138537406921387, |
| 0.7564309239387512, |
| 0.1831926703453064 |
| ], |
| "start_transitions": [ |
| -1.4693340063095093, |
| -0.9572128653526306, |
| 0.21432557702064514, |
| 0.6372387409210205, |
| 0.9176182746887207, |
| 0.3472508192062378, |
| 0.4026823341846466 |
| ], |
| "transitions": [ |
| [ |
| -0.2170916646718979, |
| -0.3218635320663452, |
| 0.18208211660385132, |
| -0.12818297743797302, |
| 0.34018412232398987, |
| -0.48195144534111023, |
| -0.04889260604977608 |
| ], |
| [ |
| 0.7136948108673096, |
| -0.2505597174167633, |
| -0.38005825877189636, |
| -0.40438419580459595, |
| 0.41300636529922485, |
| -1.0484856367111206, |
| 0.42295384407043457 |
| ], |
| [ |
| 0.25146323442459106, |
| 0.1469486951828003, |
| -0.8152630925178528, |
| -0.38353827595710754, |
| 0.04554833471775055, |
| -0.15171411633491516, |
| -0.4340072572231293 |
| ], |
| [ |
| 0.6300148963928223, |
| -0.15564066171646118, |
| -0.01947428472340107, |
| -0.06949518620967865, |
| -0.6198987364768982, |
| -0.059607941657304764, |
| 0.1473998874425888 |
| ], |
| [ |
| 0.39721569418907166, |
| 0.3447458744049072, |
| -0.2558501064777374, |
| 0.19512270390987396, |
| 0.1556796282529831, |
| -0.32401445508003235, |
| -0.2883574068546295 |
| ], |
| [ |
| 0.36640703678131104, |
| 0.24679042398929596, |
| 0.08182600140571594, |
| -0.23832643032073975, |
| -0.03733787685632706, |
| -0.12589092552661896, |
| -0.1488221138715744 |
| ], |
| [ |
| 0.21845334768295288, |
| -0.14681760966777802, |
| 0.059233345091342926, |
| -0.251701682806015, |
| 0.04643300548195839, |
| -0.06884881108999252, |
| 0.1407143473625183 |
| ] |
| ] |
| }, |
| "eos_token_id": 2, |
| "hidden_act": "gelu", |
| "hidden_dropout_prob": 0.1, |
| "hidden_size": 1024, |
| "id2label": { |
| "0": "B-LOC", |
| "1": "B-ORG", |
| "2": "B-PER", |
| "3": "I-LOC", |
| "4": "I-ORG", |
| "5": "I-PER", |
| "6": "O" |
| }, |
| "initializer_range": 0.02, |
| "intermediate_size": 4096, |
| "label2id": { |
| "B-LOC": 0, |
| "B-ORG": 1, |
| "B-PER": 2, |
| "I-LOC": 3, |
| "I-ORG": 4, |
| "I-PER": 5, |
| "O": 6 |
| }, |
| "layer_norm_eps": 1e-05, |
| "max_position_embeddings": 514, |
| "model_type": "roberta", |
| "num_attention_heads": 16, |
| "num_hidden_layers": 24, |
| "pad_token_id": 1, |
| "position_embedding_type": "absolute", |
| "torch_dtype": "float32", |
| "transformers_version": "4.11.3", |
| "type_vocab_size": 1, |
| "use_cache": true, |
| "vocab_size": 50265 |
| } |
|
|