| { |
| "architectures": [ |
| "MobileBertForTokenClassification" |
| ], |
| "attention_probs_dropout_prob": 0.1, |
| "classifier_activation": false, |
| "embedding_size": 128, |
| "hidden_act": "relu", |
| "hidden_dropout_prob": 0.0, |
| "hidden_size": 512, |
| "id2label": { |
| "0": "O", |
| "1": "CC", |
| "2": "CD", |
| "3": "DT", |
| "4": "EX", |
| "5": "FW", |
| "6": "IN", |
| "7": "JJ", |
| "8": "JJR", |
| "9": "JJS", |
| "10": "MD", |
| "11": "NN", |
| "12": "NNP", |
| "13": "NNPS", |
| "14": "NNS", |
| "15": "PDT", |
| "16": "POS", |
| "17": "PRP", |
| "18": "RB", |
| "19": "RBR", |
| "20": "RBS", |
| "21": "RP", |
| "22": "SYM", |
| "23": "TO", |
| "24": "UH", |
| "25": "VB", |
| "26": "VBD", |
| "27": "VBG", |
| "28": "VBN", |
| "29": "VBP", |
| "30": "VBZ", |
| "31": "WDT", |
| "32": "WP", |
| "33": "WRB" |
| }, |
| "initializer_range": 0.02, |
| "intermediate_size": 512, |
| "intra_bottleneck_size": 128, |
| "key_query_shared_bottleneck": true, |
| "label2id": { |
| "CC": 1, |
| "CD": 2, |
| "DT": 3, |
| "EX": 4, |
| "FW": 5, |
| "IN": 6, |
| "JJ": 7, |
| "JJR": 8, |
| "JJS": 9, |
| "MD": 10, |
| "NN": 11, |
| "NNP": 12, |
| "NNPS": 13, |
| "NNS": 14, |
| "O": 0, |
| "PDT": 15, |
| "POS": 16, |
| "PRP": 17, |
| "RB": 18, |
| "RBR": 19, |
| "RBS": 20, |
| "RP": 21, |
| "SYM": 22, |
| "TO": 23, |
| "UH": 24, |
| "VB": 25, |
| "VBD": 26, |
| "VBG": 27, |
| "VBN": 28, |
| "VBP": 29, |
| "VBZ": 30, |
| "WDT": 31, |
| "WP": 32, |
| "WRB": 33 |
| }, |
| "layer_norm_eps": 1e-12, |
| "max_position_embeddings": 512, |
| "model_type": "mobilebert", |
| "normalization_type": "no_norm", |
| "num_attention_heads": 4, |
| "num_feedforward_networks": 4, |
| "num_hidden_layers": 24, |
| "pad_token_id": 0, |
| "trigram_input": true, |
| "true_hidden_size": 128, |
| "type_vocab_size": 2, |
| "use_bottleneck": true, |
| "use_bottleneck_attention": false, |
| "vocab_size": 30522 |
| } |
|
|