{ "architectures": [ "AlbertForSequenceClassification" ], "attention_probs_dropout_prob": 0, "bos_token_id": 2, "classifier_dropout_prob": 0.1, "down_scale_factor": 1, "dtype": "float32", "embedding_size": 128, "eos_token_id": 3, "gap_size": 0, "hidden_act": "gelu", "hidden_dropout_prob": 0, "hidden_size": 768, "id2label": { "0": "ACKNOWLEDGMENT", "1": "AFFIRMATION", "2": "AFIRMATION", "3": "APOLOGY", "4": "COMMAND", "5": "COMMENT", "6": "COMMENTS", "7": "COMPLAINT", "8": "COMPLIMENT", "9": "CONFIRMATION", "10": "DATA_ENTITY", "11": "DAT_ENTITY", "12": "EXPLANATION", "13": "FILLER", "14": "GREETING", "15": "GREETINGS", "16": "INFO", "17": "INFORMATION", "18": "NEGATE", "19": "NEGATICITY", "20": "NEGATIITY", "21": "NEGATIOITY", "22": "NEGATION", "23": "NEGATIVE", "24": "NEGATIVITY", "25": "NEGICATION", "26": "NEGIGATION", "27": "NEGILITY", "28": "NEGITION", "29": "NEGITIVE", "30": "NEGLECT", "31": "NEGLECTION", "32": "NEGLICATION", "33": "NEGLIGATION", "34": "NEGLIGENCE", "35": "OFFER_HELP", "36": "POSITION", "37": "POSITIVE", "38": "POSITIVITY", "39": "QUERY", "40": "QUESTION", "41": "QUESTIONS", "42": "REPEAT_REQUEST", "43": "REQUEST", "44": "STAAEMENT", "45": "STAEMENT", "46": "STATEMENT", "47": "STATUS", "48": "THANKS", "49": "TODO" }, "initializer_range": 0.02, "inner_group_num": 1, "intermediate_size": 3072, "label2id": { "ACKNOWLEDGMENT": 0, "AFFIRMATION": 1, "AFIRMATION": 2, "APOLOGY": 3, "COMMAND": 4, "COMMENT": 5, "COMMENTS": 6, "COMPLAINT": 7, "COMPLIMENT": 8, "CONFIRMATION": 9, "DATA_ENTITY": 10, "DAT_ENTITY": 11, "EXPLANATION": 12, "FILLER": 13, "GREETING": 14, "GREETINGS": 15, "INFO": 16, "INFORMATION": 17, "NEGATE": 18, "NEGATICITY": 19, "NEGATIITY": 20, "NEGATIOITY": 21, "NEGATION": 22, "NEGATIVE": 23, "NEGATIVITY": 24, "NEGICATION": 25, "NEGIGATION": 26, "NEGILITY": 27, "NEGITION": 28, "NEGITIVE": 29, "NEGLECT": 30, "NEGLECTION": 31, "NEGLICATION": 32, "NEGLIGATION": 33, "NEGLIGENCE": 34, "OFFER_HELP": 35, "POSITION": 36, "POSITIVE": 37, "POSITIVITY": 38, "QUERY": 39, "QUESTION": 40, "QUESTIONS": 41, "REPEAT_REQUEST": 42, "REQUEST": 43, "STAAEMENT": 44, "STAEMENT": 45, "STATEMENT": 46, "STATUS": 47, "THANKS": 48, "TODO": 49 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "albert", "net_structure_type": 0, "num_attention_heads": 12, "num_hidden_groups": 1, "num_hidden_layers": 12, "num_memory_blocks": 0, "pad_token_id": 0, "problem_type": "single_label_classification", "tie_word_embeddings": true, "transformers_version": "5.0.0", "type_vocab_size": 2, "use_cache": false, "vocab_size": 200000 }