{ "model_type": "labin", "architecture": "BiLSTM + SeqWeightedAttention (Keras)", "vocab_size": 40, "embedding_dim": 128, "maxlen": 64, "lstm_size": 128, "dropout": 0.5, "padding": "left", "num_classes": 2, "id2label": {"0": "legit", "1": "dga"}, "label2id": {"legit": 0, "dga": 1}, "framework": "keras", "weights_file": "LABin_best_model.keras", "train_families": 54, "train_rows": 845639, "dependencies": ["keras", "keras_self_attention"] }