{ "speculators_model_type": "mtp", "architectures": [ "MTPSpeculator" ], "num_speculative_steps": 3, "num_nextn_predict_layers": 1, "mtp_loss_step_weights": [ 0.51, 0.31, 0.18 ], "hidden_size": 2048, "intermediate_size": 5120, "num_attention_heads": 16, "num_key_value_heads": 2, "vocab_size": 151936, "max_position_embeddings": 262144, "rms_norm_eps": 1e-06, "transformer_config": { "model_type": "qwen3_next", "vocab_size": 151936, "hidden_size": 2048, "intermediate_size": 5120, "num_hidden_layers": 1, "num_attention_heads": 16, "num_key_value_heads": 2, "max_position_embeddings": 262144, "rms_norm_eps": 1e-06, "rope_theta": 10000000, "tie_word_embeddings": false, "attention_bias": true }, "speculators_config": { "algorithm": "mtp", "proposal_methods": [ { "proposal_type": "greedy", "speculative_tokens": 3, "verifier_accept_k": 1, "accept_tolerance": 0.0 } ], "default_proposal_method": "greedy", "verifier": { "name_or_path": "Qwen/Qwen3-Next-80B-A3B-Instruct", "architectures": ["Qwen3MoeForCausalLM"] } } }