{ "archi_type": "decoder", "attn_implementation": "flash_attention_2", "attn_type": "causal", "bidirectional": true, "lexical": false, "lora": true, "lora_r": 64, "model_name_or_path": "Qwen/Qwen3-8B", "model_type": "splade", "n_layers": null, "padding_side": "left", "train_head": false, "transformers_version": "4.53.3", "auto_map": { "AutoConfig": "splade.SpladeConfig", "AutoModelForCausalLM": "splade.Splade" } }