splade-code-8B / config.json
maxoul's picture
Update config.json
0d7419b verified
raw
history blame contribute delete
455 Bytes
{
"archi_type": "decoder",
"attn_implementation": "flash_attention_2",
"attn_type": "causal",
"bidirectional": true,
"lexical": false,
"lora": true,
"lora_r": 64,
"model_name_or_path": "Qwen/Qwen3-8B",
"model_type": "splade",
"n_layers": null,
"padding_side": "left",
"train_head": false,
"transformers_version": "4.53.3",
"auto_map": {
"AutoConfig": "splade.SpladeConfig",
"AutoModelForCausalLM": "splade.Splade"
}
}