Meta-13-Qwen-7B / config.json
meta13sphere's picture
Upload folder using huggingface_hub
5064bc1 verified
raw
history blame contribute delete
288 Bytes
{
"architectures": [
"Meta13ForCausalLM"
],
"auto_map": {
"AutoConfig": "configuration_meta13.Meta13Config",
"AutoModelForCausalLM": "modeling_meta13.Meta13ForCausalLM"
},
"dtype": "float32",
"model_type": "meta13",
"transformers_version": "5.3.0"
}