Push model using huggingface_hub.
Browse files- config.json +7 -1
- model.safetensors +1 -1
config.json
CHANGED
|
@@ -25,6 +25,7 @@
|
|
| 25 |
"num_shared_experts": 2,
|
| 26 |
"rope_base": 100000,
|
| 27 |
"router_amp": true,
|
|
|
|
| 28 |
"self_att_type": "sqa",
|
| 29 |
"seq_len": 8192,
|
| 30 |
"shared_expert_dim": 384,
|
|
@@ -85,12 +86,17 @@
|
|
| 85 |
"use_tanh_residual_gate": false
|
| 86 |
},
|
| 87 |
"memory_attention_variant": "grouped-self-interlayer",
|
|
|
|
| 88 |
"tokenizer": null,
|
| 89 |
"tokenizer_config": {
|
| 90 |
"answer_token_id": 6,
|
| 91 |
"bos_token_id": 2,
|
| 92 |
"eos_token_id": 3,
|
|
|
|
| 93 |
"pad_token_id": 0,
|
| 94 |
-
"query_token_id": 5
|
|
|
|
|
|
|
|
|
|
| 95 |
}
|
| 96 |
}
|
|
|
|
| 25 |
"num_shared_experts": 2,
|
| 26 |
"rope_base": 100000,
|
| 27 |
"router_amp": true,
|
| 28 |
+
"router_dtype": "bfloat16",
|
| 29 |
"self_att_type": "sqa",
|
| 30 |
"seq_len": 8192,
|
| 31 |
"shared_expert_dim": 384,
|
|
|
|
| 86 |
"use_tanh_residual_gate": false
|
| 87 |
},
|
| 88 |
"memory_attention_variant": "grouped-self-interlayer",
|
| 89 |
+
"system_prompt_title": "SYSTEM INSTRUCTIONS",
|
| 90 |
"tokenizer": null,
|
| 91 |
"tokenizer_config": {
|
| 92 |
"answer_token_id": 6,
|
| 93 |
"bos_token_id": 2,
|
| 94 |
"eos_token_id": 3,
|
| 95 |
+
"internal_token_id": 8,
|
| 96 |
"pad_token_id": 0,
|
| 97 |
+
"query_token_id": 5,
|
| 98 |
+
"think_token_id": 7,
|
| 99 |
+
"tool_call_token_id": 9,
|
| 100 |
+
"tool_use_token_id": 10
|
| 101 |
}
|
| 102 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 6099558592
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:432593c2cd1d4b547d599ba3768d89cc5b3b04b400bb28730020e5fa2f64b178
|
| 3 |
size 6099558592
|