Add files using upload-large-folder tool
Browse files- config.json +1 -1
- model-00072-of-00072.safetensors +2 -2
- model.safetensors.index.json +1 -1
- recipe.yaml +1 -1
config.json
CHANGED
|
@@ -87,4 +87,4 @@
|
|
| 87 |
"use_cache": true,
|
| 88 |
"use_qk_norm": true,
|
| 89 |
"vocab_size": 151552
|
| 90 |
-
}
|
|
|
|
| 87 |
"use_cache": true,
|
| 88 |
"use_qk_norm": true,
|
| 89 |
"vocab_size": 151552
|
| 90 |
+
}
|
model-00072-of-00072.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:696c035ae5eafddf7f688b04a41af9a14b35f22aee6396a61179e0715820d744
|
| 3 |
+
size 4210893856
|
model.safetensors.index.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
"total_parameters": 356906484992,
|
| 4 |
-
"total_size":
|
| 5 |
},
|
| 6 |
"weight_map": {
|
| 7 |
"lm_head.weight": "model-00072-of-00072.safetensors",
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
"total_parameters": 356906484992,
|
| 4 |
+
"total_size": 358812379904
|
| 5 |
},
|
| 6 |
"weight_map": {
|
| 7 |
"lm_head.weight": "model-00072-of-00072.safetensors",
|
recipe.yaml
CHANGED
|
@@ -3,5 +3,5 @@ default_stage:
|
|
| 3 |
QuantizationModifier:
|
| 4 |
targets: [Linear]
|
| 5 |
ignore: [lm_head, 're:.*mlp.gate$', 're:.*mlp.shared_expert_gate$', 're:.*linear_attn.*',
|
| 6 |
-
're:.*
|
| 7 |
scheme: FP8_DYNAMIC
|
|
|
|
| 3 |
QuantizationModifier:
|
| 4 |
targets: [Linear]
|
| 5 |
ignore: [lm_head, 're:.*mlp.gate$', 're:.*mlp.shared_expert_gate$', 're:.*linear_attn.*',
|
| 6 |
+
're:.*layers.92.shared_head.*', 're:.*layers.92.eh_proj.*']
|
| 7 |
scheme: FP8_DYNAMIC
|