Upload folder using huggingface_hub
Browse files- config.json +6 -2
- model-00001-of-00002.safetensors +1 -1
- model-00002-of-00002.safetensors +1 -1
- recipe.yaml +0 -2
config.json
CHANGED
|
@@ -88,9 +88,11 @@
|
|
| 88 |
"num_bits": 8,
|
| 89 |
"observer": null,
|
| 90 |
"observer_kwargs": {},
|
|
|
|
| 91 |
"strategy": "token",
|
| 92 |
"symmetric": true,
|
| 93 |
-
"type": "int"
|
|
|
|
| 94 |
},
|
| 95 |
"output_activations": null,
|
| 96 |
"targets": [
|
|
@@ -104,9 +106,11 @@
|
|
| 104 |
"num_bits": 8,
|
| 105 |
"observer": "minmax",
|
| 106 |
"observer_kwargs": {},
|
|
|
|
| 107 |
"strategy": "channel",
|
| 108 |
"symmetric": true,
|
| 109 |
-
"type": "int"
|
|
|
|
| 110 |
}
|
| 111 |
}
|
| 112 |
},
|
|
|
|
| 88 |
"num_bits": 8,
|
| 89 |
"observer": null,
|
| 90 |
"observer_kwargs": {},
|
| 91 |
+
"scale_dtype": null,
|
| 92 |
"strategy": "token",
|
| 93 |
"symmetric": true,
|
| 94 |
+
"type": "int",
|
| 95 |
+
"zp_dtype": null
|
| 96 |
},
|
| 97 |
"output_activations": null,
|
| 98 |
"targets": [
|
|
|
|
| 106 |
"num_bits": 8,
|
| 107 |
"observer": "minmax",
|
| 108 |
"observer_kwargs": {},
|
| 109 |
+
"scale_dtype": null,
|
| 110 |
"strategy": "channel",
|
| 111 |
"symmetric": true,
|
| 112 |
+
"type": "int",
|
| 113 |
+
"zp_dtype": null
|
| 114 |
}
|
| 115 |
}
|
| 116 |
},
|
model-00001-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3819693440
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:102ff3a935bb6c6431301567344814dc0d85cfd3069705334bc745e27b178db7
|
| 3 |
size 3819693440
|
model-00002-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3818189840
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22b246939898d4825ea90dd3ccc841a3a4a44519f7bdfd0df7ef820c7ba61ef9
|
| 3 |
size 3818189840
|
recipe.yaml
CHANGED
|
@@ -5,8 +5,6 @@ default_stage:
|
|
| 5 |
mappings:
|
| 6 |
- - ['re:.*self_attn.q_proj', 're:.*self_attn.k_proj', 're:.*self_attn.v_proj']
|
| 7 |
- re:.*input_layernorm
|
| 8 |
-
- - ['re:.*mamba.in_proj']
|
| 9 |
-
- re:.*input_layernorm
|
| 10 |
- - ['re:.*shared_mlp.input_linear', 're:.*block_sparse_moe\.input_linear\.experts\.\d+']
|
| 11 |
- re:.*post_attention_layernorm
|
| 12 |
ignore: []
|
|
|
|
| 5 |
mappings:
|
| 6 |
- - ['re:.*self_attn.q_proj', 're:.*self_attn.k_proj', 're:.*self_attn.v_proj']
|
| 7 |
- re:.*input_layernorm
|
|
|
|
|
|
|
| 8 |
- - ['re:.*shared_mlp.input_linear', 're:.*block_sparse_moe\.input_linear\.experts\.\d+']
|
| 9 |
- re:.*post_attention_layernorm
|
| 10 |
ignore: []
|