reissbaker commited on
Commit
2c9171e
·
verified ·
1 Parent(s): 9af24b5

Add files using upload-large-folder tool

Browse files
config.json CHANGED
@@ -87,4 +87,4 @@
87
  "use_cache": true,
88
  "use_qk_norm": true,
89
  "vocab_size": 151552
90
- }
 
87
  "use_cache": true,
88
  "use_qk_norm": true,
89
  "vocab_size": 151552
90
+ }
model-00072-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf4cb8265754dfb34c5a74c1d2c4b9d25e8b71f6f88d0f8b343684ed4737b5d5
3
- size 4210924568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:696c035ae5eafddf7f688b04a41af9a14b35f22aee6396a61179e0715820d744
3
+ size 4210893856
model.safetensors.index.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "metadata": {
3
  "total_parameters": 356906484992,
4
- "total_size": 358812410624
5
  },
6
  "weight_map": {
7
  "lm_head.weight": "model-00072-of-00072.safetensors",
 
1
  {
2
  "metadata": {
3
  "total_parameters": 356906484992,
4
+ "total_size": 358812379904
5
  },
6
  "weight_map": {
7
  "lm_head.weight": "model-00072-of-00072.safetensors",
recipe.yaml CHANGED
@@ -3,5 +3,5 @@ default_stage:
3
  QuantizationModifier:
4
  targets: [Linear]
5
  ignore: [lm_head, 're:.*mlp.gate$', 're:.*mlp.shared_expert_gate$', 're:.*linear_attn.*',
6
- 're:.*shared_head.*', 're:.*eh_proj.*', 're:.*enorm.*', 're:.*hnorm.*']
7
  scheme: FP8_DYNAMIC
 
3
  QuantizationModifier:
4
  targets: [Linear]
5
  ignore: [lm_head, 're:.*mlp.gate$', 're:.*mlp.shared_expert_gate$', 're:.*linear_attn.*',
6
+ 're:.*layers.92.shared_head.*', 're:.*layers.92.eh_proj.*']
7
  scheme: FP8_DYNAMIC