krishnateja95 commited on
Commit
59f88be
·
verified ·
1 Parent(s): 097532a

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -88,9 +88,11 @@
88
  "num_bits": 8,
89
  "observer": null,
90
  "observer_kwargs": {},
 
91
  "strategy": "token",
92
  "symmetric": true,
93
- "type": "int"
 
94
  },
95
  "output_activations": null,
96
  "targets": [
@@ -104,9 +106,11 @@
104
  "num_bits": 8,
105
  "observer": "minmax",
106
  "observer_kwargs": {},
 
107
  "strategy": "channel",
108
  "symmetric": true,
109
- "type": "int"
 
110
  }
111
  }
112
  },
 
88
  "num_bits": 8,
89
  "observer": null,
90
  "observer_kwargs": {},
91
+ "scale_dtype": null,
92
  "strategy": "token",
93
  "symmetric": true,
94
+ "type": "int",
95
+ "zp_dtype": null
96
  },
97
  "output_activations": null,
98
  "targets": [
 
106
  "num_bits": 8,
107
  "observer": "minmax",
108
  "observer_kwargs": {},
109
+ "scale_dtype": null,
110
  "strategy": "channel",
111
  "symmetric": true,
112
+ "type": "int",
113
+ "zp_dtype": null
114
  }
115
  }
116
  },
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3e7a8a9aa0924de09973bd0fc1b37aa4ef75b78a7c73508c788467caa9ca4b8
3
  size 3819693440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:102ff3a935bb6c6431301567344814dc0d85cfd3069705334bc745e27b178db7
3
  size 3819693440
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c11d03fd82edfc467006a0dff085e4ca3225f87b0032fc934e3e9bc9083cd0c
3
  size 3818189840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22b246939898d4825ea90dd3ccc841a3a4a44519f7bdfd0df7ef820c7ba61ef9
3
  size 3818189840
recipe.yaml CHANGED
@@ -5,8 +5,6 @@ default_stage:
5
  mappings:
6
  - - ['re:.*self_attn.q_proj', 're:.*self_attn.k_proj', 're:.*self_attn.v_proj']
7
  - re:.*input_layernorm
8
- - - ['re:.*mamba.in_proj']
9
- - re:.*input_layernorm
10
  - - ['re:.*shared_mlp.input_linear', 're:.*block_sparse_moe\.input_linear\.experts\.\d+']
11
  - re:.*post_attention_layernorm
12
  ignore: []
 
5
  mappings:
6
  - - ['re:.*self_attn.q_proj', 're:.*self_attn.k_proj', 're:.*self_attn.v_proj']
7
  - re:.*input_layernorm
 
 
8
  - - ['re:.*shared_mlp.input_linear', 're:.*block_sparse_moe\.input_linear\.experts\.\d+']
9
  - re:.*post_attention_layernorm
10
  ignore: []