lbs1163 commited on
Commit
4abc63d
·
verified ·
1 Parent(s): 6f250a5

sync main to W4V8

Browse files
.gitattributes CHANGED
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Llama-3.1-8B-Instruct.mxq filter=lfs diff=lfs merge=lfs -text
37
  Llama-3.1-8B-Instruct-W8.mxq filter=lfs diff=lfs merge=lfs -text
 
 
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Llama-3.1-8B-Instruct.mxq filter=lfs diff=lfs merge=lfs -text
37
  Llama-3.1-8B-Instruct-W8.mxq filter=lfs diff=lfs merge=lfs -text
38
+ Llama-3.1-8B-Instruct-W4V8.mxq filter=lfs diff=lfs merge=lfs -text
Llama-3.1-8B-Instruct-W8.mxq → Llama-3.1-8B-Instruct-W4V8.mxq RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38bfd2bb5101236b8e7a95210c226c26eefd99763afd9cb0ed1eff32d8088b7a
3
- size 7551802853
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e55d74b63c342473ac7afe1a175e823071c2cc9708e754c6941eb2c6c63e5ac6
3
+ size 3866355461
config.json CHANGED
@@ -22,7 +22,7 @@
22
  "max_position_embeddings": 4096,
23
  "mlp_bias": false,
24
  "model_type": "mobilint-llama",
25
- "mxq_path": "Llama-3.1-8B-Instruct-W8.mxq",
26
  "num_attention_heads": 32,
27
  "num_hidden_layers": 32,
28
  "num_key_value_heads": 8,
 
22
  "max_position_embeddings": 4096,
23
  "mlp_bias": false,
24
  "model_type": "mobilint-llama",
25
+ "mxq_path": "Llama-3.1-8B-Instruct-W4V8.mxq",
26
  "num_attention_heads": 32,
27
  "num_hidden_layers": 32,
28
  "num_key_value_heads": 8,