bdellabe commited on
Commit
c0a9c19
·
verified ·
1 Parent(s): a28bbd7

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -40,7 +40,7 @@
40
  "re:.*self_attn\\.(k|q|o|v)_proj$"
41
  ],
42
  "weights": {
43
- "actorder": "static",
44
  "block_structure": null,
45
  "dynamic": false,
46
  "group_size": null,
@@ -83,13 +83,13 @@
83
  "quantization_status": "compressed",
84
  "sparsity_config": {},
85
  "transform_config": {},
86
- "version": "0.11.1.dev11+g7f2c5de"
87
  },
88
  "rms_norm_eps": 1e-05,
89
  "rope_scaling": null,
90
  "rope_theta": 500000.0,
91
  "tie_word_embeddings": false,
92
- "transformers_version": "4.56.1",
93
  "use_cache": true,
94
  "vocab_size": 128256
95
  }
 
40
  "re:.*self_attn\\.(k|q|o|v)_proj$"
41
  ],
42
  "weights": {
43
+ "actorder": null,
44
  "block_structure": null,
45
  "dynamic": false,
46
  "group_size": null,
 
83
  "quantization_status": "compressed",
84
  "sparsity_config": {},
85
  "transform_config": {},
86
+ "version": "0.11.1.dev28+gdfd069b.d20250922"
87
  },
88
  "rms_norm_eps": 1e-05,
89
  "rope_scaling": null,
90
  "rope_theta": 500000.0,
91
  "tie_word_embeddings": false,
92
+ "transformers_version": "4.56.2",
93
  "use_cache": true,
94
  "vocab_size": 128256
95
  }
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  "max_length": 4096,
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
- "transformers_version": "4.56.1"
12
  }
 
8
  "max_length": 4096,
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
+ "transformers_version": "4.56.2"
12
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e2fca601892750197e3d7bf4c849cd6e8c9092cb3af26c88590b7e805dd3bab
3
  size 4974551800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0185fe6180d944380eb76a5cbc19a7890fa9e4d542630c70e2357da5a0950712
3
  size 4974551800
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0c9364ee8fa8b44ff2b784b74cd73b60a323411fff879e9c1d7c465c9d92d06
3
  size 1376884080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d452a4cfc97c983e2aa69c6b45c41e54118af83a1e0600e935864b03dd15a6d
3
  size 1376884080
recipe.yaml CHANGED
@@ -7,7 +7,7 @@ default_stage:
7
  sequential_update: true
8
  block_size: 128
9
  dampening_frac: 0.01
10
- actorder: !!python/object/apply:llmcompressor.sentinel.Sentinel [static, /home/bdellabe/projects/llm-compressor/src/llmcompressor/sentinel.py]
11
  offload_hessians: false
12
  AWQModifier:
13
  targets: ['re:.*mlp\.(down|gate|up)_proj$']
 
7
  sequential_update: true
8
  block_size: 128
9
  dampening_frac: 0.01
10
+ actorder: static
11
  offload_hessians: false
12
  AWQModifier:
13
  targets: ['re:.*mlp\.(down|gate|up)_proj$']