Kiy-K commited on
Commit
6c71433
·
verified ·
1 Parent(s): 4deb51f

Push model via Rescue Script

Browse files
config.json CHANGED
@@ -5,7 +5,7 @@
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 128000,
8
- "dtype": "bfloat16",
9
  "eos_token_id": 128012,
10
  "hidden_act": "silu",
11
  "hidden_size": 2048,
@@ -101,7 +101,7 @@
101
  "rope_scaling": null,
102
  "rope_theta": 5000000.0,
103
  "sliding_window": null,
104
- "transformers_version": "4.57.1",
105
  "use_cache": false,
106
  "use_sliding_window": false,
107
  "vocab_size": 128256
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 128000,
8
+ "dtype": "float16",
9
  "eos_token_id": 128012,
10
  "hidden_act": "silu",
11
  "hidden_size": 2048,
 
101
  "rope_scaling": null,
102
  "rope_theta": 5000000.0,
103
  "sliding_window": null,
104
+ "transformers_version": "4.57.2",
105
  "use_cache": false,
106
  "use_sliding_window": false,
107
  "vocab_size": 128256
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "pad_token_id": 128004,
6
  "temperature": 0.6,
7
  "top_p": 0.95,
8
- "transformers_version": "4.57.1"
9
  }
 
5
  "pad_token_id": 128004,
6
  "temperature": 0.6,
7
  "top_p": 0.95,
8
+ "transformers_version": "4.57.2"
9
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:842306e0e09a9903ffb2b8aa4a57cebd73941af3444ea744308f668c4fa5c990
3
- size 4966315264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3152cde72d12e632e806b0e862161fce792a526397f2125b1d244e4dae862ac
3
+ size 4966315008
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1cca596968ab73877da6e1a728916c48c010538565764f29c41185d50f4f9b79
3
- size 1709256488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1ef7dfd0c4f2aacb704215ea859f217c1cf8eb09e87d2b8c7dc239e5f3603a9
3
+ size 1183919680
model.safetensors.index.json CHANGED
@@ -1,10 +1,9 @@
1
  {
2
  "metadata": {
3
- "total_parameters": 3337766912,
4
- "total_size": 6675533824
5
  },
6
  "weight_map": {
7
- "lm_head.weight": "model-00002-of-00002.safetensors",
8
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
9
  "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
10
  "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_parameters": 3075098624,
4
+ "total_size": 6150197248
5
  },
6
  "weight_map": {
 
7
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
8
  "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
9
  "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",