Upload LlamaForCausalLM
Browse files- config.json +43 -0
- generation_config.json +7 -0
- model-00001-of-00009.safetensors +3 -0
- model-00002-of-00009.safetensors +3 -0
- model-00003-of-00009.safetensors +3 -0
- model-00004-of-00009.safetensors +3 -0
- model-00005-of-00009.safetensors +3 -0
- model-00006-of-00009.safetensors +3 -0
- model-00007-of-00009.safetensors +3 -0
- model-00008-of-00009.safetensors +3 -0
- model-00009-of-00009.safetensors +3 -0
- model.safetensors.index.json +0 -0
config.json
ADDED
|
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "MultiVerse_70B_quant",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"LlamaForCausalLM"
|
| 5 |
+
],
|
| 6 |
+
"attention_bias": true,
|
| 7 |
+
"attention_dropout": 0.0,
|
| 8 |
+
"bos_token_id": 1,
|
| 9 |
+
"eos_token_id": 151643,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 8192,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 24576,
|
| 14 |
+
"max_position_embeddings": 32768,
|
| 15 |
+
"model_type": "llama",
|
| 16 |
+
"num_attention_heads": 64,
|
| 17 |
+
"num_hidden_layers": 80,
|
| 18 |
+
"num_key_value_heads": 64,
|
| 19 |
+
"pad_token_id": 151643,
|
| 20 |
+
"pretraining_tp": 1,
|
| 21 |
+
"quantization_config": {
|
| 22 |
+
"backend": "autoawq",
|
| 23 |
+
"bits": 4,
|
| 24 |
+
"do_fuse": false,
|
| 25 |
+
"exllama_config": null,
|
| 26 |
+
"fuse_max_seq_len": null,
|
| 27 |
+
"group_size": 128,
|
| 28 |
+
"modules_to_fuse": null,
|
| 29 |
+
"modules_to_not_convert": null,
|
| 30 |
+
"quant_method": "awq",
|
| 31 |
+
"version": "gemm",
|
| 32 |
+
"zero_point": true
|
| 33 |
+
},
|
| 34 |
+
"rms_norm_eps": 1e-06,
|
| 35 |
+
"rope_scaling": null,
|
| 36 |
+
"rope_theta": 1000000,
|
| 37 |
+
"seq_length": 32768,
|
| 38 |
+
"tie_word_embeddings": false,
|
| 39 |
+
"torch_dtype": "float16",
|
| 40 |
+
"transformers_version": "4.39.3",
|
| 41 |
+
"use_cache": false,
|
| 42 |
+
"vocab_size": 152064
|
| 43 |
+
}
|
generation_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"do_sample": true,
|
| 4 |
+
"eos_token_id": 151643,
|
| 5 |
+
"pad_token_id": 151643,
|
| 6 |
+
"transformers_version": "4.39.3"
|
| 7 |
+
}
|
model-00001-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:85a44056167fba6178f58a365498b9df6aef4e2b5375ee259c2adc32b65e8111
|
| 3 |
+
size 4999136232
|
model-00002-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a207677e05c5afc098a7a28690cc4c6e07c29d3d2ac58aab04874ce99d2026c9
|
| 3 |
+
size 4989977648
|
model-00003-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c040c9634ef84c58516929f75234b03fd3641c39080a702fe0bcf604bc55c405
|
| 3 |
+
size 4986831928
|
model-00004-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:533a079d179a245bbd3df2e4b735e04e285123ff1438ff613637434573faa2b8
|
| 3 |
+
size 4986831928
|
model-00005-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6f4985b071d424cb60356e28c7346046b98c6a4b76cb956244aae6443386808
|
| 3 |
+
size 4986831928
|
model-00006-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d3786d2df19622f49a9f1a6c2def16679fe8b51d53e9567e51f79d1ec83acb6
|
| 3 |
+
size 4986831928
|
model-00007-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:718fe87b78659e1fb035fc7bde269ed34b7f0b38ed1e84baa57d8edaa382ff6c
|
| 3 |
+
size 4986831928
|
model-00008-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2797d37201c744438df966f93e8d531001c04ac5fe02f3147dcaceec1eea6f5d
|
| 3 |
+
size 3836027912
|
model-00009-of-00009.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c6ac26139d540e4c56780d2e0cfb894b384fe5f33030e20d781c6be1939c0322
|
| 3 |
+
size 2491416704
|
model.safetensors.index.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|