Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- config.json +2 -2
- generation_config.json +2 -4
- pytorch_model-00001-of-00004.bin +3 -0
- pytorch_model-00002-of-00004.bin +3 -0
- pytorch_model-00003-of-00004.bin +3 -0
- pytorch_model-00004-of-00004.bin +3 -0
- pytorch_model.bin.index.json +0 -0
- unsloth.F16.gguf +3 -0
- unsloth.Q4_K_M.gguf +2 -2
.gitattributes
CHANGED
|
@@ -36,3 +36,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 36 |
unsloth.BF16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
unsloth.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 36 |
unsloth.BF16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
unsloth.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
unsloth.F16.gguf filter=lfs diff=lfs merge=lfs -text
|
config.json
CHANGED
|
@@ -5,7 +5,6 @@
|
|
| 5 |
"attention_bias": false,
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 128000,
|
| 8 |
-
"dtype": "float16",
|
| 9 |
"eos_token_id": 128001,
|
| 10 |
"head_dim": 128,
|
| 11 |
"hidden_act": "silu",
|
|
@@ -30,7 +29,8 @@
|
|
| 30 |
},
|
| 31 |
"rope_theta": 500000.0,
|
| 32 |
"tie_word_embeddings": false,
|
| 33 |
-
"
|
|
|
|
| 34 |
"unsloth_fixed": true,
|
| 35 |
"unsloth_version": "2025.9.2",
|
| 36 |
"use_cache": true,
|
|
|
|
| 5 |
"attention_bias": false,
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 128000,
|
|
|
|
| 8 |
"eos_token_id": 128001,
|
| 9 |
"head_dim": 128,
|
| 10 |
"hidden_act": "silu",
|
|
|
|
| 29 |
},
|
| 30 |
"rope_theta": 500000.0,
|
| 31 |
"tie_word_embeddings": false,
|
| 32 |
+
"torch_dtype": "float16",
|
| 33 |
+
"transformers_version": "4.55.4",
|
| 34 |
"unsloth_fixed": true,
|
| 35 |
"unsloth_version": "2025.9.2",
|
| 36 |
"use_cache": true,
|
generation_config.json
CHANGED
|
@@ -2,12 +2,10 @@
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 128000,
|
| 4 |
"do_sample": true,
|
| 5 |
-
"eos_token_id":
|
| 6 |
-
128001
|
| 7 |
-
],
|
| 8 |
"max_length": 131072,
|
| 9 |
"pad_token_id": 128004,
|
| 10 |
"temperature": 0.6,
|
| 11 |
"top_p": 0.9,
|
| 12 |
-
"transformers_version": "4.
|
| 13 |
}
|
|
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 128000,
|
| 4 |
"do_sample": true,
|
| 5 |
+
"eos_token_id": 128001,
|
|
|
|
|
|
|
| 6 |
"max_length": 131072,
|
| 7 |
"pad_token_id": 128004,
|
| 8 |
"temperature": 0.6,
|
| 9 |
"top_p": 0.9,
|
| 10 |
+
"transformers_version": "4.55.4"
|
| 11 |
}
|
pytorch_model-00001-of-00004.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e06a7b9f33c9b6b28d9eae43359d4d9f77579c114cb2725005ce8f3c3e25064f
|
| 3 |
+
size 4976718881
|
pytorch_model-00002-of-00004.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7aaf07f08a985c2570bd36e60cbf68d12a8841e1dc96b523a487f0976d3ba82c
|
| 3 |
+
size 4999827861
|
pytorch_model-00003-of-00004.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e50724e53a98faacbef92eaa91c4d49d21c628a8dd412349340dfe4786b879c
|
| 3 |
+
size 4915939585
|
pytorch_model-00004-of-00004.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ffce2e6aabe0303a299e4b7ead2a2e67ce82040893912b0555d39260aeab893d
|
| 3 |
+
size 1168141312
|
pytorch_model.bin.index.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
unsloth.F16.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a3af8226d8f6ad9d1c7a902872a44ba7300015472009a30f82ae12b644d85c0
|
| 3 |
+
size 16068891296
|
unsloth.Q4_K_M.gguf
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1bd4a1a903fde54ddd970ad107f99ef3d5b7624edabf8265bc70bc23f1028cb3
|
| 3 |
+
size 4920734368
|