Orpheus (es)
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +26 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/.gitattributes +47 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/README.md +79 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/source.txt +1 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/.gitattributes +36 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/README.md +45 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/config.json +35 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/generation_config.json +9 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00001-of-00003.safetensors +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00002-of-00003.safetensors +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00003-of-00003.safetensors +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/model.safetensors.index.json +261 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/source.txt +1 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/special_tokens_map.json +20 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer.json +3 -0
- es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer_config.json +0 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/.gitattributes +47 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/README.md +79 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/source.txt +1 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795/.gitattributes +36 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795/README.md +47 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795/config.json +35 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795/generation_config.json +9 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00001-of-00003.safetensors +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00002-of-00003.safetensors +3 -0
- es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00003-of-00003.safetensors +3 -0
.gitattributes
CHANGED
|
@@ -138,3 +138,29 @@ bn/orpheus-bangla-tts-gguf-8bit/model.gguf filter=lfs diff=lfs merge=lfs -text
|
|
| 138 |
bn/orpheus-bangla-tts-gguf/model.gguf filter=lfs diff=lfs merge=lfs -text
|
| 139 |
bn/orpheus-bangla-tts/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 140 |
pt/Orpheus_PTBR_FT_Unsloth_Chk-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 138 |
bn/orpheus-bangla-tts-gguf/model.gguf filter=lfs diff=lfs merge=lfs -text
|
| 139 |
bn/orpheus-bangla-tts/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 140 |
pt/Orpheus_PTBR_FT_Unsloth_Chk-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 141 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 142 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
| 143 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 144 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
| 145 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 146 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 147 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 148 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 149 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 150 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 151 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 152 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 153 |
+
es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 154 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 155 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
| 156 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 157 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
| 158 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 159 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 160 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 161 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 162 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 163 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 164 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 165 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 166 |
+
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/.gitattributes
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:996178dc6465f3f719be894020a9b8a6333743b6f1f3622f38682c7518237452
|
| 3 |
+
size 1914098112
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:74919e21f9a98bd9eb099c160de4a78a4da35a75b3baca03162f7b66c2661625
|
| 3 |
+
size 1437127104
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f88e9f8d61c783c7a4b3abc5d972f032bc3b71099b96f9fb32c50116eef4ca97
|
| 3 |
+
size 1888539072
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fa0008e9ee30fb8821c081fa1350d8f0a2e268d16ab98db660aae33300043385
|
| 3 |
+
size 1760350656
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f98df2ee39862529c602efa047f79ff75e21881a1c1fb8e97759412d6c30ed0a
|
| 3 |
+
size 1616040384
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b3e431008bb4674aa2726a92245e673499ede2d783e9a180e264d744f6f0623
|
| 3 |
+
size 2092569024
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:82a2f6527147bf55d9afe2b40a686a26d91353246943ea1099f1dba37c348b05
|
| 3 |
+
size 2001392064
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4487916de82f59107a98516792c07f85df7a7bfa71a6400ca4e093fc0a209f32
|
| 3 |
+
size 2395345344
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b65b902c6932857c255cacaaeb8ab5f80787f4a9b85072781320839d384f6b9
|
| 3 |
+
size 2342703552
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d33dc27859bbc1e7f57f6f5de65a24f0e1320f4ee2e0f517fb245c5a20c89282
|
| 3 |
+
size 2717045184
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f576d18f6a30040f7ccd194eaedf3aea5d437aaa26bf30bc4a4e6b602c8b4b72
|
| 3 |
+
size 3516431520
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec07fd901a912e415fb110851ee6e350148e2e2667a85f5d3b6cb15dadde8b18
|
| 3 |
+
size 6610830240
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/README.md
ADDED
|
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: marianbasti/Llama-3.2-3B-Orpheus-Chilean-1795
|
| 3 |
+
datasets:
|
| 4 |
+
- ylacombe/google-chilean-spanish
|
| 5 |
+
language:
|
| 6 |
+
- es
|
| 7 |
+
library_name: transformers
|
| 8 |
+
license: llama3.2
|
| 9 |
+
mradermacher:
|
| 10 |
+
readme_rev: 1
|
| 11 |
+
quantized_by: mradermacher
|
| 12 |
+
tags:
|
| 13 |
+
- text-to-speech
|
| 14 |
+
- chilean
|
| 15 |
+
- argentinian
|
| 16 |
+
- speech
|
| 17 |
+
---
|
| 18 |
+
## About
|
| 19 |
+
|
| 20 |
+
<!-- ### quantize_version: 2 -->
|
| 21 |
+
<!-- ### output_tensor_quantised: 1 -->
|
| 22 |
+
<!-- ### convert_type: hf -->
|
| 23 |
+
<!-- ### vocab_type: -->
|
| 24 |
+
<!-- ### tags: -->
|
| 25 |
+
<!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
|
| 26 |
+
<!-- ### quants_skip: -->
|
| 27 |
+
<!-- ### skip_mmproj: -->
|
| 28 |
+
static quants of https://huggingface.co/marianbasti/Llama-3.2-3B-Orpheus-Chilean-1795
|
| 29 |
+
|
| 30 |
+
<!-- provided-files -->
|
| 31 |
+
|
| 32 |
+
***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Llama-3.2-3B-Orpheus-Chilean-1795-GGUF).***
|
| 33 |
+
|
| 34 |
+
weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
|
| 35 |
+
## Usage
|
| 36 |
+
|
| 37 |
+
If you are unsure how to use GGUF files, refer to one of [TheBloke's
|
| 38 |
+
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
|
| 39 |
+
more details, including on how to concatenate multi-part files.
|
| 40 |
+
|
| 41 |
+
## Provided Quants
|
| 42 |
+
|
| 43 |
+
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
|
| 44 |
+
|
| 45 |
+
| Link | Type | Size/GB | Notes |
|
| 46 |
+
|:-----|:-----|--------:|:------|
|
| 47 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf) | Q2_K | 1.5 | |
|
| 48 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf) | Q3_K_S | 1.7 | |
|
| 49 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf) | Q3_K_M | 1.9 | lower quality |
|
| 50 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf) | Q3_K_L | 2.0 | |
|
| 51 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf) | IQ4_XS | 2.0 | |
|
| 52 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf) | Q4_K_S | 2.1 | fast, recommended |
|
| 53 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf) | Q4_K_M | 2.2 | fast, recommended |
|
| 54 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf) | Q5_K_S | 2.4 | |
|
| 55 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf) | Q5_K_M | 2.5 | |
|
| 56 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf) | Q6_K | 2.8 | very good quality |
|
| 57 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf) | Q8_0 | 3.6 | fast, best quality |
|
| 58 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf) | f16 | 6.7 | 16 bpw, overkill |
|
| 59 |
+
|
| 60 |
+
Here is a handy graph by ikawrakow comparing some lower-quality quant
|
| 61 |
+
types (lower is better):
|
| 62 |
+
|
| 63 |
+

|
| 64 |
+
|
| 65 |
+
And here are Artefact2's thoughts on the matter:
|
| 66 |
+
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
|
| 67 |
+
|
| 68 |
+
## FAQ / Model Request
|
| 69 |
+
|
| 70 |
+
See https://huggingface.co/mradermacher/model_requests for some answers to
|
| 71 |
+
questions you might have and/or if you want some other model quantized.
|
| 72 |
+
|
| 73 |
+
## Thanks
|
| 74 |
+
|
| 75 |
+
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
|
| 76 |
+
me use its servers and providing upgrades to my workstation to enable
|
| 77 |
+
this work in my free time.
|
| 78 |
+
|
| 79 |
+
<!-- end -->
|
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/source.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/.gitattributes
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/README.md
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: llama3.2
|
| 3 |
+
datasets:
|
| 4 |
+
- ylacombe/google-chilean-spanish
|
| 5 |
+
language:
|
| 6 |
+
- es
|
| 7 |
+
base_model:
|
| 8 |
+
- canopylabs/3b-es_it-ft-research_release
|
| 9 |
+
pipeline_tag: text-to-speech
|
| 10 |
+
library_name: transformers
|
| 11 |
+
tags:
|
| 12 |
+
- text-to-speech
|
| 13 |
+
- chilean
|
| 14 |
+
- argentinian
|
| 15 |
+
- speech
|
| 16 |
+
---
|
| 17 |
+
|
| 18 |
+
# Orpheus 3B Chilean Spanish finetune
|
| 19 |
+
|
| 20 |
+
Orpheus TTS is a Llama-based Speech-LLM designed for high-quality, empathetic text-to-speech generation. This model has been finetuned from [canopylabs/3b-es_it-ft-research_release](https://huggingface.co/canopylabs/3b-es_it-ft-research_release) with [ylacombe/google-argentinian-spanish](https://huggingface.co/datasets/ylacombe/google-argentinian-spanish) to deliver human-level, Chilean accent speech synthesis, achieving clarity, expressiveness, and real-time streaming performances.
|
| 21 |
+
|
| 22 |
+
# Model Details
|
| 23 |
+
|
| 24 |
+
### Model Capabilities
|
| 25 |
+
|
| 26 |
+
- **Human-Like Speech**: Natural intonation, emotion, and rhythm that is superior to SOTA closed source models
|
| 27 |
+
- **Zero-Shot Voice Cloning**: Clone voices without prior fine-tuning
|
| 28 |
+
- **Guided Emotion and Intonation**: Control speech and emotion characteristics with simple tags
|
| 29 |
+
- **Low Latency**: ~200ms streaming latency for realtime applications, reducible to ~100ms with input streaming
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
### Model Sources
|
| 33 |
+
|
| 34 |
+
- **GitHub Repo:** [https://github.com/canopyai/Orpheus-TTS](https://github.com/canopyai/Orpheus-TTS)
|
| 35 |
+
- **Blog Post:** [https://canopylabs.ai/model-releases](https://canopylabs.ai/releases/orpheus_can_speak_any_language)
|
| 36 |
+
- **Colab Inference Notebook:** [notebook link](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
# Usage
|
| 40 |
+
|
| 41 |
+
Check out Colab ([link to Colab](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)) or GitHub ([link to GitHub](https://github.com/canopyai/Orpheus-TTS)) on how to run easy inference on our finetuned models.
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
# Model Misuse
|
| 45 |
+
Do not use our models for impersonation without consent, misinformation or deception (including fake news or fraudulent calls), or any illegal or harmful activity. By using this model, you agree to follow all applicable laws and ethical guidelines. We disclaim responsibility for any use.
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/config.json
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"LlamaForCausalLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_bias": false,
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 128000,
|
| 8 |
+
"eos_token_id": 128001,
|
| 9 |
+
"head_dim": 128,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 3072,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 8192,
|
| 14 |
+
"max_position_embeddings": 131072,
|
| 15 |
+
"mlp_bias": false,
|
| 16 |
+
"model_type": "llama",
|
| 17 |
+
"num_attention_heads": 24,
|
| 18 |
+
"num_hidden_layers": 28,
|
| 19 |
+
"num_key_value_heads": 8,
|
| 20 |
+
"pretraining_tp": 1,
|
| 21 |
+
"rms_norm_eps": 1e-05,
|
| 22 |
+
"rope_scaling": {
|
| 23 |
+
"factor": 32.0,
|
| 24 |
+
"high_freq_factor": 4.0,
|
| 25 |
+
"low_freq_factor": 1.0,
|
| 26 |
+
"original_max_position_embeddings": 8192,
|
| 27 |
+
"rope_type": "llama3"
|
| 28 |
+
},
|
| 29 |
+
"rope_theta": 500000.0,
|
| 30 |
+
"tie_word_embeddings": true,
|
| 31 |
+
"torch_dtype": "float32",
|
| 32 |
+
"transformers_version": "4.51.3",
|
| 33 |
+
"use_cache": true,
|
| 34 |
+
"vocab_size": 156940
|
| 35 |
+
}
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/generation_config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 128000,
|
| 4 |
+
"do_sample": true,
|
| 5 |
+
"eos_token_id": 128001,
|
| 6 |
+
"temperature": 0.6,
|
| 7 |
+
"top_p": 0.9,
|
| 8 |
+
"transformers_version": "4.51.3"
|
| 9 |
+
}
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00001-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cd29794de25f926aec6810be3b4b16f34686fbbc9b0cb3f6da079c387001f267
|
| 3 |
+
size 4948557560
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00002-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f81717bad70f55bd9e331455dca986bfd01f5832f3cf089521ac0702bebd6871
|
| 3 |
+
size 4932808960
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00003-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c5e0647eb4c750c5279a379a6b447c036288de1ecd437936c2f08c083ad3a5db
|
| 3 |
+
size 3322130968
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/model.safetensors.index.json
ADDED
|
@@ -0,0 +1,261 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 13203468288
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
| 7 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 8 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 9 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 10 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 11 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 12 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 13 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 14 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 15 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 16 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 17 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 18 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 19 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 20 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 21 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 22 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 23 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 24 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 25 |
+
"model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 26 |
+
"model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 27 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 28 |
+
"model.layers.10.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 29 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 30 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 31 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 32 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 33 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 34 |
+
"model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 35 |
+
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 36 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 37 |
+
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 38 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 39 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 40 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 41 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 42 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 43 |
+
"model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 44 |
+
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 45 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 46 |
+
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 47 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 48 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 49 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 50 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 51 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 52 |
+
"model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 53 |
+
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 54 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 55 |
+
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 56 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 57 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 58 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 59 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 60 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 61 |
+
"model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 62 |
+
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 63 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 64 |
+
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 65 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 66 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 67 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 68 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 69 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 70 |
+
"model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 71 |
+
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 72 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 73 |
+
"model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 74 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 75 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 76 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 77 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 78 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 79 |
+
"model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 80 |
+
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 81 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 82 |
+
"model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 83 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 84 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 85 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 86 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 87 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 88 |
+
"model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 89 |
+
"model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 90 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 91 |
+
"model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 92 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 93 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 94 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 95 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 96 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 97 |
+
"model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 98 |
+
"model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 99 |
+
"model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 100 |
+
"model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 101 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 102 |
+
"model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 103 |
+
"model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 104 |
+
"model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 105 |
+
"model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 106 |
+
"model.layers.19.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 107 |
+
"model.layers.19.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 108 |
+
"model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 109 |
+
"model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 110 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 111 |
+
"model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 112 |
+
"model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 113 |
+
"model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 114 |
+
"model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 115 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 116 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 117 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 118 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 119 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 120 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 121 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 122 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 123 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 124 |
+
"model.layers.20.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 125 |
+
"model.layers.20.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 126 |
+
"model.layers.20.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 127 |
+
"model.layers.20.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 128 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 129 |
+
"model.layers.20.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 130 |
+
"model.layers.20.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 131 |
+
"model.layers.20.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 132 |
+
"model.layers.20.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 133 |
+
"model.layers.21.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 134 |
+
"model.layers.21.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 135 |
+
"model.layers.21.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 136 |
+
"model.layers.21.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 137 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 138 |
+
"model.layers.21.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 139 |
+
"model.layers.21.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 140 |
+
"model.layers.21.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 141 |
+
"model.layers.21.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 142 |
+
"model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 143 |
+
"model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 144 |
+
"model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 145 |
+
"model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 146 |
+
"model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 147 |
+
"model.layers.22.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 148 |
+
"model.layers.22.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 149 |
+
"model.layers.22.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 150 |
+
"model.layers.22.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 151 |
+
"model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 152 |
+
"model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 153 |
+
"model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 154 |
+
"model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 155 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 156 |
+
"model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 157 |
+
"model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 158 |
+
"model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 159 |
+
"model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 160 |
+
"model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 161 |
+
"model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 162 |
+
"model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 163 |
+
"model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 164 |
+
"model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 165 |
+
"model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 166 |
+
"model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 167 |
+
"model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 168 |
+
"model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 169 |
+
"model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 170 |
+
"model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 171 |
+
"model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 172 |
+
"model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 173 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 174 |
+
"model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 175 |
+
"model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 176 |
+
"model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 177 |
+
"model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 178 |
+
"model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 179 |
+
"model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 180 |
+
"model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 181 |
+
"model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 182 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 183 |
+
"model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 184 |
+
"model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 185 |
+
"model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 186 |
+
"model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 187 |
+
"model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 188 |
+
"model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 189 |
+
"model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 190 |
+
"model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 191 |
+
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 192 |
+
"model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 193 |
+
"model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 194 |
+
"model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 195 |
+
"model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 196 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 197 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 198 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 199 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 200 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 201 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 202 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 203 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 204 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 205 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 206 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 207 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 208 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 209 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 210 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 211 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 212 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 213 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 214 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 215 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 216 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 217 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 218 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 219 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 220 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 221 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 222 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 223 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 224 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 225 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 226 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 227 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 228 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 229 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 230 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 231 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 232 |
+
"model.layers.7.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 233 |
+
"model.layers.7.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 234 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 235 |
+
"model.layers.7.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 236 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 237 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 238 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 239 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 240 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 241 |
+
"model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 242 |
+
"model.layers.8.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 243 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 244 |
+
"model.layers.8.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 245 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 246 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 247 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 248 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 249 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 250 |
+
"model.layers.9.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 251 |
+
"model.layers.9.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 252 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 253 |
+
"model.layers.9.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 254 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 255 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 256 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 257 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 258 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 259 |
+
"model.norm.weight": "model-00003-of-00003.safetensors"
|
| 260 |
+
}
|
| 261 |
+
}
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/source.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
https://huggingface.co/surus-ai/Llama-3.2-3B-Orpheus-Chilean-1795
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/special_tokens_map.json
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
"<|audio|>"
|
| 4 |
+
],
|
| 5 |
+
"bos_token": {
|
| 6 |
+
"content": "<|begin_of_text|>",
|
| 7 |
+
"lstrip": false,
|
| 8 |
+
"normalized": false,
|
| 9 |
+
"rstrip": false,
|
| 10 |
+
"single_word": false
|
| 11 |
+
},
|
| 12 |
+
"eos_token": {
|
| 13 |
+
"content": "<|eot_id|>",
|
| 14 |
+
"lstrip": false,
|
| 15 |
+
"normalized": false,
|
| 16 |
+
"rstrip": false,
|
| 17 |
+
"single_word": false
|
| 18 |
+
},
|
| 19 |
+
"pad_token": "<custom_token_7>"
|
| 20 |
+
}
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fc3fecb199b4170636dbfab986d25f628157268d37b861f9cadaca60b1353bce
|
| 3 |
+
size 22849547
|
es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/.gitattributes
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b527db3e16a7c5cad98c45cd4401c4085072634609149e9eea923aea77ee73b2
|
| 3 |
+
size 1914098144
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9508cfe57a6fa532e445d27ec5d6ad7bca2d37ded14d34c88d8b2ff59e07997a
|
| 3 |
+
size 1437127136
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:720bb84dc1e38d25170993666281d7593f0e749ac56cba9aeec22c00b2c54659
|
| 3 |
+
size 1888539104
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:93c155656958444b4821d84fa54953b1af0b5b36d44f3ecb21a482f777e70b50
|
| 3 |
+
size 1760350688
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2575b5b4a0369ddaa652e0aac24cf40f3f3985f2e5155ab30d9538887362a2f8
|
| 3 |
+
size 1616040416
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:21f2e560a03a4ced4465b35aeb17ce79213ed0877eaf7b9e3999e8293153a63e
|
| 3 |
+
size 2092569056
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:53b46cf71b6ffcce203cffb7aead42e38d3a2b0ec73553e4dd0fa9395cc65d08
|
| 3 |
+
size 2001392096
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ea681229fc060241fa1631af7bbb2d221d5b531b064b54615787ee434b6d94a4
|
| 3 |
+
size 2395345376
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f84896516a69fa77d51b60115f84ab0c7f12c7c8ed6b041a99c6f5c2afcbefc
|
| 3 |
+
size 2342703584
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54a511153912ef2657b3f6abf039f6fc40e8c5e7150c5c4f73aad42121664054
|
| 3 |
+
size 2717045216
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ec829cc3cf7897bec66541d02c4164848cecbc37b48ac6f1b528f54f04ef850
|
| 3 |
+
size 3516431552
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ac99443dfe5d1f4b6c91d16aa89d2b90340d201b65edee7e41c0453592f1204e
|
| 3 |
+
size 6610830272
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/README.md
ADDED
|
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: surus-ai/Llama-3.2-3B-Orpheus-Rioplatense-1795
|
| 3 |
+
datasets:
|
| 4 |
+
- ylacombe/google-argentinian-spanish
|
| 5 |
+
language:
|
| 6 |
+
- es
|
| 7 |
+
library_name: transformers
|
| 8 |
+
license: llama3.2
|
| 9 |
+
mradermacher:
|
| 10 |
+
readme_rev: 1
|
| 11 |
+
quantized_by: mradermacher
|
| 12 |
+
tags:
|
| 13 |
+
- text-to-speech
|
| 14 |
+
- rioplatense
|
| 15 |
+
- argentinian
|
| 16 |
+
- speech
|
| 17 |
+
---
|
| 18 |
+
## About
|
| 19 |
+
|
| 20 |
+
<!-- ### quantize_version: 2 -->
|
| 21 |
+
<!-- ### output_tensor_quantised: 1 -->
|
| 22 |
+
<!-- ### convert_type: hf -->
|
| 23 |
+
<!-- ### vocab_type: -->
|
| 24 |
+
<!-- ### tags: -->
|
| 25 |
+
<!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
|
| 26 |
+
<!-- ### quants_skip: -->
|
| 27 |
+
<!-- ### skip_mmproj: -->
|
| 28 |
+
static quants of https://huggingface.co/surus-ai/Llama-3.2-3B-Orpheus-Rioplatense-1795
|
| 29 |
+
|
| 30 |
+
<!-- provided-files -->
|
| 31 |
+
|
| 32 |
+
***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF).***
|
| 33 |
+
|
| 34 |
+
weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
|
| 35 |
+
## Usage
|
| 36 |
+
|
| 37 |
+
If you are unsure how to use GGUF files, refer to one of [TheBloke's
|
| 38 |
+
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
|
| 39 |
+
more details, including on how to concatenate multi-part files.
|
| 40 |
+
|
| 41 |
+
## Provided Quants
|
| 42 |
+
|
| 43 |
+
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
|
| 44 |
+
|
| 45 |
+
| Link | Type | Size/GB | Notes |
|
| 46 |
+
|:-----|:-----|--------:|:------|
|
| 47 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf) | Q2_K | 1.5 | |
|
| 48 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf) | Q3_K_S | 1.7 | |
|
| 49 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf) | Q3_K_M | 1.9 | lower quality |
|
| 50 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf) | Q3_K_L | 2.0 | |
|
| 51 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf) | IQ4_XS | 2.0 | |
|
| 52 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf) | Q4_K_S | 2.1 | fast, recommended |
|
| 53 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf) | Q4_K_M | 2.2 | fast, recommended |
|
| 54 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf) | Q5_K_S | 2.4 | |
|
| 55 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf) | Q5_K_M | 2.5 | |
|
| 56 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf) | Q6_K | 2.8 | very good quality |
|
| 57 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf) | Q8_0 | 3.6 | fast, best quality |
|
| 58 |
+
| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf) | f16 | 6.7 | 16 bpw, overkill |
|
| 59 |
+
|
| 60 |
+
Here is a handy graph by ikawrakow comparing some lower-quality quant
|
| 61 |
+
types (lower is better):
|
| 62 |
+
|
| 63 |
+

|
| 64 |
+
|
| 65 |
+
And here are Artefact2's thoughts on the matter:
|
| 66 |
+
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
|
| 67 |
+
|
| 68 |
+
## FAQ / Model Request
|
| 69 |
+
|
| 70 |
+
See https://huggingface.co/mradermacher/model_requests for some answers to
|
| 71 |
+
questions you might have and/or if you want some other model quantized.
|
| 72 |
+
|
| 73 |
+
## Thanks
|
| 74 |
+
|
| 75 |
+
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
|
| 76 |
+
me use its servers and providing upgrades to my workstation to enable
|
| 77 |
+
this work in my free time.
|
| 78 |
+
|
| 79 |
+
<!-- end -->
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/source.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/.gitattributes
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/README.md
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: llama3.2
|
| 3 |
+
datasets:
|
| 4 |
+
- ylacombe/google-argentinian-spanish
|
| 5 |
+
language:
|
| 6 |
+
- es
|
| 7 |
+
base_model:
|
| 8 |
+
- canopylabs/3b-es_it-ft-research_release
|
| 9 |
+
pipeline_tag: text-to-speech
|
| 10 |
+
library_name: transformers
|
| 11 |
+
tags:
|
| 12 |
+
- text-to-speech
|
| 13 |
+
- rioplatense
|
| 14 |
+
- argentinian
|
| 15 |
+
- speech
|
| 16 |
+
---
|
| 17 |
+
|
| 18 |
+
# Orpheus 3B Rioplatense Spanish finetune
|
| 19 |
+
|
| 20 |
+
Orpheus TTS is a Llama-based Speech-LLM designed for high-quality, empathetic text-to-speech generation. This model has been finetuned from [canopylabs/3b-es_it-ft-research_release](https://huggingface.co/canopylabs/3b-es_it-ft-research_release) with [ylacombe/google-argentinian-spanish](https://huggingface.co/datasets/ylacombe/google-argentinian-spanish) to deliver human-level, Rioplatense accent speech synthesis, achieving clarity, expressiveness, and real-time streaming performances.
|
| 21 |
+
|
| 22 |
+
Find an inference API endpoint for this model at [surus](https://surus.dev/)! Sign up and get free credits to test out this and other SoTA LM models!
|
| 23 |
+
|
| 24 |
+
# Model Details
|
| 25 |
+
|
| 26 |
+
### Model Capabilities
|
| 27 |
+
|
| 28 |
+
- **Human-Like Speech**: Natural intonation, emotion, and rhythm that is superior to SOTA closed source models
|
| 29 |
+
- **Zero-Shot Voice Cloning**: Clone voices without prior fine-tuning
|
| 30 |
+
- **Guided Emotion and Intonation**: Control speech and emotion characteristics with simple tags
|
| 31 |
+
- **Low Latency**: ~200ms streaming latency for realtime applications, reducible to ~100ms with input streaming
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
### Model Sources
|
| 35 |
+
|
| 36 |
+
- **GitHub Repo:** [https://github.com/canopyai/Orpheus-TTS](https://github.com/canopyai/Orpheus-TTS)
|
| 37 |
+
- **Blog Post:** [https://canopylabs.ai/model-releases](https://canopylabs.ai/releases/orpheus_can_speak_any_language)
|
| 38 |
+
- **Colab Inference Notebook:** [notebook link](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
# Usage
|
| 42 |
+
|
| 43 |
+
Check out Colab ([link to Colab](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)) or GitHub ([link to GitHub](https://github.com/canopyai/Orpheus-TTS)) on how to run easy inference on our finetuned models.
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
# Model Misuse
|
| 47 |
+
Do not use our models for impersonation without consent, misinformation or deception (including fake news or fraudulent calls), or any illegal or harmful activity. By using this model, you agree to follow all applicable laws and ethical guidelines. We disclaim responsibility for any use.
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/config.json
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"LlamaForCausalLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_bias": false,
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 128000,
|
| 8 |
+
"eos_token_id": 128001,
|
| 9 |
+
"head_dim": 128,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 3072,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 8192,
|
| 14 |
+
"max_position_embeddings": 131072,
|
| 15 |
+
"mlp_bias": false,
|
| 16 |
+
"model_type": "llama",
|
| 17 |
+
"num_attention_heads": 24,
|
| 18 |
+
"num_hidden_layers": 28,
|
| 19 |
+
"num_key_value_heads": 8,
|
| 20 |
+
"pretraining_tp": 1,
|
| 21 |
+
"rms_norm_eps": 1e-05,
|
| 22 |
+
"rope_scaling": {
|
| 23 |
+
"factor": 32.0,
|
| 24 |
+
"high_freq_factor": 4.0,
|
| 25 |
+
"low_freq_factor": 1.0,
|
| 26 |
+
"original_max_position_embeddings": 8192,
|
| 27 |
+
"rope_type": "llama3"
|
| 28 |
+
},
|
| 29 |
+
"rope_theta": 500000.0,
|
| 30 |
+
"tie_word_embeddings": true,
|
| 31 |
+
"torch_dtype": "float32",
|
| 32 |
+
"transformers_version": "4.51.3",
|
| 33 |
+
"use_cache": true,
|
| 34 |
+
"vocab_size": 156940
|
| 35 |
+
}
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/generation_config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 128000,
|
| 4 |
+
"do_sample": true,
|
| 5 |
+
"eos_token_id": 128001,
|
| 6 |
+
"temperature": 0.6,
|
| 7 |
+
"top_p": 0.9,
|
| 8 |
+
"transformers_version": "4.51.3"
|
| 9 |
+
}
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00001-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:21ad2c7e6c38f16944a9dfdd97344fcaf07101b1e3182498af02f5e87af45c88
|
| 3 |
+
size 4948557560
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00002-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ae8f40f3c4c8d11b39f3e4763438b7a805a9bc437b90abc8c252b15d613ae74
|
| 3 |
+
size 4932808960
|
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00003-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7733b0bb5134ad01ecd2f4f312a52897e2373a6503d78340dab46d2a83fc12a5
|
| 3 |
+
size 3322130968
|