niobures commited on
Commit
44e1a4b
·
verified ·
1 Parent(s): 762d607

Orpheus (es)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +26 -0
  2. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/.gitattributes +47 -0
  3. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf +3 -0
  4. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf +3 -0
  5. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf +3 -0
  6. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf +3 -0
  7. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf +3 -0
  8. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf +3 -0
  9. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf +3 -0
  10. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf +3 -0
  11. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf +3 -0
  12. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf +3 -0
  13. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf +3 -0
  14. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf +3 -0
  15. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/README.md +79 -0
  16. es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/source.txt +1 -0
  17. es/Llama-3.2-3B-Orpheus-Chilean-1795/.gitattributes +36 -0
  18. es/Llama-3.2-3B-Orpheus-Chilean-1795/README.md +45 -0
  19. es/Llama-3.2-3B-Orpheus-Chilean-1795/config.json +35 -0
  20. es/Llama-3.2-3B-Orpheus-Chilean-1795/generation_config.json +9 -0
  21. es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00001-of-00003.safetensors +3 -0
  22. es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00002-of-00003.safetensors +3 -0
  23. es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00003-of-00003.safetensors +3 -0
  24. es/Llama-3.2-3B-Orpheus-Chilean-1795/model.safetensors.index.json +261 -0
  25. es/Llama-3.2-3B-Orpheus-Chilean-1795/source.txt +1 -0
  26. es/Llama-3.2-3B-Orpheus-Chilean-1795/special_tokens_map.json +20 -0
  27. es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer.json +3 -0
  28. es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer_config.json +0 -0
  29. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/.gitattributes +47 -0
  30. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf +3 -0
  31. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf +3 -0
  32. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf +3 -0
  33. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf +3 -0
  34. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf +3 -0
  35. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf +3 -0
  36. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf +3 -0
  37. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf +3 -0
  38. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf +3 -0
  39. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf +3 -0
  40. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf +3 -0
  41. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf +3 -0
  42. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/README.md +79 -0
  43. es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/source.txt +1 -0
  44. es/Llama-3.2-3B-Orpheus-Rioplatense-1795/.gitattributes +36 -0
  45. es/Llama-3.2-3B-Orpheus-Rioplatense-1795/README.md +47 -0
  46. es/Llama-3.2-3B-Orpheus-Rioplatense-1795/config.json +35 -0
  47. es/Llama-3.2-3B-Orpheus-Rioplatense-1795/generation_config.json +9 -0
  48. es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00001-of-00003.safetensors +3 -0
  49. es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00002-of-00003.safetensors +3 -0
  50. es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00003-of-00003.safetensors +3 -0
.gitattributes CHANGED
@@ -138,3 +138,29 @@ bn/orpheus-bangla-tts-gguf-8bit/model.gguf filter=lfs diff=lfs merge=lfs -text
138
  bn/orpheus-bangla-tts-gguf/model.gguf filter=lfs diff=lfs merge=lfs -text
139
  bn/orpheus-bangla-tts/tokenizer.json filter=lfs diff=lfs merge=lfs -text
140
  pt/Orpheus_PTBR_FT_Unsloth_Chk-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
138
  bn/orpheus-bangla-tts-gguf/model.gguf filter=lfs diff=lfs merge=lfs -text
139
  bn/orpheus-bangla-tts/tokenizer.json filter=lfs diff=lfs merge=lfs -text
140
  pt/Orpheus_PTBR_FT_Unsloth_Chk-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
141
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
142
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
143
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
144
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
145
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
146
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
147
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
148
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
149
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
150
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
151
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
152
+ es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
153
+ es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer.json filter=lfs diff=lfs merge=lfs -text
154
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
155
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
156
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
157
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
158
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
159
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
160
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
161
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
162
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
163
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
164
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
165
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
166
+ es/Llama-3.2-3B-Orpheus-Rioplatense-1795/tokenizer.json filter=lfs diff=lfs merge=lfs -text
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/.gitattributes ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:996178dc6465f3f719be894020a9b8a6333743b6f1f3622f38682c7518237452
3
+ size 1914098112
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74919e21f9a98bd9eb099c160de4a78a4da35a75b3baca03162f7b66c2661625
3
+ size 1437127104
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f88e9f8d61c783c7a4b3abc5d972f032bc3b71099b96f9fb32c50116eef4ca97
3
+ size 1888539072
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa0008e9ee30fb8821c081fa1350d8f0a2e268d16ab98db660aae33300043385
3
+ size 1760350656
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f98df2ee39862529c602efa047f79ff75e21881a1c1fb8e97759412d6c30ed0a
3
+ size 1616040384
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b3e431008bb4674aa2726a92245e673499ede2d783e9a180e264d744f6f0623
3
+ size 2092569024
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82a2f6527147bf55d9afe2b40a686a26d91353246943ea1099f1dba37c348b05
3
+ size 2001392064
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4487916de82f59107a98516792c07f85df7a7bfa71a6400ca4e093fc0a209f32
3
+ size 2395345344
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b65b902c6932857c255cacaaeb8ab5f80787f4a9b85072781320839d384f6b9
3
+ size 2342703552
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d33dc27859bbc1e7f57f6f5de65a24f0e1320f4ee2e0f517fb245c5a20c89282
3
+ size 2717045184
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f576d18f6a30040f7ccd194eaedf3aea5d437aaa26bf30bc4a4e6b602c8b4b72
3
+ size 3516431520
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec07fd901a912e415fb110851ee6e350148e2e2667a85f5d3b6cb15dadde8b18
3
+ size 6610830240
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: marianbasti/Llama-3.2-3B-Orpheus-Chilean-1795
3
+ datasets:
4
+ - ylacombe/google-chilean-spanish
5
+ language:
6
+ - es
7
+ library_name: transformers
8
+ license: llama3.2
9
+ mradermacher:
10
+ readme_rev: 1
11
+ quantized_by: mradermacher
12
+ tags:
13
+ - text-to-speech
14
+ - chilean
15
+ - argentinian
16
+ - speech
17
+ ---
18
+ ## About
19
+
20
+ <!-- ### quantize_version: 2 -->
21
+ <!-- ### output_tensor_quantised: 1 -->
22
+ <!-- ### convert_type: hf -->
23
+ <!-- ### vocab_type: -->
24
+ <!-- ### tags: -->
25
+ <!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
26
+ <!-- ### quants_skip: -->
27
+ <!-- ### skip_mmproj: -->
28
+ static quants of https://huggingface.co/marianbasti/Llama-3.2-3B-Orpheus-Chilean-1795
29
+
30
+ <!-- provided-files -->
31
+
32
+ ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Llama-3.2-3B-Orpheus-Chilean-1795-GGUF).***
33
+
34
+ weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
35
+ ## Usage
36
+
37
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
38
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
39
+ more details, including on how to concatenate multi-part files.
40
+
41
+ ## Provided Quants
42
+
43
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
44
+
45
+ | Link | Type | Size/GB | Notes |
46
+ |:-----|:-----|--------:|:------|
47
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q2_K.gguf) | Q2_K | 1.5 | |
48
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_S.gguf) | Q3_K_S | 1.7 | |
49
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_M.gguf) | Q3_K_M | 1.9 | lower quality |
50
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q3_K_L.gguf) | Q3_K_L | 2.0 | |
51
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.IQ4_XS.gguf) | IQ4_XS | 2.0 | |
52
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_S.gguf) | Q4_K_S | 2.1 | fast, recommended |
53
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q4_K_M.gguf) | Q4_K_M | 2.2 | fast, recommended |
54
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_S.gguf) | Q5_K_S | 2.4 | |
55
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q5_K_M.gguf) | Q5_K_M | 2.5 | |
56
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q6_K.gguf) | Q6_K | 2.8 | very good quality |
57
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.Q8_0.gguf) | Q8_0 | 3.6 | fast, best quality |
58
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Chilean-1795.f16.gguf) | f16 | 6.7 | 16 bpw, overkill |
59
+
60
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
61
+ types (lower is better):
62
+
63
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
64
+
65
+ And here are Artefact2's thoughts on the matter:
66
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
67
+
68
+ ## FAQ / Model Request
69
+
70
+ See https://huggingface.co/mradermacher/model_requests for some answers to
71
+ questions you might have and/or if you want some other model quantized.
72
+
73
+ ## Thanks
74
+
75
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
76
+ me use its servers and providing upgrades to my workstation to enable
77
+ this work in my free time.
78
+
79
+ <!-- end -->
es/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF/source.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Chilean-1795-GGUF
es/Llama-3.2-3B-Orpheus-Chilean-1795/.gitattributes ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
es/Llama-3.2-3B-Orpheus-Chilean-1795/README.md ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.2
3
+ datasets:
4
+ - ylacombe/google-chilean-spanish
5
+ language:
6
+ - es
7
+ base_model:
8
+ - canopylabs/3b-es_it-ft-research_release
9
+ pipeline_tag: text-to-speech
10
+ library_name: transformers
11
+ tags:
12
+ - text-to-speech
13
+ - chilean
14
+ - argentinian
15
+ - speech
16
+ ---
17
+
18
+ # Orpheus 3B Chilean Spanish finetune
19
+
20
+ Orpheus TTS is a Llama-based Speech-LLM designed for high-quality, empathetic text-to-speech generation. This model has been finetuned from [canopylabs/3b-es_it-ft-research_release](https://huggingface.co/canopylabs/3b-es_it-ft-research_release) with [ylacombe/google-argentinian-spanish](https://huggingface.co/datasets/ylacombe/google-argentinian-spanish) to deliver human-level, Chilean accent speech synthesis, achieving clarity, expressiveness, and real-time streaming performances.
21
+
22
+ # Model Details
23
+
24
+ ### Model Capabilities
25
+
26
+ - **Human-Like Speech**: Natural intonation, emotion, and rhythm that is superior to SOTA closed source models
27
+ - **Zero-Shot Voice Cloning**: Clone voices without prior fine-tuning
28
+ - **Guided Emotion and Intonation**: Control speech and emotion characteristics with simple tags
29
+ - **Low Latency**: ~200ms streaming latency for realtime applications, reducible to ~100ms with input streaming
30
+
31
+
32
+ ### Model Sources
33
+
34
+ - **GitHub Repo:** [https://github.com/canopyai/Orpheus-TTS](https://github.com/canopyai/Orpheus-TTS)
35
+ - **Blog Post:** [https://canopylabs.ai/model-releases](https://canopylabs.ai/releases/orpheus_can_speak_any_language)
36
+ - **Colab Inference Notebook:** [notebook link](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)
37
+
38
+
39
+ # Usage
40
+
41
+ Check out Colab ([link to Colab](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)) or GitHub ([link to GitHub](https://github.com/canopyai/Orpheus-TTS)) on how to run easy inference on our finetuned models.
42
+
43
+
44
+ # Model Misuse
45
+ Do not use our models for impersonation without consent, misinformation or deception (including fake news or fraudulent calls), or any illegal or harmful activity. By using this model, you agree to follow all applicable laws and ethical guidelines. We disclaim responsibility for any use.
es/Llama-3.2-3B-Orpheus-Chilean-1795/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "eos_token_id": 128001,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 3072,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 8192,
14
+ "max_position_embeddings": 131072,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 24,
18
+ "num_hidden_layers": 28,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": {
23
+ "factor": 32.0,
24
+ "high_freq_factor": 4.0,
25
+ "low_freq_factor": 1.0,
26
+ "original_max_position_embeddings": 8192,
27
+ "rope_type": "llama3"
28
+ },
29
+ "rope_theta": 500000.0,
30
+ "tie_word_embeddings": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.51.3",
33
+ "use_cache": true,
34
+ "vocab_size": 156940
35
+ }
es/Llama-3.2-3B-Orpheus-Chilean-1795/generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "do_sample": true,
5
+ "eos_token_id": 128001,
6
+ "temperature": 0.6,
7
+ "top_p": 0.9,
8
+ "transformers_version": "4.51.3"
9
+ }
es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd29794de25f926aec6810be3b4b16f34686fbbc9b0cb3f6da079c387001f267
3
+ size 4948557560
es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f81717bad70f55bd9e331455dca986bfd01f5832f3cf089521ac0702bebd6871
3
+ size 4932808960
es/Llama-3.2-3B-Orpheus-Chilean-1795/model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5e0647eb4c750c5279a379a6b447c036288de1ecd437936c2f08c083ad3a5db
3
+ size 3322130968
es/Llama-3.2-3B-Orpheus-Chilean-1795/model.safetensors.index.json ADDED
@@ -0,0 +1,261 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 13203468288
4
+ },
5
+ "weight_map": {
6
+ "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
7
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
8
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
9
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
10
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
11
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
12
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
13
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
14
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
15
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
16
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
17
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
18
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
19
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
20
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
21
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
22
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
23
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
24
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
25
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
26
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
27
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
28
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
29
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
30
+ "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
31
+ "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
32
+ "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
33
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
34
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
35
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
36
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
37
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
38
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
39
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
40
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
41
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
42
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
43
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
44
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
45
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
46
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
47
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
48
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
49
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
50
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
51
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
52
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
53
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
54
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
55
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
56
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
57
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
58
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
59
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
60
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
61
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
62
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
63
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
64
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
65
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
66
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
67
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
68
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
69
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
70
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
71
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
72
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
73
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
74
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
75
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
76
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
77
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
78
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
79
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
80
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
81
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
82
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
83
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
84
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
85
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
86
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
88
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
89
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
90
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
91
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
92
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
93
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
94
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
95
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
96
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
97
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
98
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
99
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
100
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
101
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
102
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
103
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
104
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
105
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
106
+ "model.layers.19.input_layernorm.weight": "model-00003-of-00003.safetensors",
107
+ "model.layers.19.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
108
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
109
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
110
+ "model.layers.19.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
111
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
112
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
113
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
114
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
115
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
116
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
117
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
118
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
119
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
120
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
121
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
122
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
123
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
124
+ "model.layers.20.input_layernorm.weight": "model-00003-of-00003.safetensors",
125
+ "model.layers.20.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
126
+ "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
127
+ "model.layers.20.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
128
+ "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
129
+ "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
130
+ "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
131
+ "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
132
+ "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
133
+ "model.layers.21.input_layernorm.weight": "model-00003-of-00003.safetensors",
134
+ "model.layers.21.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
135
+ "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
136
+ "model.layers.21.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
137
+ "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
138
+ "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
139
+ "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
140
+ "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
141
+ "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
142
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
143
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
144
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
146
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
147
+ "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
148
+ "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
149
+ "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
150
+ "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
151
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
152
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
153
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
154
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
155
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
156
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
157
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
158
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
159
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
160
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
161
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
162
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
163
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
164
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
165
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
166
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
167
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
168
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
169
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
170
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
171
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
172
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
173
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
174
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
175
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
176
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
177
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
178
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
179
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
180
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
181
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
182
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
183
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
184
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
185
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
186
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
187
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
188
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
189
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
190
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
191
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
192
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
193
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
194
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
195
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
196
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
197
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
198
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
199
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
200
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
201
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
202
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
203
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
204
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
205
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
206
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
207
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
208
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
209
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
210
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
211
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
212
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
213
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
214
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
215
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
216
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
217
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
218
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
219
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
220
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
221
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
222
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
223
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
224
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
225
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
226
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
227
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
228
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
229
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
230
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
231
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
232
+ "model.layers.7.input_layernorm.weight": "model-00002-of-00003.safetensors",
233
+ "model.layers.7.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
234
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
235
+ "model.layers.7.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
236
+ "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
237
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
238
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
239
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
240
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
241
+ "model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
242
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
243
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
244
+ "model.layers.8.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
245
+ "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
246
+ "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
247
+ "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
248
+ "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
249
+ "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
250
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00003.safetensors",
251
+ "model.layers.9.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
252
+ "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
253
+ "model.layers.9.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
254
+ "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
255
+ "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
256
+ "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
257
+ "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
258
+ "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
259
+ "model.norm.weight": "model-00003-of-00003.safetensors"
260
+ }
261
+ }
es/Llama-3.2-3B-Orpheus-Chilean-1795/source.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ https://huggingface.co/surus-ai/Llama-3.2-3B-Orpheus-Chilean-1795
es/Llama-3.2-3B-Orpheus-Chilean-1795/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|audio|>"
4
+ ],
5
+ "bos_token": {
6
+ "content": "<|begin_of_text|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eos_token": {
13
+ "content": "<|eot_id|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "pad_token": "<custom_token_7>"
20
+ }
es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc3fecb199b4170636dbfab986d25f628157268d37b861f9cadaca60b1353bce
3
+ size 22849547
es/Llama-3.2-3B-Orpheus-Chilean-1795/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/.gitattributes ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf filter=lfs diff=lfs merge=lfs -text
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b527db3e16a7c5cad98c45cd4401c4085072634609149e9eea923aea77ee73b2
3
+ size 1914098144
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9508cfe57a6fa532e445d27ec5d6ad7bca2d37ded14d34c88d8b2ff59e07997a
3
+ size 1437127136
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:720bb84dc1e38d25170993666281d7593f0e749ac56cba9aeec22c00b2c54659
3
+ size 1888539104
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93c155656958444b4821d84fa54953b1af0b5b36d44f3ecb21a482f777e70b50
3
+ size 1760350688
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2575b5b4a0369ddaa652e0aac24cf40f3f3985f2e5155ab30d9538887362a2f8
3
+ size 1616040416
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21f2e560a03a4ced4465b35aeb17ce79213ed0877eaf7b9e3999e8293153a63e
3
+ size 2092569056
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53b46cf71b6ffcce203cffb7aead42e38d3a2b0ec73553e4dd0fa9395cc65d08
3
+ size 2001392096
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea681229fc060241fa1631af7bbb2d221d5b531b064b54615787ee434b6d94a4
3
+ size 2395345376
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f84896516a69fa77d51b60115f84ab0c7f12c7c8ed6b041a99c6f5c2afcbefc
3
+ size 2342703584
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54a511153912ef2657b3f6abf039f6fc40e8c5e7150c5c4f73aad42121664054
3
+ size 2717045216
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ec829cc3cf7897bec66541d02c4164848cecbc37b48ac6f1b528f54f04ef850
3
+ size 3516431552
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac99443dfe5d1f4b6c91d16aa89d2b90340d201b65edee7e41c0453592f1204e
3
+ size 6610830272
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: surus-ai/Llama-3.2-3B-Orpheus-Rioplatense-1795
3
+ datasets:
4
+ - ylacombe/google-argentinian-spanish
5
+ language:
6
+ - es
7
+ library_name: transformers
8
+ license: llama3.2
9
+ mradermacher:
10
+ readme_rev: 1
11
+ quantized_by: mradermacher
12
+ tags:
13
+ - text-to-speech
14
+ - rioplatense
15
+ - argentinian
16
+ - speech
17
+ ---
18
+ ## About
19
+
20
+ <!-- ### quantize_version: 2 -->
21
+ <!-- ### output_tensor_quantised: 1 -->
22
+ <!-- ### convert_type: hf -->
23
+ <!-- ### vocab_type: -->
24
+ <!-- ### tags: -->
25
+ <!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
26
+ <!-- ### quants_skip: -->
27
+ <!-- ### skip_mmproj: -->
28
+ static quants of https://huggingface.co/surus-ai/Llama-3.2-3B-Orpheus-Rioplatense-1795
29
+
30
+ <!-- provided-files -->
31
+
32
+ ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF).***
33
+
34
+ weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
35
+ ## Usage
36
+
37
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
38
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
39
+ more details, including on how to concatenate multi-part files.
40
+
41
+ ## Provided Quants
42
+
43
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
44
+
45
+ | Link | Type | Size/GB | Notes |
46
+ |:-----|:-----|--------:|:------|
47
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q2_K.gguf) | Q2_K | 1.5 | |
48
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_S.gguf) | Q3_K_S | 1.7 | |
49
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_M.gguf) | Q3_K_M | 1.9 | lower quality |
50
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q3_K_L.gguf) | Q3_K_L | 2.0 | |
51
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.IQ4_XS.gguf) | IQ4_XS | 2.0 | |
52
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_S.gguf) | Q4_K_S | 2.1 | fast, recommended |
53
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q4_K_M.gguf) | Q4_K_M | 2.2 | fast, recommended |
54
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_S.gguf) | Q5_K_S | 2.4 | |
55
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q5_K_M.gguf) | Q5_K_M | 2.5 | |
56
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q6_K.gguf) | Q6_K | 2.8 | very good quality |
57
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.Q8_0.gguf) | Q8_0 | 3.6 | fast, best quality |
58
+ | [GGUF](https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/resolve/main/Llama-3.2-3B-Orpheus-Rioplatense-1795.f16.gguf) | f16 | 6.7 | 16 bpw, overkill |
59
+
60
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
61
+ types (lower is better):
62
+
63
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
64
+
65
+ And here are Artefact2's thoughts on the matter:
66
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
67
+
68
+ ## FAQ / Model Request
69
+
70
+ See https://huggingface.co/mradermacher/model_requests for some answers to
71
+ questions you might have and/or if you want some other model quantized.
72
+
73
+ ## Thanks
74
+
75
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
76
+ me use its servers and providing upgrades to my workstation to enable
77
+ this work in my free time.
78
+
79
+ <!-- end -->
es/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF/source.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ https://huggingface.co/mradermacher/Llama-3.2-3B-Orpheus-Rioplatense-1795-GGUF
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/.gitattributes ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.2
3
+ datasets:
4
+ - ylacombe/google-argentinian-spanish
5
+ language:
6
+ - es
7
+ base_model:
8
+ - canopylabs/3b-es_it-ft-research_release
9
+ pipeline_tag: text-to-speech
10
+ library_name: transformers
11
+ tags:
12
+ - text-to-speech
13
+ - rioplatense
14
+ - argentinian
15
+ - speech
16
+ ---
17
+
18
+ # Orpheus 3B Rioplatense Spanish finetune
19
+
20
+ Orpheus TTS is a Llama-based Speech-LLM designed for high-quality, empathetic text-to-speech generation. This model has been finetuned from [canopylabs/3b-es_it-ft-research_release](https://huggingface.co/canopylabs/3b-es_it-ft-research_release) with [ylacombe/google-argentinian-spanish](https://huggingface.co/datasets/ylacombe/google-argentinian-spanish) to deliver human-level, Rioplatense accent speech synthesis, achieving clarity, expressiveness, and real-time streaming performances.
21
+
22
+ Find an inference API endpoint for this model at [surus](https://surus.dev/)! Sign up and get free credits to test out this and other SoTA LM models!
23
+
24
+ # Model Details
25
+
26
+ ### Model Capabilities
27
+
28
+ - **Human-Like Speech**: Natural intonation, emotion, and rhythm that is superior to SOTA closed source models
29
+ - **Zero-Shot Voice Cloning**: Clone voices without prior fine-tuning
30
+ - **Guided Emotion and Intonation**: Control speech and emotion characteristics with simple tags
31
+ - **Low Latency**: ~200ms streaming latency for realtime applications, reducible to ~100ms with input streaming
32
+
33
+
34
+ ### Model Sources
35
+
36
+ - **GitHub Repo:** [https://github.com/canopyai/Orpheus-TTS](https://github.com/canopyai/Orpheus-TTS)
37
+ - **Blog Post:** [https://canopylabs.ai/model-releases](https://canopylabs.ai/releases/orpheus_can_speak_any_language)
38
+ - **Colab Inference Notebook:** [notebook link](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)
39
+
40
+
41
+ # Usage
42
+
43
+ Check out Colab ([link to Colab](https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing)) or GitHub ([link to GitHub](https://github.com/canopyai/Orpheus-TTS)) on how to run easy inference on our finetuned models.
44
+
45
+
46
+ # Model Misuse
47
+ Do not use our models for impersonation without consent, misinformation or deception (including fake news or fraudulent calls), or any illegal or harmful activity. By using this model, you agree to follow all applicable laws and ethical guidelines. We disclaim responsibility for any use.
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "eos_token_id": 128001,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 3072,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 8192,
14
+ "max_position_embeddings": 131072,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 24,
18
+ "num_hidden_layers": 28,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": {
23
+ "factor": 32.0,
24
+ "high_freq_factor": 4.0,
25
+ "low_freq_factor": 1.0,
26
+ "original_max_position_embeddings": 8192,
27
+ "rope_type": "llama3"
28
+ },
29
+ "rope_theta": 500000.0,
30
+ "tie_word_embeddings": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.51.3",
33
+ "use_cache": true,
34
+ "vocab_size": 156940
35
+ }
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "do_sample": true,
5
+ "eos_token_id": 128001,
6
+ "temperature": 0.6,
7
+ "top_p": 0.9,
8
+ "transformers_version": "4.51.3"
9
+ }
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21ad2c7e6c38f16944a9dfdd97344fcaf07101b1e3182498af02f5e87af45c88
3
+ size 4948557560
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ae8f40f3c4c8d11b39f3e4763438b7a805a9bc437b90abc8c252b15d613ae74
3
+ size 4932808960
es/Llama-3.2-3B-Orpheus-Rioplatense-1795/model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7733b0bb5134ad01ecd2f4f312a52897e2373a6503d78340dab46d2a83fc12a5
3
+ size 3322130968