niobures commited on
Commit
dfdea44
·
verified ·
1 Parent(s): 2127851

Kani TTS ( en)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +36 -0
  2. en/kani-tts-450m-0.1-ft-GGUF/.gitattributes +47 -0
  3. en/kani-tts-450m-0.1-ft-GGUF/README.md +75 -0
  4. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.IQ4_XS.gguf +3 -0
  5. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q2_K.gguf +3 -0
  6. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_L.gguf +3 -0
  7. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_M.gguf +3 -0
  8. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_S.gguf +3 -0
  9. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q4_K_M.gguf +3 -0
  10. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q4_K_S.gguf +3 -0
  11. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q5_K_M.gguf +3 -0
  12. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q5_K_S.gguf +3 -0
  13. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q6_K.gguf +3 -0
  14. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q8_0.gguf +3 -0
  15. en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.f16.gguf +3 -0
  16. en/kani-tts-450m-0.1-ft-GGUF/source.txt +1 -0
  17. en/kani-tts-450m-0.1-ft/.gitattributes +35 -0
  18. en/kani-tts-450m-0.1-ft/README.md +113 -0
  19. en/kani-tts-450m-0.1-ft/chat_template.jinja +37 -0
  20. en/kani-tts-450m-0.1-ft/config.json +57 -0
  21. en/kani-tts-450m-0.1-ft/generation_config.json +7 -0
  22. en/kani-tts-450m-0.1-ft/model.safetensors +3 -0
  23. en/kani-tts-450m-0.1-ft/source.txt +1 -0
  24. en/kani-tts-450m-0.1-ft/special_tokens_map.json +23 -0
  25. en/kani-tts-450m-0.1-ft/tokenizer.json +0 -0
  26. en/kani-tts-450m-0.1-ft/tokenizer_config.json +0 -0
  27. en/kani-tts-450m-0.1-pt-GGUF/.gitattributes +47 -0
  28. en/kani-tts-450m-0.1-pt-GGUF/README.md +72 -0
  29. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.IQ4_XS.gguf +3 -0
  30. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q2_K.gguf +3 -0
  31. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_L.gguf +3 -0
  32. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_M.gguf +3 -0
  33. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_S.gguf +3 -0
  34. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q4_K_M.gguf +3 -0
  35. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q4_K_S.gguf +3 -0
  36. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q5_K_M.gguf +3 -0
  37. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q5_K_S.gguf +3 -0
  38. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q6_K.gguf +3 -0
  39. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q8_0.gguf +3 -0
  40. en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.f16.gguf +3 -0
  41. en/kani-tts-450m-0.1-pt-GGUF/source.txt +1 -0
  42. en/kani-tts-450m-0.2-ft-GGUF/.gitattributes +47 -0
  43. en/kani-tts-450m-0.2-ft-GGUF/README.md +75 -0
  44. en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.IQ4_XS.gguf +3 -0
  45. en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q2_K.gguf +3 -0
  46. en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_L.gguf +3 -0
  47. en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_M.gguf +3 -0
  48. en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_S.gguf +3 -0
  49. en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q4_K_M.gguf +3 -0
  50. en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q4_K_S.gguf +3 -0
.gitattributes CHANGED
@@ -58,3 +58,39 @@ ar,de,en,fr,ja,ko,zh/kani-tts-450m-0.2-pt-GGUF/kani-tts-450m-0.2-pt.Q5_K_S.gguf
58
  ar,de,en,fr,ja,ko,zh/kani-tts-450m-0.2-pt-GGUF/kani-tts-450m-0.2-pt.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
59
  ar,de,en,fr,ja,ko,zh/kani-tts-450m-0.2-pt-GGUF/kani-tts-450m-0.2-pt.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
60
  ar,de,en,fr,ja,ko,zh/kani-tts-450m-0.2-pt/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  ar,de,en,fr,ja,ko,zh/kani-tts-450m-0.2-pt-GGUF/kani-tts-450m-0.2-pt.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
59
  ar,de,en,fr,ja,ko,zh/kani-tts-450m-0.2-pt-GGUF/kani-tts-450m-0.2-pt.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
60
  ar,de,en,fr,ja,ko,zh/kani-tts-450m-0.2-pt/trainer_state.json filter=lfs diff=lfs merge=lfs -text
61
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.f16.gguf filter=lfs diff=lfs merge=lfs -text
62
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
63
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
64
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
65
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
66
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
67
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
68
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
69
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
70
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
71
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
72
+ en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
73
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.f16.gguf filter=lfs diff=lfs merge=lfs -text
74
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
75
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
76
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
77
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
78
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
79
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
80
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
81
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
82
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
83
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
84
+ en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
85
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.f16.gguf filter=lfs diff=lfs merge=lfs -text
86
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
87
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
88
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
89
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
90
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
91
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
92
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
93
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
94
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
95
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
96
+ en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
en/kani-tts-450m-0.1-ft-GGUF/.gitattributes ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ kani-tts-450m-0.1-ft.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
37
+ kani-tts-450m-0.1-ft.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
38
+ kani-tts-450m-0.1-ft.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
39
+ kani-tts-450m-0.1-ft.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ kani-tts-450m-0.1-ft.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
41
+ kani-tts-450m-0.1-ft.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ kani-tts-450m-0.1-ft.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ kani-tts-450m-0.1-ft.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ kani-tts-450m-0.1-ft.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
45
+ kani-tts-450m-0.1-ft.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
46
+ kani-tts-450m-0.1-ft.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
47
+ kani-tts-450m-0.1-ft.f16.gguf filter=lfs diff=lfs merge=lfs -text
en/kani-tts-450m-0.1-ft-GGUF/README.md ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: nineninesix/kani-tts-450m-0.1-ft
3
+ datasets:
4
+ - nytopop/expresso-conversational
5
+ - nineninesix/expresso-conversational-en-nano-codec-dataset
6
+ language:
7
+ - en
8
+ library_name: transformers
9
+ license: apache-2.0
10
+ mradermacher:
11
+ readme_rev: 1
12
+ quantized_by: mradermacher
13
+ ---
14
+ ## About
15
+
16
+ <!-- ### quantize_version: 2 -->
17
+ <!-- ### output_tensor_quantised: 1 -->
18
+ <!-- ### convert_type: hf -->
19
+ <!-- ### vocab_type: -->
20
+ <!-- ### tags: -->
21
+ <!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
22
+ <!-- ### quants_skip: -->
23
+ <!-- ### skip_mmproj: -->
24
+ static quants of https://huggingface.co/nineninesix/kani-tts-450m-0.1-ft
25
+
26
+ <!-- provided-files -->
27
+
28
+ ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#kani-tts-450m-0.1-ft-GGUF).***
29
+
30
+ weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
31
+ ## Usage
32
+
33
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
34
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
35
+ more details, including on how to concatenate multi-part files.
36
+
37
+ ## Provided Quants
38
+
39
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
40
+
41
+ | Link | Type | Size/GB | Notes |
42
+ |:-----|:-----|--------:|:------|
43
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q2_K.gguf) | Q2_K | 0.3 | |
44
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q3_K_S.gguf) | Q3_K_S | 0.3 | |
45
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q3_K_M.gguf) | Q3_K_M | 0.3 | lower quality |
46
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q3_K_L.gguf) | Q3_K_L | 0.3 | |
47
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.IQ4_XS.gguf) | IQ4_XS | 0.3 | |
48
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q4_K_S.gguf) | Q4_K_S | 0.3 | fast, recommended |
49
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q4_K_M.gguf) | Q4_K_M | 0.3 | fast, recommended |
50
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q5_K_S.gguf) | Q5_K_S | 0.4 | |
51
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q5_K_M.gguf) | Q5_K_M | 0.4 | |
52
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q6_K.gguf) | Q6_K | 0.4 | very good quality |
53
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.Q8_0.gguf) | Q8_0 | 0.5 | fast, best quality |
54
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF/resolve/main/kani-tts-450m-0.1-ft.f16.gguf) | f16 | 0.8 | 16 bpw, overkill |
55
+
56
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
57
+ types (lower is better):
58
+
59
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
60
+
61
+ And here are Artefact2's thoughts on the matter:
62
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
63
+
64
+ ## FAQ / Model Request
65
+
66
+ See https://huggingface.co/mradermacher/model_requests for some answers to
67
+ questions you might have and/or if you want some other model quantized.
68
+
69
+ ## Thanks
70
+
71
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
72
+ me use its servers and providing upgrades to my workstation to enable
73
+ this work in my free time.
74
+
75
+ <!-- end -->
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a0d3b6919af765a50a30e2ab0b37e9b37d8156930ac49757db2a009641a1192
3
+ size 224888032
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e6adc6f14990cfada350a960216b0383c8cdab508855e5391d2af2e2c96d42b
3
+ size 173679840
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d644da2daf3ba8eb81813090e6638bf2321263878ff76e73dbd3d135bf7ac19a
3
+ size 216130784
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcb128dcf9bb5dcb20e739e12d7768e83a42e4ed98ff47c3ab9dc0e125498d72
3
+ size 206234848
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f33258444f115880a4da59e02843d846d63f39829d57be01c39898059baeb8e
3
+ size 194233568
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51e8c9b5049daa12ce9aa625bd8e426bb98b65d09bc5010f0974a8effcc510a3
3
+ size 242394336
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6cbfbfb3ce32f190e5f3e8baa3e7bd03e805732054e3c1b0a7b65e3e8799271
3
+ size 233833696
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf292c11d029e6aa14466bac8a78bbdec3b31b27e28b031f74ba360fe1485b71
3
+ size 273458400
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2534cedbe4f691967f5a6ca4cadf62dd122dd9a1b5715b040f95f2919545976
3
+ size 268305632
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fee5ebfb2e3b7786003e0f524262a2cf03035a51bf0ec1ace09587d8c52f3df9
3
+ size 306463968
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e4492eb368e415f57a6454539fe4acf1d57b2c6eafa5e05efe32b531673514d
3
+ size 396020480
en/kani-tts-450m-0.1-ft-GGUF/kani-tts-450m-0.1-ft.f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cf303e7b5c9aa2c05d10026cdc33e94a9beb497f7dd19e46c6c61b0d13f8558
3
+ size 742690880
en/kani-tts-450m-0.1-ft-GGUF/source.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ https://huggingface.co/mradermacher/kani-tts-450m-0.1-ft-GGUF
en/kani-tts-450m-0.1-ft/.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
en/kani-tts-450m-0.1-ft/README.md ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - nytopop/expresso-conversational
5
+ - nineninesix/expresso-conversational-en-nano-codec-dataset
6
+ language:
7
+ - en
8
+ pipeline_tag: text-to-speech
9
+ library_name: transformers
10
+ base_model:
11
+ - nineninesix/kani-tts-450m-0.1-pt
12
+ ---
13
+
14
+ <p>
15
+ <img src="https://www.nineninesix.ai/kitty.png" alt="Logo" width="200" height="200">
16
+ </p>
17
+
18
+ # KaniTTS Finetuned
19
+
20
+ Text-to-Speech (TTS) model designed for high-speed, high-fidelity audio generation.
21
+
22
+ KaniTTS is built on a novel architecture that combines a powerful language model with a highly efficient audio codec, enabling it to deliver exceptional performance for real-time applications.
23
+
24
+
25
+ ## Model Details
26
+
27
+ KaniTTS operates on a two-stage pipeline, leveraging a large foundation model for token generation and a compact, efficient codec for waveform synthesis.
28
+
29
+ The two-stage design of KaniTTS provides a significant advantage in terms of speed and efficiency. The backbone LLM generates a compressed token representation, which is then rapidly expanded into an audio waveform by the NanoCodec. This architecture bypasses the computational overhead associated with generating waveforms directly from large-scale language models, resulting in extremely low latency.
30
+
31
+
32
+ ## Features
33
+ This model trained primarily on English for robust core capabilities and the tokenizer supports these languages: English, Arabic, Chinese, French, German, Japanese, Korean, and Spanish.
34
+
35
+ The base model can be continually pretrained on the multilingual dataset producing high-fidelity audio at sample rates 22kHz.
36
+
37
+ This model powers voice interactions in the modern agentic systems, enabling seamless, human-like conversations.
38
+
39
+ - Model Size: 450M parameters (pretrained version)
40
+ - License: [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt)
41
+
42
+ ## Examples
43
+
44
+ | Text | Audio |
45
+ |---|---|
46
+ | I do believe Marsellus Wallace, MY husband, YOUR boss, told you to take me out and do WHATEVER I WANTED. | <audio controls><source src="https://github.com/nineninesix-ai/kani-tts/raw/refs/heads/main/public/mia.wav" type="audio/mpeg"></audio> |
47
+ | What do we say the the god of death? Not today! | <audio controls><source src="https://github.com/nineninesix-ai/kani-tts/raw/refs/heads/main/public/arya.wav" type="audio/wav"></audio> |
48
+ | What do you call a lawyer with an IQ of 60? Your honor | <audio controls><source src="https://github.com/nineninesix-ai/kani-tts/raw/refs/heads/main/public/saul.wav" type="audio/wav"></audio> |
49
+ | You mean, let me understand this cause, you know maybe it's me, it's a little fucked up maybe, but I'm funny how, I mean funny like I'm a clown, I amuse you? I make you laugh, I'm here to fucking amuse you? | <audio controls><source src="https://github.com/nineninesix-ai/kani-tts/raw/refs/heads/main/public/tommy.wav" type="audio/wav"></audio> |
50
+
51
+
52
+
53
+ ### Sources
54
+
55
+ - Website: [nineninesix.ai](https://www.nineninesix.ai/)
56
+ - GitHub Repo: [https://github.com/nineninesix-ai/kani-tts](https://github.com/nineninesix-ai/kani-tts)
57
+ - Base Model Card on HF: [nineninesix/kani-tts-450m-0.1-pt](https://huggingface.co/nineninesix/kani-tts-450m-0.1-pt)
58
+ - FT Model Card on HuggingFace: [nineninesix/kani-tts-450m-0.2-ft](https://huggingface.co/nineninesix/kani-tts-450m-0.2-ft)
59
+ - Link to HF Space: [nineninesix/KaniTTS](https://huggingface.co/spaces/nineninesix/KaniTTS)
60
+ - Inference Example: [Colab Notebook](https://colab.research.google.com/drive/1mvzGs7jtAMSUz8wvNlL5uFmgFEyAPjDh?usp=sharing)
61
+ - Finetuning Example: [Colab Notebook](https://colab.research.google.com/drive/1oDIPOSHW2kUoP3CGafvh9lM6j03Z-vE6?usp=sharing)
62
+ - Example Dataset for Fine-tuning: [Expresso Conversational](https://huggingface.co/datasets/nineninesix/expresso-conversational-en-nano-codec-dataset)
63
+ - [Waiting List](https://airtable.com/appX2G2TpoRk4M5Bf/pagO2xbIOjiwulPcP/form) for Pro Version
64
+
65
+ ## Recommended Uses
66
+ - Conversational AI: Integrate into chatbots, virtual assistants, or voice-enabled apps for real-time speech output.
67
+ - Edge and Server Deployment: Optimized for low-latency inference on edge devices or affordable servers, enabling scalable, resource-efficient voice applications.
68
+ - Accessibility Tools: Support screen readers or language learning apps with expressive prosody.
69
+ - Research: Fine-tune for domain-specific voices (e.g., accents, emotions) or benchmark against other TTS systems.
70
+
71
+ ## Limitations
72
+ - Performance may vary with finetuned variants, long inputs ( > 2000 tokens) or rare languages/accents.
73
+ - Emotion control is basic; advanced expressivity requires fine-tuning.
74
+ - Trained on public datasets; may inherit biases in prosody or pronunciation from training data.
75
+
76
+
77
+ ## Training Data
78
+
79
+ - Dataset: Curated from LibriTTS, Common Voice and Emilia (~50k hours).
80
+ - Pretrained mostly on English speech for robust core capabilities, with multilingual fine-tuning for supported languages.
81
+ - Metrics: MOS (Mean Opinion Score) 4.3/5 for naturalness; WER (Word Error Rate) < 5% on benchmark texts.
82
+ - Hardware: Pretrained on 8x H200 over 8 hours.
83
+
84
+ ## Inference on Nvidia RTX 5080:
85
+ - **Latency**: < 1s to generate 15 seconds of audio
86
+ - **Memory Usage**: 2GB GPU VRAM
87
+
88
+ > This performance makes KaniTTS suitable for real-time conversational AI applications and low-latency voice synthesis.
89
+
90
+
91
+ ## Tips & Tricks
92
+
93
+ - Language Optimization: For the best results in non-English languages, continually pretrain this model on datasets from your desired language set to improve prosody, accents, and pronunciation accuracy. Additionally, finetune NanoCodec for desired set of languages.
94
+ - Batch Processing: For high-throughput applications, process texts in batches of 8-16 to leverage parallel computation, reducing per-sample latency.
95
+ - **Blackwell GPU Optimization**: This model runs efficiently on NVIDIA's Blackwell architecture GPUs for faster inference and reduced latency in real-time applications.
96
+
97
+
98
+ ## Credits
99
+ - This project was inspired by the works of [Orpheus TTS](https://huggingface.co/canopylabs/orpheus-3b-0.1-pretrained) and [Sesame CSM](https://huggingface.co/sesame/csm-1b).
100
+ - It utilizes the [LiquidAI LFM2 350M](https://huggingface.co/LiquidAI/LFM2-350M) as its core backbone and
101
+ - [Nvidia NanoCodec](https://huggingface.co/nvidia/nemo-nano-codec-22khz-0.6kbps-12.5fps) for efficient audio processing.
102
+
103
+ ## Responsible Use and Prohibited Activities
104
+ The model is designed for ethical and responsible use. The following activities are strictly prohibited:
105
+ - The model may not be used for any illegal purposes or to create content that is harmful, threatening, defamatory, or obscene. This includes, but is not limited to, the generation of hate speech, harassment, or incitement of violence.
106
+ - You may not use the model to generate or disseminate false or misleading information. This includes creating deceptive audio content that impersonates individuals without their consent or misrepresents facts.
107
+ - The model is not to be used for any malicious activities, such as spamming, phishing, or the creation of content intended to deceive or defraud.
108
+
109
+ By using this model, you agree to abide by these restrictions and all applicable laws and regulations.
110
+
111
+
112
+ ## Contact
113
+ Have a question, feedback, or need support? Please fill out our [contact form](https://airtable.com/appX2G2TpoRk4M5Bf/pagO2xbIOjiwulPcP/form) and we'll get back to you as soon as possible.
en/kani-tts-450m-0.1-ft/chat_template.jinja ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{- bos_token -}}
2
+ {%- set system_prompt = "" -%}
3
+ {%- set ns = namespace(system_prompt="") -%}
4
+ {%- if messages[0]["role"] == "system" -%}
5
+ {%- set ns.system_prompt = messages[0]["content"] -%}
6
+ {%- set messages = messages[1:] -%}
7
+ {%- endif -%}
8
+ {%- if tools -%}
9
+ {%- set ns.system_prompt = ns.system_prompt + ("\n" if ns.system_prompt else "") + "List of tools: <|tool_list_start|>[" -%}
10
+ {%- for tool in tools -%}
11
+ {%- if tool is not string -%}
12
+ {%- set tool = tool | tojson -%}
13
+ {%- endif -%}
14
+ {%- set ns.system_prompt = ns.system_prompt + tool -%}
15
+ {%- if not loop.last -%}
16
+ {%- set ns.system_prompt = ns.system_prompt + ", " -%}
17
+ {%- endif -%}
18
+ {%- endfor -%}
19
+ {%- set ns.system_prompt = ns.system_prompt + "]<|tool_list_end|>" -%}
20
+ {%- endif -%}
21
+ {%- if ns.system_prompt -%}
22
+ {{- "<|im_start|>system\n" + ns.system_prompt + "<|im_end|>\n" -}}
23
+ {%- endif -%}
24
+ {%- for message in messages -%}
25
+ {{- "<|im_start|>" + message["role"] + "\n" -}}
26
+ {%- set content = message["content"] -%}
27
+ {%- if content is not string -%}
28
+ {%- set content = content | tojson -%}
29
+ {%- endif -%}
30
+ {%- if message["role"] == "tool" -%}
31
+ {%- set content = "<|tool_response_start|>" + content + "<|tool_response_end|>" -%}
32
+ {%- endif -%}
33
+ {{- content + "<|im_end|>\n" -}}
34
+ {%- endfor -%}
35
+ {%- if add_generation_prompt -%}
36
+ {{- "<|im_start|>assistant\n" -}}
37
+ {%- endif -%}
en/kani-tts-450m-0.1-ft/config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Lfm2ForCausalLM"
4
+ ],
5
+ "block_auto_adjust_ff_dim": true,
6
+ "block_dim": 1024,
7
+ "block_ff_dim": 6656,
8
+ "block_ffn_dim_multiplier": 1.0,
9
+ "block_mlp_init_scale": 1.0,
10
+ "block_multiple_of": 256,
11
+ "block_norm_eps": 1e-05,
12
+ "block_out_init_scale": 1.0,
13
+ "block_use_swiglu": true,
14
+ "block_use_xavier_init": true,
15
+ "bos_token_id": 1,
16
+ "conv_L_cache": 3,
17
+ "conv_bias": false,
18
+ "conv_dim": 1024,
19
+ "conv_dim_out": 1024,
20
+ "conv_use_xavier_init": true,
21
+ "dtype": "float32",
22
+ "eos_token_id": 7,
23
+ "hidden_size": 1024,
24
+ "initializer_range": 0.02,
25
+ "intermediate_size": 6656,
26
+ "layer_types": [
27
+ "conv",
28
+ "conv",
29
+ "full_attention",
30
+ "conv",
31
+ "conv",
32
+ "full_attention",
33
+ "conv",
34
+ "conv",
35
+ "full_attention",
36
+ "conv",
37
+ "full_attention",
38
+ "conv",
39
+ "full_attention",
40
+ "conv",
41
+ "full_attention",
42
+ "conv"
43
+ ],
44
+ "max_position_embeddings": 128000,
45
+ "model_type": "lfm2",
46
+ "norm_eps": 1e-05,
47
+ "num_attention_heads": 16,
48
+ "num_heads": 16,
49
+ "num_hidden_layers": 16,
50
+ "num_key_value_heads": 8,
51
+ "pad_token_id": 0,
52
+ "rope_theta": 1000000.0,
53
+ "transformers_version": "4.56.0",
54
+ "use_cache": true,
55
+ "use_pos_enc": true,
56
+ "vocab_size": 80539
57
+ }
en/kani-tts-450m-0.1-ft/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 7,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.56.0"
7
+ }
en/kani-tts-450m-0.1-ft/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b236870bd7f813138bf2dbe237c7e8bcc35df3f687e4eba9deb11e6be1c66731
3
+ size 1479404656
en/kani-tts-450m-0.1-ft/source.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ https://huggingface.co/nineninesix/kani-tts-450m-0.1-ft
en/kani-tts-450m-0.1-ft/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|im_end|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|pad|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
en/kani-tts-450m-0.1-ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
en/kani-tts-450m-0.1-ft/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
en/kani-tts-450m-0.1-pt-GGUF/.gitattributes ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ kani-tts-450m-0.1-pt.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
37
+ kani-tts-450m-0.1-pt.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
38
+ kani-tts-450m-0.1-pt.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
39
+ kani-tts-450m-0.1-pt.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ kani-tts-450m-0.1-pt.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
41
+ kani-tts-450m-0.1-pt.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ kani-tts-450m-0.1-pt.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ kani-tts-450m-0.1-pt.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ kani-tts-450m-0.1-pt.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
45
+ kani-tts-450m-0.1-pt.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
46
+ kani-tts-450m-0.1-pt.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
47
+ kani-tts-450m-0.1-pt.f16.gguf filter=lfs diff=lfs merge=lfs -text
en/kani-tts-450m-0.1-pt-GGUF/README.md ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: nineninesix/kani-tts-450m-0.1-pt
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ license: apache-2.0
7
+ mradermacher:
8
+ readme_rev: 1
9
+ quantized_by: mradermacher
10
+ ---
11
+ ## About
12
+
13
+ <!-- ### quantize_version: 2 -->
14
+ <!-- ### output_tensor_quantised: 1 -->
15
+ <!-- ### convert_type: hf -->
16
+ <!-- ### vocab_type: -->
17
+ <!-- ### tags: -->
18
+ <!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
19
+ <!-- ### quants_skip: -->
20
+ <!-- ### skip_mmproj: -->
21
+ static quants of https://huggingface.co/nineninesix/kani-tts-450m-0.1-pt
22
+
23
+ <!-- provided-files -->
24
+
25
+ ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#kani-tts-450m-0.1-pt-GGUF).***
26
+
27
+ weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
28
+ ## Usage
29
+
30
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
31
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
32
+ more details, including on how to concatenate multi-part files.
33
+
34
+ ## Provided Quants
35
+
36
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
37
+
38
+ | Link | Type | Size/GB | Notes |
39
+ |:-----|:-----|--------:|:------|
40
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q2_K.gguf) | Q2_K | 0.3 | |
41
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q3_K_S.gguf) | Q3_K_S | 0.3 | |
42
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q3_K_M.gguf) | Q3_K_M | 0.3 | lower quality |
43
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q3_K_L.gguf) | Q3_K_L | 0.4 | |
44
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.IQ4_XS.gguf) | IQ4_XS | 0.4 | |
45
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q4_K_S.gguf) | Q4_K_S | 0.4 | fast, recommended |
46
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q4_K_M.gguf) | Q4_K_M | 0.4 | fast, recommended |
47
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q5_K_S.gguf) | Q5_K_S | 0.4 | |
48
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q5_K_M.gguf) | Q5_K_M | 0.4 | |
49
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q6_K.gguf) | Q6_K | 0.5 | very good quality |
50
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.Q8_0.gguf) | Q8_0 | 0.6 | fast, best quality |
51
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF/resolve/main/kani-tts-450m-0.1-pt.f16.gguf) | f16 | 1.0 | 16 bpw, overkill |
52
+
53
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
54
+ types (lower is better):
55
+
56
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
57
+
58
+ And here are Artefact2's thoughts on the matter:
59
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
60
+
61
+ ## FAQ / Model Request
62
+
63
+ See https://huggingface.co/mradermacher/model_requests for some answers to
64
+ questions you might have and/or if you want some other model quantized.
65
+
66
+ ## Thanks
67
+
68
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
69
+ me use its servers and providing upgrades to my workstation to enable
70
+ this work in my free time.
71
+
72
+ <!-- end -->
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca7b30ff82c2d7eaedaf0957a575efde7d894bac74b82f27819fe7b929153ed7
3
+ size 268700736
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5123d6383e4751732da3a77e7a7ef158b2070b851500cc8fe71a44f36bb9567b
3
+ size 200740448
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c65f02c14e3a3b06a8d325a4735671c93654ec06c8b59d774e0c307154aacce9
3
+ size 251567456
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:721b3fc05e33e7c1164f4a1572a23c07e102bf8b49e350ce2d6e7cf943af57b9
3
+ size 241671520
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d92dc71d7d4f613fcbb3be4b079d15e4a22e7380eaeeac103eeb1cf4a982dec
3
+ size 229670240
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:330ad3a8c3a65fd37b84d216391344935d3d5fd5d7c1e147e6a53ebe855a9d69
3
+ size 288784288
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42e4c9b589921c1313e94e65c901488c18b6933cb54321c2a2c9d0973397a8c7
3
+ size 280223648
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dde33d52ede5549d45a2d363ea10e31643627b321a4ee61af012731ff5d5cd17
3
+ size 330157344
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d12740bc6504fb7c1c9ef9662129dc28328ebbb14e774e00c417186d79a356d
3
+ size 325004576
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae58cf1978dfee3b1bfdec13235a4f83cb2a3a0ea75cb63d2c5c11cce5dd0ccc
3
+ size 374116224
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efbf33cdbd12bdd6e8752a307975dc17bfb0fa3660c61ac16261e561dd629635
3
+ size 483646400
en/kani-tts-450m-0.1-pt-GGUF/kani-tts-450m-0.1-pt.f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f512339178b4661b2dfb3e4e34d6759bf894e48115ad988a18035241629f2c33
3
+ size 907634240
en/kani-tts-450m-0.1-pt-GGUF/source.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ https://huggingface.co/mradermacher/kani-tts-450m-0.1-pt-GGUF
en/kani-tts-450m-0.2-ft-GGUF/.gitattributes ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ kani-tts-450m-0.2-ft.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
37
+ kani-tts-450m-0.2-ft.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
38
+ kani-tts-450m-0.2-ft.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
39
+ kani-tts-450m-0.2-ft.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ kani-tts-450m-0.2-ft.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
41
+ kani-tts-450m-0.2-ft.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ kani-tts-450m-0.2-ft.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ kani-tts-450m-0.2-ft.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ kani-tts-450m-0.2-ft.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
45
+ kani-tts-450m-0.2-ft.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
46
+ kani-tts-450m-0.2-ft.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
47
+ kani-tts-450m-0.2-ft.f16.gguf filter=lfs diff=lfs merge=lfs -text
en/kani-tts-450m-0.2-ft-GGUF/README.md ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: nineninesix/kani-tts-450m-0.2-ft
3
+ datasets:
4
+ - nytopop/expresso-conversational
5
+ - nineninesix/expresso-conversational-en-nano-codec-dataset
6
+ language:
7
+ - en
8
+ library_name: transformers
9
+ license: apache-2.0
10
+ mradermacher:
11
+ readme_rev: 1
12
+ quantized_by: mradermacher
13
+ ---
14
+ ## About
15
+
16
+ <!-- ### quantize_version: 2 -->
17
+ <!-- ### output_tensor_quantised: 1 -->
18
+ <!-- ### convert_type: hf -->
19
+ <!-- ### vocab_type: -->
20
+ <!-- ### tags: -->
21
+ <!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
22
+ <!-- ### quants_skip: -->
23
+ <!-- ### skip_mmproj: -->
24
+ static quants of https://huggingface.co/nineninesix/kani-tts-450m-0.2-ft
25
+
26
+ <!-- provided-files -->
27
+
28
+ ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#kani-tts-450m-0.2-ft-GGUF).***
29
+
30
+ weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
31
+ ## Usage
32
+
33
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
34
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
35
+ more details, including on how to concatenate multi-part files.
36
+
37
+ ## Provided Quants
38
+
39
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
40
+
41
+ | Link | Type | Size/GB | Notes |
42
+ |:-----|:-----|--------:|:------|
43
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q2_K.gguf) | Q2_K | 0.3 | |
44
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q3_K_S.gguf) | Q3_K_S | 0.3 | |
45
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q3_K_M.gguf) | Q3_K_M | 0.3 | lower quality |
46
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q3_K_L.gguf) | Q3_K_L | 0.3 | |
47
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.IQ4_XS.gguf) | IQ4_XS | 0.3 | |
48
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q4_K_S.gguf) | Q4_K_S | 0.3 | fast, recommended |
49
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q4_K_M.gguf) | Q4_K_M | 0.3 | fast, recommended |
50
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q5_K_S.gguf) | Q5_K_S | 0.4 | |
51
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q5_K_M.gguf) | Q5_K_M | 0.4 | |
52
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q6_K.gguf) | Q6_K | 0.4 | very good quality |
53
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.Q8_0.gguf) | Q8_0 | 0.5 | fast, best quality |
54
+ | [GGUF](https://huggingface.co/mradermacher/kani-tts-450m-0.2-ft-GGUF/resolve/main/kani-tts-450m-0.2-ft.f16.gguf) | f16 | 0.8 | 16 bpw, overkill |
55
+
56
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
57
+ types (lower is better):
58
+
59
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
60
+
61
+ And here are Artefact2's thoughts on the matter:
62
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
63
+
64
+ ## FAQ / Model Request
65
+
66
+ See https://huggingface.co/mradermacher/model_requests for some answers to
67
+ questions you might have and/or if you want some other model quantized.
68
+
69
+ ## Thanks
70
+
71
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
72
+ me use its servers and providing upgrades to my workstation to enable
73
+ this work in my free time.
74
+
75
+ <!-- end -->
en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02bf423f2e73823fd946a300e5ad4c672f59d2d5662d4d4204f13f65c33ce4a8
3
+ size 224888032
en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d7f6b1fe23ed4464324ad3f71db86bebc1686e82a1eabbd57ef75b0c03bd6f8
3
+ size 173679840
en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d1fb435da0f273e1f0066275c97b53bdf1199b3f64ecce99a739bfe1c92535f
3
+ size 216130784
en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cdc741e27f22f91f49c90aa9825e3ec49561d1036aa0a1ec04ae86c988cbc35
3
+ size 206234848
en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a52ebc7fdb357344c86e23303a2adb62059d937c70ade050d02ac2010b92884
3
+ size 194233568
en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5561015d0e1aa8dac02f037251d13cd70dc536d5499570fe8876d5a96e780c5d
3
+ size 242394336
en/kani-tts-450m-0.2-ft-GGUF/kani-tts-450m-0.2-ft.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:402ce7ddcbf2935bc8d82aafc2f8714950c5fe72e665f6d2ee6e537ddee91b1f
3
+ size 233833696