EduardoPacheco commited on
Commit
5b86a64
·
verified ·
1 Parent(s): f1ec7a5

Add Qwen3-TTS Custom Voice Models

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/analytics/coremldata.bin +3 -0
  2. qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/coremldata.bin +3 -0
  3. qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/metadata.json +159 -0
  4. qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/model.mil +0 -0
  5. qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/weights/weight.bin +3 -0
  6. qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/analytics/coremldata.bin +3 -0
  7. qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/coremldata.bin +3 -0
  8. qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/metadata.json +159 -0
  9. qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/model.mil +0 -0
  10. qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/weights/weight.bin +3 -0
  11. qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/analytics/coremldata.bin +3 -0
  12. qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/coremldata.bin +3 -0
  13. qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/metadata.json +66 -0
  14. qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/model.mil +26 -0
  15. qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/weights/weight.bin +3 -0
  16. qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/analytics/coremldata.bin +3 -0
  17. qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/coremldata.bin +3 -0
  18. qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/metadata.json +66 -0
  19. qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/model.mil +26 -0
  20. qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/weights/weight.bin +3 -0
  21. qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/analytics/coremldata.bin +3 -0
  22. qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/coremldata.bin +3 -0
  23. qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/metadata.json +151 -0
  24. qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/model.mil +0 -0
  25. qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/weights/weight.bin +3 -0
  26. qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/analytics/coremldata.bin +3 -0
  27. qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/coremldata.bin +3 -0
  28. qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/metadata.json +151 -0
  29. qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/model.mil +0 -0
  30. qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/weights/weight.bin +3 -0
  31. qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/analytics/coremldata.bin +3 -0
  32. qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/coremldata.bin +3 -0
  33. qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/metadata.json +66 -0
  34. qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/model.mil +26 -0
  35. qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/weights/weight.bin +3 -0
  36. qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/analytics/coremldata.bin +3 -0
  37. qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/coremldata.bin +3 -0
  38. qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/metadata.json +66 -0
  39. qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/model.mil +26 -0
  40. qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/weights/weight.bin +3 -0
  41. qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/analytics/coremldata.bin +3 -0
  42. qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/coremldata.bin +3 -0
  43. qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/metadata.json +178 -0
  44. qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/model.mil +0 -0
  45. qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/weights/weight.bin +3 -0
  46. qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/analytics/coremldata.bin +3 -0
  47. qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/coremldata.bin +3 -0
  48. qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/metadata.json +177 -0
  49. qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/model.mil +0 -0
  50. qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/weights/weight.bin +3 -0
qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbc1ea6ac3fa1e7a07a9e19a4407badc3cdf153bc2382b6a6d408a880f3964c3
3
+ size 243
qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c9e439b240cd4bf60f1a142b4c3d34ea5ae64f57c9e2fd407606cbdae76a4d2
3
+ size 671
qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Mixed (Float16, Int32, Palettized (8 bits), UInt8)",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 1 × 3072)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 3072]",
13
+ "name" : "logits",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 1)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 1024, 1, 1]",
23
+ "name" : "hidden_states",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 28672 × 1 × 1)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 28672, 1, 1]",
33
+ "name" : "key_cache_updates",
34
+ "type" : "MultiArray"
35
+ },
36
+ {
37
+ "hasShapeFlexibility" : "0",
38
+ "isOptional" : "0",
39
+ "dataType" : "Float16",
40
+ "formattedType" : "MultiArray (Float16 1 × 28672 × 1 × 1)",
41
+ "shortDescription" : "",
42
+ "shape" : "[1, 28672, 1, 1]",
43
+ "name" : "value_cache_updates",
44
+ "type" : "MultiArray"
45
+ }
46
+ ],
47
+ "modelParameters" : [
48
+
49
+ ],
50
+ "specificationVersion" : 9,
51
+ "mlProgramOperationTypeHistogram" : {
52
+ "Ios18.expandDims" : 8,
53
+ "Ios18.softmax" : 28,
54
+ "Ios18.mul" : 675,
55
+ "Ios18.matmul" : 56,
56
+ "Ios18.rsqrt" : 113,
57
+ "Ios16.reduceMean" : 113,
58
+ "Split" : 2,
59
+ "Ios18.greaterEqual" : 2,
60
+ "Select" : 2,
61
+ "Ios18.readState" : 2,
62
+ "Ios18.gather" : 2,
63
+ "Ios18.add" : 311,
64
+ "Ios18.reshape" : 224,
65
+ "Ios18.constexprLutToDense" : 199,
66
+ "Ios18.conv" : 197,
67
+ "Ios18.concat" : 114,
68
+ "Ios18.cast" : 5,
69
+ "Ios18.sub" : 1,
70
+ "Ios18.silu" : 28,
71
+ "Ios18.transpose" : 1,
72
+ "Ios18.sliceByIndex" : 560,
73
+ "Ios18.squeeze" : 1
74
+ },
75
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
76
+ "isUpdatable" : "0",
77
+ "stateSchema" : [
78
+ {
79
+ "dataType" : "Float16",
80
+ "isOptional" : "0",
81
+ "formattedType" : "State (Float16 1 × 28672 × 1 × 256)",
82
+ "shortDescription" : "",
83
+ "shape" : "[1, 28672, 1, 256]",
84
+ "name" : "self_attn_key_cache",
85
+ "type" : "State"
86
+ },
87
+ {
88
+ "dataType" : "Float16",
89
+ "isOptional" : "0",
90
+ "formattedType" : "State (Float16 1 × 28672 × 1 × 256)",
91
+ "shortDescription" : "",
92
+ "shape" : "[1, 28672, 1, 256]",
93
+ "name" : "self_attn_value_cache",
94
+ "type" : "State"
95
+ }
96
+ ],
97
+ "availability" : {
98
+ "macOS" : "15.0",
99
+ "tvOS" : "18.0",
100
+ "visionOS" : "2.0",
101
+ "watchOS" : "11.0",
102
+ "iOS" : "18.0",
103
+ "macCatalyst" : "18.0"
104
+ },
105
+ "modelType" : {
106
+ "name" : "MLModelType_mlProgram"
107
+ },
108
+ "userDefinedMetadata" : {
109
+ "com.github.apple.coremltools.conversion_date" : "2026-02-17",
110
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
111
+ "com.github.apple.coremltools.version" : "9.0",
112
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
113
+ },
114
+ "inputSchema" : [
115
+ {
116
+ "hasShapeFlexibility" : "0",
117
+ "isOptional" : "0",
118
+ "dataType" : "Float16",
119
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 1)",
120
+ "shortDescription" : "",
121
+ "shape" : "[1, 1024, 1, 1]",
122
+ "name" : "input_embeds",
123
+ "type" : "MultiArray"
124
+ },
125
+ {
126
+ "hasShapeFlexibility" : "0",
127
+ "isOptional" : "0",
128
+ "dataType" : "Int32",
129
+ "formattedType" : "MultiArray (Int32 1)",
130
+ "shortDescription" : "",
131
+ "shape" : "[1]",
132
+ "name" : "cache_length",
133
+ "type" : "MultiArray"
134
+ },
135
+ {
136
+ "hasShapeFlexibility" : "0",
137
+ "isOptional" : "0",
138
+ "dataType" : "Float16",
139
+ "formattedType" : "MultiArray (Float16 1 × 256)",
140
+ "shortDescription" : "",
141
+ "shape" : "[1, 256]",
142
+ "name" : "kv_cache_update_mask",
143
+ "type" : "MultiArray"
144
+ },
145
+ {
146
+ "hasShapeFlexibility" : "0",
147
+ "isOptional" : "0",
148
+ "dataType" : "Float16",
149
+ "formattedType" : "MultiArray (Float16 1 × 256)",
150
+ "shortDescription" : "",
151
+ "shape" : "[1, 256]",
152
+ "name" : "key_padding_mask",
153
+ "type" : "MultiArray"
154
+ }
155
+ ],
156
+ "generatedClassName" : "CodeDecoderWithStatefulSelfAttention_8_bit",
157
+ "method" : "predict"
158
+ }
159
+ ]
qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
qwen3_tts/code_decoder/12hz-0.6b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9618fdb2a96db80646b257659ec4cf01240283c211ea7fc3ad0d986d35304fa
3
+ size 443885568
qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ae0466e386343e0079070977b713c72abd29e9653af179fbac1a0c656a7408b
3
+ size 243
qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bac65aa923aecda6c84e1439769c47c440ab8f8afcdfd47b8d100b992dc7be1a
3
+ size 671
qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Mixed (Float16, Int32, Palettized (8 bits), UInt8)",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 1 × 3072)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 3072]",
13
+ "name" : "logits",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 2048 × 1 × 1)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 2048, 1, 1]",
23
+ "name" : "hidden_states",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 28672 × 1 × 1)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 28672, 1, 1]",
33
+ "name" : "key_cache_updates",
34
+ "type" : "MultiArray"
35
+ },
36
+ {
37
+ "hasShapeFlexibility" : "0",
38
+ "isOptional" : "0",
39
+ "dataType" : "Float16",
40
+ "formattedType" : "MultiArray (Float16 1 × 28672 × 1 × 1)",
41
+ "shortDescription" : "",
42
+ "shape" : "[1, 28672, 1, 1]",
43
+ "name" : "value_cache_updates",
44
+ "type" : "MultiArray"
45
+ }
46
+ ],
47
+ "modelParameters" : [
48
+
49
+ ],
50
+ "specificationVersion" : 9,
51
+ "mlProgramOperationTypeHistogram" : {
52
+ "Ios18.expandDims" : 8,
53
+ "Ios18.softmax" : 28,
54
+ "Ios18.mul" : 675,
55
+ "Ios18.matmul" : 56,
56
+ "Ios18.rsqrt" : 113,
57
+ "Ios16.reduceMean" : 113,
58
+ "Split" : 2,
59
+ "Ios18.greaterEqual" : 2,
60
+ "Select" : 2,
61
+ "Ios18.readState" : 2,
62
+ "Ios18.gather" : 2,
63
+ "Ios18.add" : 311,
64
+ "Ios18.reshape" : 224,
65
+ "Ios18.constexprLutToDense" : 199,
66
+ "Ios18.conv" : 197,
67
+ "Ios18.concat" : 114,
68
+ "Ios18.cast" : 5,
69
+ "Ios18.sub" : 1,
70
+ "Ios18.silu" : 28,
71
+ "Ios18.transpose" : 1,
72
+ "Ios18.sliceByIndex" : 560,
73
+ "Ios18.squeeze" : 1
74
+ },
75
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
76
+ "isUpdatable" : "0",
77
+ "stateSchema" : [
78
+ {
79
+ "dataType" : "Float16",
80
+ "isOptional" : "0",
81
+ "formattedType" : "State (Float16 1 × 28672 × 1 × 256)",
82
+ "shortDescription" : "",
83
+ "shape" : "[1, 28672, 1, 256]",
84
+ "name" : "self_attn_key_cache",
85
+ "type" : "State"
86
+ },
87
+ {
88
+ "dataType" : "Float16",
89
+ "isOptional" : "0",
90
+ "formattedType" : "State (Float16 1 × 28672 × 1 × 256)",
91
+ "shortDescription" : "",
92
+ "shape" : "[1, 28672, 1, 256]",
93
+ "name" : "self_attn_value_cache",
94
+ "type" : "State"
95
+ }
96
+ ],
97
+ "availability" : {
98
+ "macOS" : "15.0",
99
+ "tvOS" : "18.0",
100
+ "visionOS" : "2.0",
101
+ "watchOS" : "11.0",
102
+ "iOS" : "18.0",
103
+ "macCatalyst" : "18.0"
104
+ },
105
+ "modelType" : {
106
+ "name" : "MLModelType_mlProgram"
107
+ },
108
+ "userDefinedMetadata" : {
109
+ "com.github.apple.coremltools.conversion_date" : "2026-02-17",
110
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
111
+ "com.github.apple.coremltools.version" : "9.0",
112
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
113
+ },
114
+ "inputSchema" : [
115
+ {
116
+ "hasShapeFlexibility" : "0",
117
+ "isOptional" : "0",
118
+ "dataType" : "Float16",
119
+ "formattedType" : "MultiArray (Float16 1 × 2048 × 1 × 1)",
120
+ "shortDescription" : "",
121
+ "shape" : "[1, 2048, 1, 1]",
122
+ "name" : "input_embeds",
123
+ "type" : "MultiArray"
124
+ },
125
+ {
126
+ "hasShapeFlexibility" : "0",
127
+ "isOptional" : "0",
128
+ "dataType" : "Int32",
129
+ "formattedType" : "MultiArray (Int32 1)",
130
+ "shortDescription" : "",
131
+ "shape" : "[1]",
132
+ "name" : "cache_length",
133
+ "type" : "MultiArray"
134
+ },
135
+ {
136
+ "hasShapeFlexibility" : "0",
137
+ "isOptional" : "0",
138
+ "dataType" : "Float16",
139
+ "formattedType" : "MultiArray (Float16 1 × 256)",
140
+ "shortDescription" : "",
141
+ "shape" : "[1, 256]",
142
+ "name" : "kv_cache_update_mask",
143
+ "type" : "MultiArray"
144
+ },
145
+ {
146
+ "hasShapeFlexibility" : "0",
147
+ "isOptional" : "0",
148
+ "dataType" : "Float16",
149
+ "formattedType" : "MultiArray (Float16 1 × 256)",
150
+ "shortDescription" : "",
151
+ "shape" : "[1, 256]",
152
+ "name" : "key_padding_mask",
153
+ "type" : "MultiArray"
154
+ }
155
+ ],
156
+ "generatedClassName" : "CodeDecoderWithStatefulSelfAttention_8_bit",
157
+ "method" : "predict"
158
+ }
159
+ ]
qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
qwen3_tts/code_decoder/12hz-1.7b-customvoice/W8A16-stateful/CodeDecoder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc0cb784a9df3f11cb4c12710e962ab108d84081f25d7e0f068a0eb0aa3ace7
3
+ size 1416032256
qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5880de6256184e5a3bdbbdd631e8fccf030f2cf82d8930c36bc9542cd5d31645
3
+ size 243
qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d19846f4edbd27cef1e45f4c155aad182318620c53e94897a21622774c604b8
3
+ size 380
qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Float16",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 1)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1024, 1, 1]",
13
+ "name" : "input_embeds",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "modelParameters" : [
18
+
19
+ ],
20
+ "specificationVersion" : 9,
21
+ "mlProgramOperationTypeHistogram" : {
22
+ "Ios18.greaterEqual" : 1,
23
+ "Ios18.add" : 1,
24
+ "Ios18.cast" : 3,
25
+ "Select" : 1,
26
+ "Ios18.gather" : 1,
27
+ "Ios18.expandDims" : 2
28
+ },
29
+ "computePrecision" : "Mixed (Float16, Int16, Int32)",
30
+ "isUpdatable" : "0",
31
+ "stateSchema" : [
32
+
33
+ ],
34
+ "availability" : {
35
+ "macOS" : "15.0",
36
+ "tvOS" : "18.0",
37
+ "visionOS" : "2.0",
38
+ "watchOS" : "11.0",
39
+ "iOS" : "18.0",
40
+ "macCatalyst" : "18.0"
41
+ },
42
+ "modelType" : {
43
+ "name" : "MLModelType_mlProgram"
44
+ },
45
+ "userDefinedMetadata" : {
46
+ "com.github.apple.coremltools.conversion_date" : "2026-02-06",
47
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
48
+ "com.github.apple.coremltools.version" : "9.0",
49
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
50
+ },
51
+ "inputSchema" : [
52
+ {
53
+ "hasShapeFlexibility" : "0",
54
+ "isOptional" : "0",
55
+ "dataType" : "Int32",
56
+ "formattedType" : "MultiArray (Int32 1)",
57
+ "shortDescription" : "",
58
+ "shape" : "[1]",
59
+ "name" : "input_ids",
60
+ "type" : "MultiArray"
61
+ }
62
+ ],
63
+ "generatedClassName" : "CodeEmbedder",
64
+ "method" : "predict"
65
+ }
66
+ ]
qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/model.mil ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})]
3
+ {
4
+ func main<ios18>(tensor<int32, [1]> input_ids) {
5
+ int32 embeddings_batch_dims_0 = const()[name = string("embeddings_batch_dims_0"), val = int32(0)];
6
+ bool embeddings_validate_indices_0 = const()[name = string("embeddings_validate_indices_0"), val = bool(false)];
7
+ tensor<fp16, [3072, 1024]> codec_embedding_weight_to_fp16 = const()[name = string("codec_embedding_weight_to_fp16"), val = tensor<fp16, [3072, 1024]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
8
+ string input_ids_to_int16_dtype_0 = const()[name = string("input_ids_to_int16_dtype_0"), val = string("int16")];
9
+ string cast_2_dtype_0 = const()[name = string("cast_2_dtype_0"), val = string("int32")];
10
+ int32 greater_equal_0_y_0 = const()[name = string("greater_equal_0_y_0"), val = int32(0)];
11
+ tensor<int16, [1]> input_ids_to_int16 = cast(dtype = input_ids_to_int16_dtype_0, x = input_ids)[name = string("cast_5")];
12
+ tensor<int32, [1]> cast_2 = cast(dtype = cast_2_dtype_0, x = input_ids_to_int16)[name = string("cast_4")];
13
+ tensor<bool, [1]> greater_equal_0 = greater_equal(x = cast_2, y = greater_equal_0_y_0)[name = string("greater_equal_0")];
14
+ int32 slice_by_index_0 = const()[name = string("slice_by_index_0"), val = int32(3072)];
15
+ tensor<int32, [1]> add_0 = add(x = cast_2, y = slice_by_index_0)[name = string("add_0")];
16
+ tensor<int32, [1]> select_0 = select(a = cast_2, b = add_0, cond = greater_equal_0)[name = string("select_0")];
17
+ int32 embeddings_cast_fp16_cast_uint16_axis_0 = const()[name = string("embeddings_cast_fp16_cast_uint16_axis_0"), val = int32(0)];
18
+ string select_0_to_int16_dtype_0 = const()[name = string("select_0_to_int16_dtype_0"), val = string("int16")];
19
+ tensor<int16, [1]> select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = string("cast_3")];
20
+ tensor<fp16, [1, 1024]> embeddings_cast_fp16_cast_uint16_cast_uint16 = gather(axis = embeddings_cast_fp16_cast_uint16_axis_0, batch_dims = embeddings_batch_dims_0, indices = select_0_to_int16, validate_indices = embeddings_validate_indices_0, x = codec_embedding_weight_to_fp16)[name = string("embeddings_cast_fp16_cast_uint16_cast_uint16")];
21
+ tensor<int32, [1]> var_18_axes_0 = const()[name = string("op_18_axes_0"), val = tensor<int32, [1]>([2])];
22
+ tensor<fp16, [1, 1024, 1]> var_18_cast_fp16 = expand_dims(axes = var_18_axes_0, x = embeddings_cast_fp16_cast_uint16_cast_uint16)[name = string("op_18_cast_fp16")];
23
+ tensor<int32, [1]> var_20_axes_0 = const()[name = string("op_20_axes_0"), val = tensor<int32, [1]>([3])];
24
+ tensor<fp16, [1, 1024, 1, 1]> input_embeds = expand_dims(axes = var_20_axes_0, x = var_18_cast_fp16)[name = string("op_20_cast_fp16")];
25
+ } -> (input_embeds);
26
+ }
qwen3_tts/code_embedder/12hz-0.6b-customvoice/W16A16/CodeEmbedder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bf4059e34abbb08bea47b47e1cc305a08c690c19e42526aa97c14e9dd66b4b4
3
+ size 6291584
qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a67276394d403612b21f2042840c89e8b5feeb0d76162b0b10dc9f5be3273242
3
+ size 243
qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52bea537a5853fa5f98b869ca6c68e052febb5257d5b5a5a14be9148e4599e6b
3
+ size 380
qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Float16",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 2048 × 1 × 1)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 2048, 1, 1]",
13
+ "name" : "input_embeds",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "modelParameters" : [
18
+
19
+ ],
20
+ "specificationVersion" : 9,
21
+ "mlProgramOperationTypeHistogram" : {
22
+ "Ios18.greaterEqual" : 1,
23
+ "Ios18.add" : 1,
24
+ "Ios18.cast" : 3,
25
+ "Select" : 1,
26
+ "Ios18.gather" : 1,
27
+ "Ios18.expandDims" : 2
28
+ },
29
+ "computePrecision" : "Mixed (Float16, Int16, Int32)",
30
+ "isUpdatable" : "0",
31
+ "stateSchema" : [
32
+
33
+ ],
34
+ "availability" : {
35
+ "macOS" : "15.0",
36
+ "tvOS" : "18.0",
37
+ "visionOS" : "2.0",
38
+ "watchOS" : "11.0",
39
+ "iOS" : "18.0",
40
+ "macCatalyst" : "18.0"
41
+ },
42
+ "modelType" : {
43
+ "name" : "MLModelType_mlProgram"
44
+ },
45
+ "userDefinedMetadata" : {
46
+ "com.github.apple.coremltools.conversion_date" : "2026-02-12",
47
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
48
+ "com.github.apple.coremltools.version" : "9.0",
49
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
50
+ },
51
+ "inputSchema" : [
52
+ {
53
+ "hasShapeFlexibility" : "0",
54
+ "isOptional" : "0",
55
+ "dataType" : "Int32",
56
+ "formattedType" : "MultiArray (Int32 1)",
57
+ "shortDescription" : "",
58
+ "shape" : "[1]",
59
+ "name" : "input_ids",
60
+ "type" : "MultiArray"
61
+ }
62
+ ],
63
+ "generatedClassName" : "CodeEmbedder",
64
+ "method" : "predict"
65
+ }
66
+ ]
qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/model.mil ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})]
3
+ {
4
+ func main<ios18>(tensor<int32, [1]> input_ids) {
5
+ int32 embeddings_batch_dims_0 = const()[name = string("embeddings_batch_dims_0"), val = int32(0)];
6
+ bool embeddings_validate_indices_0 = const()[name = string("embeddings_validate_indices_0"), val = bool(false)];
7
+ tensor<fp16, [3072, 2048]> codec_embedding_weight_to_fp16 = const()[name = string("codec_embedding_weight_to_fp16"), val = tensor<fp16, [3072, 2048]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
8
+ string input_ids_to_int16_dtype_0 = const()[name = string("input_ids_to_int16_dtype_0"), val = string("int16")];
9
+ string cast_2_dtype_0 = const()[name = string("cast_2_dtype_0"), val = string("int32")];
10
+ int32 greater_equal_0_y_0 = const()[name = string("greater_equal_0_y_0"), val = int32(0)];
11
+ tensor<int16, [1]> input_ids_to_int16 = cast(dtype = input_ids_to_int16_dtype_0, x = input_ids)[name = string("cast_5")];
12
+ tensor<int32, [1]> cast_2 = cast(dtype = cast_2_dtype_0, x = input_ids_to_int16)[name = string("cast_4")];
13
+ tensor<bool, [1]> greater_equal_0 = greater_equal(x = cast_2, y = greater_equal_0_y_0)[name = string("greater_equal_0")];
14
+ int32 slice_by_index_0 = const()[name = string("slice_by_index_0"), val = int32(3072)];
15
+ tensor<int32, [1]> add_0 = add(x = cast_2, y = slice_by_index_0)[name = string("add_0")];
16
+ tensor<int32, [1]> select_0 = select(a = cast_2, b = add_0, cond = greater_equal_0)[name = string("select_0")];
17
+ int32 embeddings_cast_fp16_cast_uint16_axis_0 = const()[name = string("embeddings_cast_fp16_cast_uint16_axis_0"), val = int32(0)];
18
+ string select_0_to_int16_dtype_0 = const()[name = string("select_0_to_int16_dtype_0"), val = string("int16")];
19
+ tensor<int16, [1]> select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = string("cast_3")];
20
+ tensor<fp16, [1, 2048]> embeddings_cast_fp16_cast_uint16_cast_uint16 = gather(axis = embeddings_cast_fp16_cast_uint16_axis_0, batch_dims = embeddings_batch_dims_0, indices = select_0_to_int16, validate_indices = embeddings_validate_indices_0, x = codec_embedding_weight_to_fp16)[name = string("embeddings_cast_fp16_cast_uint16_cast_uint16")];
21
+ tensor<int32, [1]> var_18_axes_0 = const()[name = string("op_18_axes_0"), val = tensor<int32, [1]>([2])];
22
+ tensor<fp16, [1, 2048, 1]> var_18_cast_fp16 = expand_dims(axes = var_18_axes_0, x = embeddings_cast_fp16_cast_uint16_cast_uint16)[name = string("op_18_cast_fp16")];
23
+ tensor<int32, [1]> var_20_axes_0 = const()[name = string("op_20_axes_0"), val = tensor<int32, [1]>([3])];
24
+ tensor<fp16, [1, 2048, 1, 1]> input_embeds = expand_dims(axes = var_20_axes_0, x = var_18_cast_fp16)[name = string("op_20_cast_fp16")];
25
+ } -> (input_embeds);
26
+ }
qwen3_tts/code_embedder/12hz-1.7b-customvoice/W16A16/CodeEmbedder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb500c72e2fb86bbba8b89e6810c64485e177bb4bcfc3725373d2c756f28b624
3
+ size 12583040
qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7096ffa7daeda1488227df4f181dd3a628d53238752488eb4f64c08e547992a0
3
+ size 243
qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff1f98f3c58bd419916a56b58d94bf2a203efbbab376e7b12353a6eb7ec394a9
3
+ size 611
qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Mixed (Float16, Palettized (8 bits), UInt8)",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 15 × 2048)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 15, 2048]",
13
+ "name" : "all_logits",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 1)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 5120, 1, 1]",
23
+ "name" : "key_cache_updates",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 1)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 5120, 1, 1]",
33
+ "name" : "value_cache_updates",
34
+ "type" : "MultiArray"
35
+ }
36
+ ],
37
+ "modelParameters" : [
38
+
39
+ ],
40
+ "specificationVersion" : 9,
41
+ "mlProgramOperationTypeHistogram" : {
42
+ "Ios18.expandDims" : 8,
43
+ "Ios18.softmax" : 5,
44
+ "Ios18.mul" : 123,
45
+ "Ios18.matmul" : 10,
46
+ "Ios18.rsqrt" : 21,
47
+ "Ios16.reduceMean" : 21,
48
+ "Split" : 2,
49
+ "Ios18.greaterEqual" : 2,
50
+ "Select" : 2,
51
+ "Ios18.gather" : 2,
52
+ "Ios18.add" : 58,
53
+ "Ios18.reshape" : 40,
54
+ "Ios18.constexprLutToDense" : 50,
55
+ "Ios18.conv" : 50,
56
+ "Ios18.concat" : 23,
57
+ "Ios18.cast" : 5,
58
+ "Ios18.sub" : 1,
59
+ "Ios18.silu" : 5,
60
+ "Ios18.transpose" : 1,
61
+ "Ios18.sliceByIndex" : 100,
62
+ "Ios18.squeeze" : 15
63
+ },
64
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
65
+ "isUpdatable" : "0",
66
+ "stateSchema" : [
67
+
68
+ ],
69
+ "availability" : {
70
+ "macOS" : "15.0",
71
+ "tvOS" : "18.0",
72
+ "visionOS" : "2.0",
73
+ "watchOS" : "11.0",
74
+ "iOS" : "18.0",
75
+ "macCatalyst" : "18.0"
76
+ },
77
+ "modelType" : {
78
+ "name" : "MLModelType_mlProgram"
79
+ },
80
+ "userDefinedMetadata" : {
81
+ "com.github.apple.coremltools.conversion_date" : "2026-02-12",
82
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
83
+ "com.github.apple.coremltools.version" : "9.0",
84
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
85
+ },
86
+ "inputSchema" : [
87
+ {
88
+ "hasShapeFlexibility" : "0",
89
+ "isOptional" : "0",
90
+ "dataType" : "Float16",
91
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 1)",
92
+ "shortDescription" : "",
93
+ "shape" : "[1, 1024, 1, 1]",
94
+ "name" : "input_embeds",
95
+ "type" : "MultiArray"
96
+ },
97
+ {
98
+ "hasShapeFlexibility" : "0",
99
+ "isOptional" : "0",
100
+ "dataType" : "Int32",
101
+ "formattedType" : "MultiArray (Int32 1)",
102
+ "shortDescription" : "",
103
+ "shape" : "[1]",
104
+ "name" : "cache_length",
105
+ "type" : "MultiArray"
106
+ },
107
+ {
108
+ "hasShapeFlexibility" : "0",
109
+ "isOptional" : "0",
110
+ "dataType" : "Float16",
111
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 16)",
112
+ "shortDescription" : "",
113
+ "shape" : "[1, 5120, 1, 16]",
114
+ "name" : "key_cache",
115
+ "type" : "MultiArray"
116
+ },
117
+ {
118
+ "hasShapeFlexibility" : "0",
119
+ "isOptional" : "0",
120
+ "dataType" : "Float16",
121
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 16)",
122
+ "shortDescription" : "",
123
+ "shape" : "[1, 5120, 1, 16]",
124
+ "name" : "value_cache",
125
+ "type" : "MultiArray"
126
+ },
127
+ {
128
+ "hasShapeFlexibility" : "0",
129
+ "isOptional" : "0",
130
+ "dataType" : "Float16",
131
+ "formattedType" : "MultiArray (Float16 1 × 16)",
132
+ "shortDescription" : "",
133
+ "shape" : "[1, 16]",
134
+ "name" : "kv_cache_update_mask",
135
+ "type" : "MultiArray"
136
+ },
137
+ {
138
+ "hasShapeFlexibility" : "0",
139
+ "isOptional" : "0",
140
+ "dataType" : "Float16",
141
+ "formattedType" : "MultiArray (Float16 1 × 16)",
142
+ "shortDescription" : "",
143
+ "shape" : "[1, 16]",
144
+ "name" : "key_padding_mask",
145
+ "type" : "MultiArray"
146
+ }
147
+ ],
148
+ "generatedClassName" : "MultiCodeDecoder_8_bit",
149
+ "method" : "predict"
150
+ }
151
+ ]
qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
qwen3_tts/multi_code_decoder/12hz-0.6b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:469c3db16d54f85f771d5fbb6817f86c26a92c19e1545e0a452edaeda09f7a59
3
+ size 110173568
qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90d4090f7d667dc9e55ca0698f543928678e69a645d71fc343c52e7398724f65
3
+ size 243
qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91e4d1bd298329d7300432d70c3054ccc58f5f250cb500a44f0b7d4075720186
3
+ size 611
qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Mixed (Float16, Palettized (8 bits), UInt8)",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 15 × 2048)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 15, 2048]",
13
+ "name" : "all_logits",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 1)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 5120, 1, 1]",
23
+ "name" : "key_cache_updates",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 1)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 5120, 1, 1]",
33
+ "name" : "value_cache_updates",
34
+ "type" : "MultiArray"
35
+ }
36
+ ],
37
+ "modelParameters" : [
38
+
39
+ ],
40
+ "specificationVersion" : 9,
41
+ "mlProgramOperationTypeHistogram" : {
42
+ "Ios18.expandDims" : 8,
43
+ "Ios18.softmax" : 5,
44
+ "Ios18.mul" : 123,
45
+ "Ios18.matmul" : 10,
46
+ "Ios18.rsqrt" : 21,
47
+ "Ios16.reduceMean" : 21,
48
+ "Split" : 2,
49
+ "Ios18.greaterEqual" : 2,
50
+ "Select" : 2,
51
+ "Ios18.gather" : 2,
52
+ "Ios18.add" : 58,
53
+ "Ios18.reshape" : 40,
54
+ "Ios18.constexprLutToDense" : 51,
55
+ "Ios18.conv" : 51,
56
+ "Ios18.concat" : 23,
57
+ "Ios18.cast" : 5,
58
+ "Ios18.sub" : 1,
59
+ "Ios18.silu" : 5,
60
+ "Ios18.transpose" : 1,
61
+ "Ios18.sliceByIndex" : 100,
62
+ "Ios18.squeeze" : 15
63
+ },
64
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
65
+ "isUpdatable" : "0",
66
+ "stateSchema" : [
67
+
68
+ ],
69
+ "availability" : {
70
+ "macOS" : "15.0",
71
+ "tvOS" : "18.0",
72
+ "visionOS" : "2.0",
73
+ "watchOS" : "11.0",
74
+ "iOS" : "18.0",
75
+ "macCatalyst" : "18.0"
76
+ },
77
+ "modelType" : {
78
+ "name" : "MLModelType_mlProgram"
79
+ },
80
+ "userDefinedMetadata" : {
81
+ "com.github.apple.coremltools.conversion_date" : "2026-02-11",
82
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
83
+ "com.github.apple.coremltools.version" : "9.0",
84
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
85
+ },
86
+ "inputSchema" : [
87
+ {
88
+ "hasShapeFlexibility" : "0",
89
+ "isOptional" : "0",
90
+ "dataType" : "Float16",
91
+ "formattedType" : "MultiArray (Float16 1 × 2048 × 1 × 1)",
92
+ "shortDescription" : "",
93
+ "shape" : "[1, 2048, 1, 1]",
94
+ "name" : "input_embeds",
95
+ "type" : "MultiArray"
96
+ },
97
+ {
98
+ "hasShapeFlexibility" : "0",
99
+ "isOptional" : "0",
100
+ "dataType" : "Int32",
101
+ "formattedType" : "MultiArray (Int32 1)",
102
+ "shortDescription" : "",
103
+ "shape" : "[1]",
104
+ "name" : "cache_length",
105
+ "type" : "MultiArray"
106
+ },
107
+ {
108
+ "hasShapeFlexibility" : "0",
109
+ "isOptional" : "0",
110
+ "dataType" : "Float16",
111
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 16)",
112
+ "shortDescription" : "",
113
+ "shape" : "[1, 5120, 1, 16]",
114
+ "name" : "key_cache",
115
+ "type" : "MultiArray"
116
+ },
117
+ {
118
+ "hasShapeFlexibility" : "0",
119
+ "isOptional" : "0",
120
+ "dataType" : "Float16",
121
+ "formattedType" : "MultiArray (Float16 1 × 5120 × 1 × 16)",
122
+ "shortDescription" : "",
123
+ "shape" : "[1, 5120, 1, 16]",
124
+ "name" : "value_cache",
125
+ "type" : "MultiArray"
126
+ },
127
+ {
128
+ "hasShapeFlexibility" : "0",
129
+ "isOptional" : "0",
130
+ "dataType" : "Float16",
131
+ "formattedType" : "MultiArray (Float16 1 × 16)",
132
+ "shortDescription" : "",
133
+ "shape" : "[1, 16]",
134
+ "name" : "kv_cache_update_mask",
135
+ "type" : "MultiArray"
136
+ },
137
+ {
138
+ "hasShapeFlexibility" : "0",
139
+ "isOptional" : "0",
140
+ "dataType" : "Float16",
141
+ "formattedType" : "MultiArray (Float16 1 × 16)",
142
+ "shortDescription" : "",
143
+ "shape" : "[1, 16]",
144
+ "name" : "key_padding_mask",
145
+ "type" : "MultiArray"
146
+ }
147
+ ],
148
+ "generatedClassName" : "MultiCodeDecoder_8_bit",
149
+ "method" : "predict"
150
+ }
151
+ ]
qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
qwen3_tts/multi_code_decoder/12hz-1.7b-customvoice/W8A16/MultiCodeDecoder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c3d559c2413b8bffbd5f1d4e1a3cab6b084e66c369c8d02f8d1e5cee509cc85
3
+ size 112273472
qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18324dfd306ef038fcb0e0a279809ca8f8ea24e92d6bac7c15fbd81df6102220
3
+ size 243
qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08f3c2e364b7d50fd16600725143bdfe80127688d9dd05a9391a90fdcf30d29c
3
+ size 380
qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Float16",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 1)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1024, 1, 1]",
13
+ "name" : "input_embeds",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "modelParameters" : [
18
+
19
+ ],
20
+ "specificationVersion" : 9,
21
+ "mlProgramOperationTypeHistogram" : {
22
+ "Ios18.greaterEqual" : 1,
23
+ "Ios18.add" : 1,
24
+ "Ios18.cast" : 3,
25
+ "Select" : 1,
26
+ "Ios18.gather" : 1,
27
+ "Ios18.expandDims" : 2
28
+ },
29
+ "computePrecision" : "Mixed (Float16, Int16, Int32)",
30
+ "isUpdatable" : "0",
31
+ "stateSchema" : [
32
+
33
+ ],
34
+ "availability" : {
35
+ "macOS" : "15.0",
36
+ "tvOS" : "18.0",
37
+ "visionOS" : "2.0",
38
+ "watchOS" : "11.0",
39
+ "iOS" : "18.0",
40
+ "macCatalyst" : "18.0"
41
+ },
42
+ "modelType" : {
43
+ "name" : "MLModelType_mlProgram"
44
+ },
45
+ "userDefinedMetadata" : {
46
+ "com.github.apple.coremltools.conversion_date" : "2026-02-06",
47
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
48
+ "com.github.apple.coremltools.version" : "9.0",
49
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
50
+ },
51
+ "inputSchema" : [
52
+ {
53
+ "hasShapeFlexibility" : "0",
54
+ "isOptional" : "0",
55
+ "dataType" : "Int32",
56
+ "formattedType" : "MultiArray (Int32 1)",
57
+ "shortDescription" : "",
58
+ "shape" : "[1]",
59
+ "name" : "input_ids",
60
+ "type" : "MultiArray"
61
+ }
62
+ ],
63
+ "generatedClassName" : "MultiCodeEmbedder",
64
+ "method" : "predict"
65
+ }
66
+ ]
qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/model.mil ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})]
3
+ {
4
+ func main<ios18>(tensor<int32, [1]> input_ids) {
5
+ int32 embeddings_batch_dims_0 = const()[name = string("embeddings_batch_dims_0"), val = int32(0)];
6
+ bool embeddings_validate_indices_0 = const()[name = string("embeddings_validate_indices_0"), val = bool(false)];
7
+ tensor<fp16, [30720, 1024]> codec_embedding_weight_to_fp16 = const()[name = string("codec_embedding_weight_to_fp16"), val = tensor<fp16, [30720, 1024]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
8
+ string input_ids_to_int16_dtype_0 = const()[name = string("input_ids_to_int16_dtype_0"), val = string("int16")];
9
+ string cast_2_dtype_0 = const()[name = string("cast_2_dtype_0"), val = string("int32")];
10
+ int32 greater_equal_0_y_0 = const()[name = string("greater_equal_0_y_0"), val = int32(0)];
11
+ tensor<int16, [1]> input_ids_to_int16 = cast(dtype = input_ids_to_int16_dtype_0, x = input_ids)[name = string("cast_5")];
12
+ tensor<int32, [1]> cast_2 = cast(dtype = cast_2_dtype_0, x = input_ids_to_int16)[name = string("cast_4")];
13
+ tensor<bool, [1]> greater_equal_0 = greater_equal(x = cast_2, y = greater_equal_0_y_0)[name = string("greater_equal_0")];
14
+ int32 slice_by_index_0 = const()[name = string("slice_by_index_0"), val = int32(30720)];
15
+ tensor<int32, [1]> add_0 = add(x = cast_2, y = slice_by_index_0)[name = string("add_0")];
16
+ tensor<int32, [1]> select_0 = select(a = cast_2, b = add_0, cond = greater_equal_0)[name = string("select_0")];
17
+ int32 embeddings_cast_fp16_cast_uint16_axis_0 = const()[name = string("embeddings_cast_fp16_cast_uint16_axis_0"), val = int32(0)];
18
+ string select_0_to_int16_dtype_0 = const()[name = string("select_0_to_int16_dtype_0"), val = string("int16")];
19
+ tensor<int16, [1]> select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = string("cast_3")];
20
+ tensor<fp16, [1, 1024]> embeddings_cast_fp16_cast_uint16_cast_uint16 = gather(axis = embeddings_cast_fp16_cast_uint16_axis_0, batch_dims = embeddings_batch_dims_0, indices = select_0_to_int16, validate_indices = embeddings_validate_indices_0, x = codec_embedding_weight_to_fp16)[name = string("embeddings_cast_fp16_cast_uint16_cast_uint16")];
21
+ tensor<int32, [1]> var_18_axes_0 = const()[name = string("op_18_axes_0"), val = tensor<int32, [1]>([2])];
22
+ tensor<fp16, [1, 1024, 1]> var_18_cast_fp16 = expand_dims(axes = var_18_axes_0, x = embeddings_cast_fp16_cast_uint16_cast_uint16)[name = string("op_18_cast_fp16")];
23
+ tensor<int32, [1]> var_20_axes_0 = const()[name = string("op_20_axes_0"), val = tensor<int32, [1]>([3])];
24
+ tensor<fp16, [1, 1024, 1, 1]> input_embeds = expand_dims(axes = var_20_axes_0, x = var_18_cast_fp16)[name = string("op_20_cast_fp16")];
25
+ } -> (input_embeds);
26
+ }
qwen3_tts/multi_code_embedder/12hz-0.6b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f318d23d7db53b299b3aa472844a721d2b6245550fb57558b0d2c5ebbcd3fd82
3
+ size 62914688
qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59e7994cafd7a43d8c1af61044f330357ba05ee9817d11a3e51bcc3000d598f0
3
+ size 243
qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d20121b9a714012ce61f8257ebd047be23c2994a7ef0a1c40225fdbdc72c2dd8
3
+ size 380
qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Float16",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 2048 × 1 × 1)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 2048, 1, 1]",
13
+ "name" : "input_embeds",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "modelParameters" : [
18
+
19
+ ],
20
+ "specificationVersion" : 9,
21
+ "mlProgramOperationTypeHistogram" : {
22
+ "Ios18.greaterEqual" : 1,
23
+ "Ios18.add" : 1,
24
+ "Ios18.cast" : 3,
25
+ "Select" : 1,
26
+ "Ios18.gather" : 1,
27
+ "Ios18.expandDims" : 2
28
+ },
29
+ "computePrecision" : "Mixed (Float16, Int16, Int32)",
30
+ "isUpdatable" : "0",
31
+ "stateSchema" : [
32
+
33
+ ],
34
+ "availability" : {
35
+ "macOS" : "15.0",
36
+ "tvOS" : "18.0",
37
+ "visionOS" : "2.0",
38
+ "watchOS" : "11.0",
39
+ "iOS" : "18.0",
40
+ "macCatalyst" : "18.0"
41
+ },
42
+ "modelType" : {
43
+ "name" : "MLModelType_mlProgram"
44
+ },
45
+ "userDefinedMetadata" : {
46
+ "com.github.apple.coremltools.conversion_date" : "2026-02-12",
47
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
48
+ "com.github.apple.coremltools.version" : "9.0",
49
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
50
+ },
51
+ "inputSchema" : [
52
+ {
53
+ "hasShapeFlexibility" : "0",
54
+ "isOptional" : "0",
55
+ "dataType" : "Int32",
56
+ "formattedType" : "MultiArray (Int32 1)",
57
+ "shortDescription" : "",
58
+ "shape" : "[1]",
59
+ "name" : "input_ids",
60
+ "type" : "MultiArray"
61
+ }
62
+ ],
63
+ "generatedClassName" : "MultiCodeEmbedder",
64
+ "method" : "predict"
65
+ }
66
+ ]
qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/model.mil ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})]
3
+ {
4
+ func main<ios18>(tensor<int32, [1]> input_ids) {
5
+ int32 embeddings_batch_dims_0 = const()[name = string("embeddings_batch_dims_0"), val = int32(0)];
6
+ bool embeddings_validate_indices_0 = const()[name = string("embeddings_validate_indices_0"), val = bool(false)];
7
+ tensor<fp16, [30720, 2048]> codec_embedding_weight_to_fp16 = const()[name = string("codec_embedding_weight_to_fp16"), val = tensor<fp16, [30720, 2048]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
8
+ string input_ids_to_int16_dtype_0 = const()[name = string("input_ids_to_int16_dtype_0"), val = string("int16")];
9
+ string cast_2_dtype_0 = const()[name = string("cast_2_dtype_0"), val = string("int32")];
10
+ int32 greater_equal_0_y_0 = const()[name = string("greater_equal_0_y_0"), val = int32(0)];
11
+ tensor<int16, [1]> input_ids_to_int16 = cast(dtype = input_ids_to_int16_dtype_0, x = input_ids)[name = string("cast_5")];
12
+ tensor<int32, [1]> cast_2 = cast(dtype = cast_2_dtype_0, x = input_ids_to_int16)[name = string("cast_4")];
13
+ tensor<bool, [1]> greater_equal_0 = greater_equal(x = cast_2, y = greater_equal_0_y_0)[name = string("greater_equal_0")];
14
+ int32 slice_by_index_0 = const()[name = string("slice_by_index_0"), val = int32(30720)];
15
+ tensor<int32, [1]> add_0 = add(x = cast_2, y = slice_by_index_0)[name = string("add_0")];
16
+ tensor<int32, [1]> select_0 = select(a = cast_2, b = add_0, cond = greater_equal_0)[name = string("select_0")];
17
+ int32 embeddings_cast_fp16_cast_uint16_axis_0 = const()[name = string("embeddings_cast_fp16_cast_uint16_axis_0"), val = int32(0)];
18
+ string select_0_to_int16_dtype_0 = const()[name = string("select_0_to_int16_dtype_0"), val = string("int16")];
19
+ tensor<int16, [1]> select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = string("cast_3")];
20
+ tensor<fp16, [1, 2048]> embeddings_cast_fp16_cast_uint16_cast_uint16 = gather(axis = embeddings_cast_fp16_cast_uint16_axis_0, batch_dims = embeddings_batch_dims_0, indices = select_0_to_int16, validate_indices = embeddings_validate_indices_0, x = codec_embedding_weight_to_fp16)[name = string("embeddings_cast_fp16_cast_uint16_cast_uint16")];
21
+ tensor<int32, [1]> var_18_axes_0 = const()[name = string("op_18_axes_0"), val = tensor<int32, [1]>([2])];
22
+ tensor<fp16, [1, 2048, 1]> var_18_cast_fp16 = expand_dims(axes = var_18_axes_0, x = embeddings_cast_fp16_cast_uint16_cast_uint16)[name = string("op_18_cast_fp16")];
23
+ tensor<int32, [1]> var_20_axes_0 = const()[name = string("op_20_axes_0"), val = tensor<int32, [1]>([3])];
24
+ tensor<fp16, [1, 2048, 1, 1]> input_embeds = expand_dims(axes = var_20_axes_0, x = var_18_cast_fp16)[name = string("op_20_cast_fp16")];
25
+ } -> (input_embeds);
26
+ }
qwen3_tts/multi_code_embedder/12hz-1.7b-customvoice/W16A16/MultiCodeEmbedder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46f8ac7940915bf46cb15cd16ac06c1cf306044aebf6ade6253831539e61cb88
3
+ size 125829248
qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e5662e1e7fb3aa28cbbf01e35c3dd4cc11c795ce151f05068ec220092157eb8
3
+ size 243
qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:608c8c933a512f362e79193430195402b64781ee57731099225c2e83ab9dbb8d
3
+ size 681
qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Mixed (Float16, Int32, Palettized (8 bits), UInt8)",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 1 × 1 × 1920)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 1, 1920]",
13
+ "name" : "audio",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 1)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 8192, 1, 1]",
23
+ "name" : "key_cache_updates",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 1)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 8192, 1, 1]",
33
+ "name" : "value_cache_updates",
34
+ "type" : "MultiArray"
35
+ },
36
+ {
37
+ "hasShapeFlexibility" : "0",
38
+ "isOptional" : "0",
39
+ "dataType" : "Float16",
40
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 1)",
41
+ "shortDescription" : "",
42
+ "shape" : "[1, 1024, 1, 1]",
43
+ "name" : "hidden_context_update",
44
+ "type" : "MultiArray"
45
+ }
46
+ ],
47
+ "modelParameters" : [
48
+
49
+ ],
50
+ "specificationVersion" : 9,
51
+ "mlProgramOperationTypeHistogram" : {
52
+ "Ios18.expandDims" : 26,
53
+ "Ios18.mul" : 234,
54
+ "Ios18.softmax" : 8,
55
+ "Ios18.rsqrt" : 17,
56
+ "Ios18.matmul" : 16,
57
+ "Ios16.reduceMean" : 17,
58
+ "Split" : 3,
59
+ "Ios18.greaterEqual" : 1,
60
+ "Select" : 1,
61
+ "Ios18.sin" : 29,
62
+ "Tile" : 2,
63
+ "Ios18.gather" : 18,
64
+ "Ios18.add" : 132,
65
+ "Ios18.layerNorm" : 2,
66
+ "Ios18.reshape" : 56,
67
+ "Pad" : 17,
68
+ "Ios18.constexprLutToDense" : 116,
69
+ "Ios18.conv" : 93,
70
+ "Ios18.concat" : 19,
71
+ "Ios18.transpose" : 30,
72
+ "Ios18.sub" : 1,
73
+ "Ios18.cast" : 19,
74
+ "Ios18.silu" : 8,
75
+ "Ios18.gelu" : 2,
76
+ "Ios18.clip" : 1,
77
+ "Ios18.convTranspose" : 6,
78
+ "Ios18.sliceByIndex" : 39,
79
+ "Ios18.squeeze" : 18
80
+ },
81
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
82
+ "isUpdatable" : "0",
83
+ "stateSchema" : [
84
+
85
+ ],
86
+ "availability" : {
87
+ "macOS" : "15.0",
88
+ "tvOS" : "18.0",
89
+ "visionOS" : "2.0",
90
+ "watchOS" : "11.0",
91
+ "iOS" : "18.0",
92
+ "macCatalyst" : "18.0"
93
+ },
94
+ "modelType" : {
95
+ "name" : "MLModelType_mlProgram"
96
+ },
97
+ "userDefinedMetadata" : {
98
+ "com.github.apple.coremltools.conversion_date" : "2026-02-09",
99
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
100
+ "com.github.apple.coremltools.version" : "9.0",
101
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
102
+ },
103
+ "inputSchema" : [
104
+ {
105
+ "hasShapeFlexibility" : "0",
106
+ "isOptional" : "0",
107
+ "dataType" : "Int32",
108
+ "formattedType" : "MultiArray (Int32 1 × 16 × 1)",
109
+ "shortDescription" : "",
110
+ "shape" : "[1, 16, 1]",
111
+ "name" : "audio_codes",
112
+ "type" : "MultiArray"
113
+ },
114
+ {
115
+ "hasShapeFlexibility" : "0",
116
+ "isOptional" : "0",
117
+ "dataType" : "Int32",
118
+ "formattedType" : "MultiArray (Int32 1)",
119
+ "shortDescription" : "",
120
+ "shape" : "[1]",
121
+ "name" : "cache_length",
122
+ "type" : "MultiArray"
123
+ },
124
+ {
125
+ "hasShapeFlexibility" : "0",
126
+ "isOptional" : "0",
127
+ "dataType" : "Float16",
128
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 256)",
129
+ "shortDescription" : "",
130
+ "shape" : "[1, 8192, 1, 256]",
131
+ "name" : "key_cache",
132
+ "type" : "MultiArray"
133
+ },
134
+ {
135
+ "hasShapeFlexibility" : "0",
136
+ "isOptional" : "0",
137
+ "dataType" : "Float16",
138
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 256)",
139
+ "shortDescription" : "",
140
+ "shape" : "[1, 8192, 1, 256]",
141
+ "name" : "value_cache",
142
+ "type" : "MultiArray"
143
+ },
144
+ {
145
+ "hasShapeFlexibility" : "0",
146
+ "isOptional" : "0",
147
+ "dataType" : "Float16",
148
+ "formattedType" : "MultiArray (Float16 1 × 256)",
149
+ "shortDescription" : "",
150
+ "shape" : "[1, 256]",
151
+ "name" : "kv_cache_update_mask",
152
+ "type" : "MultiArray"
153
+ },
154
+ {
155
+ "hasShapeFlexibility" : "0",
156
+ "isOptional" : "0",
157
+ "dataType" : "Float16",
158
+ "formattedType" : "MultiArray (Float16 1 × 256)",
159
+ "shortDescription" : "",
160
+ "shape" : "[1, 256]",
161
+ "name" : "key_padding_mask",
162
+ "type" : "MultiArray"
163
+ },
164
+ {
165
+ "hasShapeFlexibility" : "0",
166
+ "isOptional" : "0",
167
+ "dataType" : "Float16",
168
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 4)",
169
+ "shortDescription" : "",
170
+ "shape" : "[1, 1024, 1, 4]",
171
+ "name" : "hidden_context",
172
+ "type" : "MultiArray"
173
+ }
174
+ ],
175
+ "generatedClassName" : "SpeechDecoder_8_bit",
176
+ "method" : "predict"
177
+ }
178
+ ]
qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
qwen3_tts/speech_decoder/12hz-0.6b-customvoice/W8A16/SpeechDecoder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9a352536383e01d53d1d0ab227145e6e02880b2281715bc0ff12a44f7be89f
3
+ size 114215488
qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a7c001fcef8399080af242e87d12ad44d7fc37fded8007e7fce65461346a52c
3
+ size 243
qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:554158c2df351252d84d8295d438b004a9bae5e65ae9dcd802f13de13bacd41a
3
+ size 681
qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/metadata.json ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Mixed (Float16, Int32)",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 1 × 1 × 1920)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 1, 1920]",
13
+ "name" : "audio",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 1)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 8192, 1, 1]",
23
+ "name" : "key_cache_updates",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 1)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 8192, 1, 1]",
33
+ "name" : "value_cache_updates",
34
+ "type" : "MultiArray"
35
+ },
36
+ {
37
+ "hasShapeFlexibility" : "0",
38
+ "isOptional" : "0",
39
+ "dataType" : "Float16",
40
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 1)",
41
+ "shortDescription" : "",
42
+ "shape" : "[1, 1024, 1, 1]",
43
+ "name" : "hidden_context_update",
44
+ "type" : "MultiArray"
45
+ }
46
+ ],
47
+ "modelParameters" : [
48
+
49
+ ],
50
+ "specificationVersion" : 9,
51
+ "mlProgramOperationTypeHistogram" : {
52
+ "Ios18.expandDims" : 26,
53
+ "Ios18.mul" : 234,
54
+ "Ios18.softmax" : 8,
55
+ "Ios18.matmul" : 16,
56
+ "Ios18.rsqrt" : 17,
57
+ "Ios16.reduceMean" : 17,
58
+ "Split" : 3,
59
+ "Ios18.greaterEqual" : 1,
60
+ "Select" : 1,
61
+ "Ios18.sin" : 29,
62
+ "Tile" : 2,
63
+ "Ios18.gather" : 18,
64
+ "Ios18.add" : 132,
65
+ "Ios18.layerNorm" : 2,
66
+ "Ios18.reshape" : 56,
67
+ "Pad" : 17,
68
+ "Ios18.conv" : 93,
69
+ "Ios18.concat" : 19,
70
+ "Ios18.transpose" : 30,
71
+ "Ios18.sub" : 1,
72
+ "Ios18.cast" : 19,
73
+ "Ios18.silu" : 8,
74
+ "Ios18.gelu" : 2,
75
+ "Ios18.clip" : 1,
76
+ "Ios18.convTranspose" : 6,
77
+ "Ios18.sliceByIndex" : 39,
78
+ "Ios18.squeeze" : 18
79
+ },
80
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
81
+ "isUpdatable" : "0",
82
+ "stateSchema" : [
83
+
84
+ ],
85
+ "availability" : {
86
+ "macOS" : "15.0",
87
+ "tvOS" : "18.0",
88
+ "visionOS" : "2.0",
89
+ "watchOS" : "11.0",
90
+ "iOS" : "18.0",
91
+ "macCatalyst" : "18.0"
92
+ },
93
+ "modelType" : {
94
+ "name" : "MLModelType_mlProgram"
95
+ },
96
+ "userDefinedMetadata" : {
97
+ "com.github.apple.coremltools.conversion_date" : "2026-02-12",
98
+ "com.github.apple.coremltools.source" : "torch==2.8.0",
99
+ "com.github.apple.coremltools.version" : "9.0",
100
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
101
+ },
102
+ "inputSchema" : [
103
+ {
104
+ "hasShapeFlexibility" : "0",
105
+ "isOptional" : "0",
106
+ "dataType" : "Int32",
107
+ "formattedType" : "MultiArray (Int32 1 × 16 × 1)",
108
+ "shortDescription" : "",
109
+ "shape" : "[1, 16, 1]",
110
+ "name" : "audio_codes",
111
+ "type" : "MultiArray"
112
+ },
113
+ {
114
+ "hasShapeFlexibility" : "0",
115
+ "isOptional" : "0",
116
+ "dataType" : "Int32",
117
+ "formattedType" : "MultiArray (Int32 1)",
118
+ "shortDescription" : "",
119
+ "shape" : "[1]",
120
+ "name" : "cache_length",
121
+ "type" : "MultiArray"
122
+ },
123
+ {
124
+ "hasShapeFlexibility" : "0",
125
+ "isOptional" : "0",
126
+ "dataType" : "Float16",
127
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 256)",
128
+ "shortDescription" : "",
129
+ "shape" : "[1, 8192, 1, 256]",
130
+ "name" : "key_cache",
131
+ "type" : "MultiArray"
132
+ },
133
+ {
134
+ "hasShapeFlexibility" : "0",
135
+ "isOptional" : "0",
136
+ "dataType" : "Float16",
137
+ "formattedType" : "MultiArray (Float16 1 × 8192 × 1 × 256)",
138
+ "shortDescription" : "",
139
+ "shape" : "[1, 8192, 1, 256]",
140
+ "name" : "value_cache",
141
+ "type" : "MultiArray"
142
+ },
143
+ {
144
+ "hasShapeFlexibility" : "0",
145
+ "isOptional" : "0",
146
+ "dataType" : "Float16",
147
+ "formattedType" : "MultiArray (Float16 1 × 256)",
148
+ "shortDescription" : "",
149
+ "shape" : "[1, 256]",
150
+ "name" : "kv_cache_update_mask",
151
+ "type" : "MultiArray"
152
+ },
153
+ {
154
+ "hasShapeFlexibility" : "0",
155
+ "isOptional" : "0",
156
+ "dataType" : "Float16",
157
+ "formattedType" : "MultiArray (Float16 1 × 256)",
158
+ "shortDescription" : "",
159
+ "shape" : "[1, 256]",
160
+ "name" : "key_padding_mask",
161
+ "type" : "MultiArray"
162
+ },
163
+ {
164
+ "hasShapeFlexibility" : "0",
165
+ "isOptional" : "0",
166
+ "dataType" : "Float16",
167
+ "formattedType" : "MultiArray (Float16 1 × 1024 × 1 × 4)",
168
+ "shortDescription" : "",
169
+ "shape" : "[1, 1024, 1, 4]",
170
+ "name" : "hidden_context",
171
+ "type" : "MultiArray"
172
+ }
173
+ ],
174
+ "generatedClassName" : "SpeechDecoder",
175
+ "method" : "predict"
176
+ }
177
+ ]
qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
qwen3_tts/speech_decoder/12hz-1.7b-customvoice/W8A16/SpeechDecoder.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c90dc149272173de6cdae46826e1ef4265c53b8477b1de19c02843d4e38729e7
3
+ size 228134208