seba commited on
Commit
d999afe
·
1 Parent(s): ba1bb1b

Add VibeVoice CoreML models

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitignore +1 -0
  2. acoustic_connector.mlmodelc/analytics/coremldata.bin +3 -0
  3. acoustic_connector.mlmodelc/coremldata.bin +3 -0
  4. acoustic_connector.mlmodelc/model.mil +39 -0
  5. acoustic_connector.mlmodelc/weights/weight.bin +3 -0
  6. acoustic_connector.mlpackage/Data/com.apple.CoreML/model.mlmodel +3 -0
  7. acoustic_connector.mlpackage/Data/com.apple.CoreML/weights/weight.bin +3 -0
  8. acoustic_connector.mlpackage/Manifest.json +18 -0
  9. decoder_coreml_12_ne.mlmodelc/analytics/coremldata.bin +3 -0
  10. decoder_coreml_12_ne.mlmodelc/coremldata.bin +3 -0
  11. decoder_coreml_12_ne.mlmodelc/model.mil +0 -0
  12. decoder_coreml_12_ne.mlmodelc/weights/weight.bin +3 -0
  13. diffusion_head_model.mlmodelc/analytics/coremldata.bin +3 -0
  14. diffusion_head_model.mlmodelc/coremldata.bin +3 -0
  15. diffusion_head_model.mlmodelc/model.mil +285 -0
  16. diffusion_head_model.mlmodelc/weights/weight.bin +3 -0
  17. tts_eos_classifier.mlmodelc/analytics/coremldata.bin +3 -0
  18. tts_eos_classifier.mlmodelc/coremldata.bin +3 -0
  19. tts_eos_classifier.mlmodelc/model.mil +23 -0
  20. tts_eos_classifier.mlmodelc/weights/weight.bin +3 -0
  21. tts_input_types.npy +3 -0
  22. vibe_voice_lm_model_seqlen_32.mlmodelc/analytics/coremldata.bin +3 -0
  23. vibe_voice_lm_model_seqlen_32.mlmodelc/coremldata.bin +3 -0
  24. vibe_voice_lm_model_seqlen_32.mlmodelc/model.mil +0 -0
  25. vibe_voice_lm_model_seqlen_32.mlmodelc/weights/weight.bin +3 -0
  26. vibe_voice_tts_lm_model_seqlen_8.mlmodelc/analytics/coremldata.bin +3 -0
  27. vibe_voice_tts_lm_model_seqlen_8.mlmodelc/coremldata.bin +3 -0
  28. vibe_voice_tts_lm_model_seqlen_8.mlmodelc/model.mil +0 -0
  29. vibe_voice_tts_lm_model_seqlen_8.mlmodelc/weights/weight.bin +3 -0
  30. vibevoice_embeddings.npy +3 -0
  31. vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/analytics/coremldata.bin +3 -0
  32. vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/coremldata.bin +3 -0
  33. vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/model.mil +0 -0
  34. vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/weights/weight.bin +3 -0
  35. voices/streaming_model/de-Spk0_man.npz +3 -0
  36. voices/streaming_model/de-Spk1_woman.npz +3 -0
  37. voices/streaming_model/en-Carter_man.npz +3 -0
  38. voices/streaming_model/en-Davis_man.npz +3 -0
  39. voices/streaming_model/en-Emma_woman.npz +3 -0
  40. voices/streaming_model/en-Frank_man.npz +3 -0
  41. voices/streaming_model/en-Grace_woman.npz +3 -0
  42. voices/streaming_model/en-Mike_man.npz +3 -0
  43. voices/streaming_model/fr-Spk0_man.npz +3 -0
  44. voices/streaming_model/fr-Spk1_woman.npz +3 -0
  45. voices/streaming_model/in-Samuel_man.npz +3 -0
  46. voices/streaming_model/it-Spk0_woman.npz +3 -0
  47. voices/streaming_model/it-Spk1_man.npz +3 -0
  48. voices/streaming_model/jp-Spk0_man.npz +3 -0
  49. voices/streaming_model/jp-Spk1_woman.npz +3 -0
  50. voices/streaming_model/kr-Spk0_woman.npz +3 -0
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ *.DS_Store
acoustic_connector.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2249a24e4950fe95fee340914d53f3bab173945d674fcaaff1fe3ef8e8b8b91d
3
+ size 243
acoustic_connector.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e8bc76edab8d50f1a45b48c600a7417bbc08feed346a25be77a291660dfa9eb
3
+ size 311
acoustic_connector.mlmodelc/model.mil ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3405.2.1"}, {"coremlc-version", "3405.2.1"}, {"coremltools-component-milinternal", ""}, {"coremltools-version", "9.0"}})]
3
+ {
4
+ func main<ios18>(tensor<fp16, [1, 64, 1]> input_x) {
5
+ string x_1_pad_type_0 = const()[name = string("x_1_pad_type_0"), val = string("valid")];
6
+ tensor<int32, [1]> x_1_strides_0 = const()[name = string("x_1_strides_0"), val = tensor<int32, [1]>([1])];
7
+ tensor<int32, [2]> x_1_pad_0 = const()[name = string("x_1_pad_0"), val = tensor<int32, [2]>([0, 0])];
8
+ tensor<int32, [1]> x_1_dilations_0 = const()[name = string("x_1_dilations_0"), val = tensor<int32, [1]>([1])];
9
+ int32 x_1_groups_0 = const()[name = string("x_1_groups_0"), val = int32(1)];
10
+ tensor<fp16, [896, 64, 1]> var_9_to_fp16 = const()[name = string("op_9_to_fp16"), val = tensor<fp16, [896, 64, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
11
+ tensor<fp16, [896]> fc1_bias_to_fp16 = const()[name = string("fc1_bias_to_fp16"), val = tensor<fp16, [896]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(114816)))];
12
+ tensor<fp16, [1, 896, 1]> x_1_cast_fp16 = conv(bias = fc1_bias_to_fp16, dilations = x_1_dilations_0, groups = x_1_groups_0, pad = x_1_pad_0, pad_type = x_1_pad_type_0, strides = x_1_strides_0, weight = var_9_to_fp16, x = input_x)[name = string("x_1_cast_fp16")];
13
+ int32 var_20 = const()[name = string("op_20"), val = int32(1)];
14
+ tensor<int32, [1]> x_3_axes_0 = const()[name = string("x_3_axes_0"), val = tensor<int32, [1]>([-2])];
15
+ tensor<fp16, [1, 896, 1, 1]> x_3_cast_fp16 = expand_dims(axes = x_3_axes_0, x = x_1_cast_fp16)[name = string("x_3_cast_fp16")];
16
+ fp16 const_0_promoted_to_fp16 = const()[name = string("const_0_promoted_to_fp16"), val = fp16(-0x1p+0)];
17
+ tensor<fp16, [1, 896, 1, 1]> var_25_cast_fp16 = mul(x = x_3_cast_fp16, y = const_0_promoted_to_fp16)[name = string("op_25_cast_fp16")];
18
+ bool x_5_interleave_0 = const()[name = string("x_5_interleave_0"), val = bool(false)];
19
+ tensor<fp16, [1, 1792, 1, 1]> x_5_cast_fp16 = concat(axis = var_20, interleave = x_5_interleave_0, values = (x_3_cast_fp16, var_25_cast_fp16))[name = string("x_5_cast_fp16")];
20
+ tensor<int32, [1]> out_1_axes_0 = const()[name = string("out_1_axes_0"), val = tensor<int32, [1]>([1])];
21
+ fp16 var_35_to_fp16 = const()[name = string("op_35_to_fp16"), val = fp16(0x1.1p-20)];
22
+ tensor<fp16, [1, 1792, 1, 1]> out_1_cast_fp16 = layer_norm(axes = out_1_axes_0, epsilon = var_35_to_fp16, x = x_5_cast_fp16)[name = string("out_1_cast_fp16")];
23
+ tensor<fp16, [1, 1792, 1, 1]> norm_weight_to_fp16 = const()[name = string("norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(116672)))];
24
+ tensor<fp16, [1, 1792, 1, 1]> out_3_cast_fp16 = mul(x = out_1_cast_fp16, y = norm_weight_to_fp16)[name = string("out_3_cast_fp16")];
25
+ tensor<int32, [2]> var_41_split_sizes_0 = const()[name = string("op_41_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
26
+ int32 var_41_axis_0 = const()[name = string("op_41_axis_0"), val = int32(1)];
27
+ tensor<fp16, [1, 896, 1, 1]> var_41_cast_fp16_0, tensor<fp16, [1, 896, 1, 1]> var_41_cast_fp16_1 = split(axis = var_41_axis_0, split_sizes = var_41_split_sizes_0, x = out_3_cast_fp16)[name = string("op_41_cast_fp16")];
28
+ tensor<int32, [1]> x_axes_0 = const()[name = string("x_axes_0"), val = tensor<int32, [1]>([-2])];
29
+ tensor<fp16, [1, 896, 1]> x_cast_fp16 = squeeze(axes = x_axes_0, x = var_41_cast_fp16_0)[name = string("x_cast_fp16")];
30
+ string var_55_pad_type_0 = const()[name = string("op_55_pad_type_0"), val = string("valid")];
31
+ tensor<int32, [1]> var_55_strides_0 = const()[name = string("op_55_strides_0"), val = tensor<int32, [1]>([1])];
32
+ tensor<int32, [2]> var_55_pad_0 = const()[name = string("op_55_pad_0"), val = tensor<int32, [2]>([0, 0])];
33
+ tensor<int32, [1]> var_55_dilations_0 = const()[name = string("op_55_dilations_0"), val = tensor<int32, [1]>([1])];
34
+ int32 var_55_groups_0 = const()[name = string("op_55_groups_0"), val = int32(1)];
35
+ tensor<fp16, [896, 896, 1]> var_49_to_fp16 = const()[name = string("op_49_to_fp16"), val = tensor<fp16, [896, 896, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(120320)))];
36
+ tensor<fp16, [896]> fc2_bias_to_fp16 = const()[name = string("fc2_bias_to_fp16"), val = tensor<fp16, [896]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1726016)))];
37
+ tensor<fp16, [1, 896, 1]> output = conv(bias = fc2_bias_to_fp16, dilations = var_55_dilations_0, groups = var_55_groups_0, pad = var_55_pad_0, pad_type = var_55_pad_type_0, strides = var_55_strides_0, weight = var_49_to_fp16, x = x_cast_fp16)[name = string("op_55_cast_fp16")];
38
+ } -> (output);
39
+ }
acoustic_connector.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c32f4b059cd82c4eac2af379ff3686ca657f13c8a5b96809719351ab6f1ebb35
3
+ size 1727872
acoustic_connector.mlpackage/Data/com.apple.CoreML/model.mlmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:037d6874f5deca07a926cf619475199d686a8ca143ec6252a215a65ac56e8957
3
+ size 5151
acoustic_connector.mlpackage/Data/com.apple.CoreML/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c32f4b059cd82c4eac2af379ff3686ca657f13c8a5b96809719351ab6f1ebb35
3
+ size 1727872
acoustic_connector.mlpackage/Manifest.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "fileFormatVersion": "1.0.0",
3
+ "itemInfoEntries": {
4
+ "966eb60d-af76-4b59-ab8e-302dfe8a26ac": {
5
+ "author": "com.apple.CoreML",
6
+ "description": "CoreML Model Weights",
7
+ "name": "weights",
8
+ "path": "com.apple.CoreML/weights"
9
+ },
10
+ "c3a8450e-db98-4818-b49d-61a61abf029a": {
11
+ "author": "com.apple.CoreML",
12
+ "description": "CoreML Model Specification",
13
+ "name": "model.mlmodel",
14
+ "path": "com.apple.CoreML/model.mlmodel"
15
+ }
16
+ },
17
+ "rootModelIdentifier": "c3a8450e-db98-4818-b49d-61a61abf029a"
18
+ }
decoder_coreml_12_ne.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b968c0e7f610fbc5fac13cf6044682ba49b62b2f24021430ab81379dd5246080
3
+ size 243
decoder_coreml_12_ne.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f76e909cf4859b460963ef38d863a85005db09d3def7305a5ac17ce56cc718ca
3
+ size 738
decoder_coreml_12_ne.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
decoder_coreml_12_ne.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2db7f6ae03e24acf862b5f21c73a8ce5c1932d3a75423766141457a21db2a74f
3
+ size 687406272
diffusion_head_model.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a61f48b53134e0a5bbefb076fbe166269691a90a00faef7841288f725d0caf
3
+ size 243
diffusion_head_model.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab6d23550b62e579a3d527e42a8f40f955ad7c173ffb3f0a7e11ea8e04b2f151
3
+ size 443
diffusion_head_model.mlmodelc/model.mil ADDED
@@ -0,0 +1,285 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3405.2.1"}, {"coremlc-version", "3405.2.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})]
3
+ {
4
+ func main<ios18>(tensor<fp16, [2, 896, 1, 1]> condition, tensor<fp16, [2, 64, 1, 1]> noisy_images, tensor<fp16, [256, 1, 1]> timesteps) {
5
+ string x_5_pad_type_0 = const()[name = string("x_5_pad_type_0"), val = string("valid")];
6
+ tensor<int32, [2]> x_5_strides_0 = const()[name = string("x_5_strides_0"), val = tensor<int32, [2]>([1, 1])];
7
+ tensor<int32, [4]> x_5_pad_0 = const()[name = string("x_5_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
8
+ tensor<int32, [2]> x_5_dilations_0 = const()[name = string("x_5_dilations_0"), val = tensor<int32, [2]>([1, 1])];
9
+ int32 x_5_groups_0 = const()[name = string("x_5_groups_0"), val = int32(1)];
10
+ tensor<fp16, [896, 64, 1, 1]> var_21_to_fp16 = const()[name = string("op_21_to_fp16"), val = tensor<fp16, [896, 64, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
11
+ tensor<fp16, [2, 896, 1, 1]> x_5_cast_fp16 = conv(dilations = x_5_dilations_0, groups = x_5_groups_0, pad = x_5_pad_0, pad_type = x_5_pad_type_0, strides = x_5_strides_0, weight = var_21_to_fp16, x = noisy_images)[name = string("x_5_cast_fp16")];
12
+ tensor<int32, [1]> var_34_axes_0 = const()[name = string("op_34_axes_0"), val = tensor<int32, [1]>([0])];
13
+ tensor<fp16, [1, 256, 1, 1]> var_34_cast_fp16 = expand_dims(axes = var_34_axes_0, x = timesteps)[name = string("op_34_cast_fp16")];
14
+ string var_39_pad_type_0 = const()[name = string("op_39_pad_type_0"), val = string("valid")];
15
+ tensor<int32, [2]> var_39_strides_0 = const()[name = string("op_39_strides_0"), val = tensor<int32, [2]>([1, 1])];
16
+ tensor<int32, [4]> var_39_pad_0 = const()[name = string("op_39_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
17
+ tensor<int32, [2]> var_39_dilations_0 = const()[name = string("op_39_dilations_0"), val = tensor<int32, [2]>([1, 1])];
18
+ int32 var_39_groups_0 = const()[name = string("op_39_groups_0"), val = int32(1)];
19
+ tensor<fp16, [896, 256, 1, 1]> var_29_to_fp16 = const()[name = string("op_29_to_fp16"), val = tensor<fp16, [896, 256, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(114816)))];
20
+ tensor<fp16, [1, 896, 1, 1]> var_39_cast_fp16 = conv(dilations = var_39_dilations_0, groups = var_39_groups_0, pad = var_39_pad_0, pad_type = var_39_pad_type_0, strides = var_39_strides_0, weight = var_29_to_fp16, x = var_34_cast_fp16)[name = string("op_39_cast_fp16")];
21
+ tensor<int32, [1]> input_1_axes_0 = const()[name = string("input_1_axes_0"), val = tensor<int32, [1]>([0])];
22
+ tensor<fp16, [896, 1, 1]> input_1_cast_fp16 = squeeze(axes = input_1_axes_0, x = var_39_cast_fp16)[name = string("input_1_cast_fp16")];
23
+ tensor<fp16, [896, 1, 1]> x_1_cast_fp16 = silu(x = input_1_cast_fp16)[name = string("x_1_cast_fp16")];
24
+ tensor<int32, [1]> var_42_axes_0 = const()[name = string("op_42_axes_0"), val = tensor<int32, [1]>([0])];
25
+ tensor<fp16, [1, 896, 1, 1]> var_42_cast_fp16 = expand_dims(axes = var_42_axes_0, x = x_1_cast_fp16)[name = string("op_42_cast_fp16")];
26
+ string var_47_pad_type_0 = const()[name = string("op_47_pad_type_0"), val = string("valid")];
27
+ tensor<int32, [2]> var_47_strides_0 = const()[name = string("op_47_strides_0"), val = tensor<int32, [2]>([1, 1])];
28
+ tensor<int32, [4]> var_47_pad_0 = const()[name = string("op_47_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
29
+ tensor<int32, [2]> var_47_dilations_0 = const()[name = string("op_47_dilations_0"), val = tensor<int32, [2]>([1, 1])];
30
+ int32 var_47_groups_0 = const()[name = string("op_47_groups_0"), val = int32(1)];
31
+ tensor<fp16, [896, 896, 1, 1]> var_27_to_fp16 = const()[name = string("op_27_to_fp16"), val = tensor<fp16, [896, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(573632)))];
32
+ tensor<fp16, [1, 896, 1, 1]> var_47_cast_fp16 = conv(dilations = var_47_dilations_0, groups = var_47_groups_0, pad = var_47_pad_0, pad_type = var_47_pad_type_0, strides = var_47_strides_0, weight = var_27_to_fp16, x = var_42_cast_fp16)[name = string("op_47_cast_fp16")];
33
+ tensor<int32, [1]> t_axes_0 = const()[name = string("t_axes_0"), val = tensor<int32, [1]>([0])];
34
+ tensor<fp16, [896, 1, 1]> t_cast_fp16 = squeeze(axes = t_axes_0, x = var_47_cast_fp16)[name = string("t_cast_fp16")];
35
+ string condition_pad_type_0 = const()[name = string("condition_pad_type_0"), val = string("valid")];
36
+ tensor<int32, [2]> condition_strides_0 = const()[name = string("condition_strides_0"), val = tensor<int32, [2]>([1, 1])];
37
+ tensor<int32, [4]> condition_pad_0 = const()[name = string("condition_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
38
+ tensor<int32, [2]> condition_dilations_0 = const()[name = string("condition_dilations_0"), val = tensor<int32, [2]>([1, 1])];
39
+ int32 condition_groups_0 = const()[name = string("condition_groups_0"), val = int32(1)];
40
+ tensor<fp16, [896, 896, 1, 1]> var_54_to_fp16 = const()[name = string("op_54_to_fp16"), val = tensor<fp16, [896, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2179328)))];
41
+ tensor<fp16, [2, 896, 1, 1]> condition_cast_fp16 = conv(dilations = condition_dilations_0, groups = condition_groups_0, pad = condition_pad_0, pad_type = condition_pad_type_0, strides = condition_strides_0, weight = var_54_to_fp16, x = condition)[name = string("condition_cast_fp16")];
42
+ tensor<fp16, [2, 896, 1, 1]> input_3_cast_fp16 = add(x = condition_cast_fp16, y = t_cast_fp16)[name = string("input_3_cast_fp16")];
43
+ int32 var_69 = const()[name = string("op_69"), val = int32(1)];
44
+ tensor<fp16, [2, 896, 1, 1]> x_3_cast_fp16 = silu(x = input_3_cast_fp16)[name = string("x_3_cast_fp16")];
45
+ string var_79_pad_type_0 = const()[name = string("op_79_pad_type_0"), val = string("valid")];
46
+ tensor<int32, [2]> var_79_strides_0 = const()[name = string("op_79_strides_0"), val = tensor<int32, [2]>([1, 1])];
47
+ tensor<int32, [4]> var_79_pad_0 = const()[name = string("op_79_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
48
+ tensor<int32, [2]> var_79_dilations_0 = const()[name = string("op_79_dilations_0"), val = tensor<int32, [2]>([1, 1])];
49
+ int32 var_79_groups_0 = const()[name = string("op_79_groups_0"), val = int32(1)];
50
+ tensor<fp16, [2688, 896, 1, 1]> var_71_to_fp16 = const()[name = string("op_71_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3785024)))];
51
+ tensor<fp16, [2, 2688, 1, 1]> var_79_cast_fp16 = conv(dilations = var_79_dilations_0, groups = var_79_groups_0, pad = var_79_pad_0, pad_type = var_79_pad_type_0, strides = var_79_strides_0, weight = var_71_to_fp16, x = x_3_cast_fp16)[name = string("op_79_cast_fp16")];
52
+ tensor<int32, [3]> var_80_split_sizes_0 = const()[name = string("op_80_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
53
+ int32 var_80_axis_0 = const()[name = string("op_80_axis_0"), val = int32(1)];
54
+ tensor<fp16, [2, 896, 1, 1]> var_80_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_80_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_80_cast_fp16_2 = split(axis = var_80_axis_0, split_sizes = var_80_split_sizes_0, x = var_79_cast_fp16)[name = string("op_80_cast_fp16")];
55
+ fp16 const_0_promoted_to_fp16 = const()[name = string("const_0_promoted_to_fp16"), val = fp16(-0x1p+0)];
56
+ tensor<fp16, [2, 896, 1, 1]> var_85_cast_fp16 = mul(x = x_5_cast_fp16, y = const_0_promoted_to_fp16)[name = string("op_85_cast_fp16")];
57
+ bool x_7_interleave_0 = const()[name = string("x_7_interleave_0"), val = bool(false)];
58
+ tensor<fp16, [2, 1792, 1, 1]> x_7_cast_fp16 = concat(axis = var_69, interleave = x_7_interleave_0, values = (x_5_cast_fp16, var_85_cast_fp16))[name = string("x_7_cast_fp16")];
59
+ tensor<int32, [1]> out_1_axes_0 = const()[name = string("out_1_axes_0"), val = tensor<int32, [1]>([1])];
60
+ fp16 var_95_to_fp16 = const()[name = string("op_95_to_fp16"), val = fp16(0x1.5p-17)];
61
+ tensor<fp16, [2, 1792, 1, 1]> out_1_cast_fp16 = layer_norm(axes = out_1_axes_0, epsilon = var_95_to_fp16, x = x_7_cast_fp16)[name = string("out_1_cast_fp16")];
62
+ tensor<fp16, [1, 1792, 1, 1]> layer_layers_0_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_0_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(8601984)))];
63
+ tensor<fp16, [2, 1792, 1, 1]> out_3_cast_fp16 = mul(x = out_1_cast_fp16, y = layer_layers_0_layer_norm_weight_to_fp16)[name = string("out_3_cast_fp16")];
64
+ tensor<int32, [2]> var_101_split_sizes_0 = const()[name = string("op_101_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
65
+ int32 var_101_axis_0 = const()[name = string("op_101_axis_0"), val = int32(1)];
66
+ tensor<fp16, [2, 896, 1, 1]> var_101_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_101_cast_fp16_1 = split(axis = var_101_axis_0, split_sizes = var_101_split_sizes_0, x = out_3_cast_fp16)[name = string("op_101_cast_fp16")];
67
+ fp16 var_104_promoted_to_fp16 = const()[name = string("op_104_promoted_to_fp16"), val = fp16(0x1p+0)];
68
+ tensor<fp16, [2, 896, 1, 1]> var_105_cast_fp16 = add(x = var_80_cast_fp16_1, y = var_104_promoted_to_fp16)[name = string("op_105_cast_fp16")];
69
+ tensor<fp16, [2, 896, 1, 1]> var_106_cast_fp16 = mul(x = var_101_cast_fp16_0, y = var_105_cast_fp16)[name = string("op_106_cast_fp16")];
70
+ tensor<fp16, [2, 896, 1, 1]> x_13_cast_fp16 = add(x = var_106_cast_fp16, y = var_80_cast_fp16_0)[name = string("x_13_cast_fp16")];
71
+ string input_5_pad_type_0 = const()[name = string("input_5_pad_type_0"), val = string("valid")];
72
+ tensor<int32, [2]> input_5_strides_0 = const()[name = string("input_5_strides_0"), val = tensor<int32, [2]>([1, 1])];
73
+ tensor<int32, [4]> input_5_pad_0 = const()[name = string("input_5_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
74
+ tensor<int32, [2]> input_5_dilations_0 = const()[name = string("input_5_dilations_0"), val = tensor<int32, [2]>([1, 1])];
75
+ int32 input_5_groups_0 = const()[name = string("input_5_groups_0"), val = int32(1)];
76
+ tensor<fp16, [2688, 896, 1, 1]> var_62_to_fp16 = const()[name = string("op_62_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(8605632)))];
77
+ tensor<fp16, [2, 2688, 1, 1]> input_5_cast_fp16 = conv(dilations = input_5_dilations_0, groups = input_5_groups_0, pad = input_5_pad_0, pad_type = input_5_pad_type_0, strides = input_5_strides_0, weight = var_62_to_fp16, x = x_13_cast_fp16)[name = string("input_5_cast_fp16")];
78
+ string up_1_pad_type_0 = const()[name = string("up_1_pad_type_0"), val = string("valid")];
79
+ tensor<int32, [2]> up_1_strides_0 = const()[name = string("up_1_strides_0"), val = tensor<int32, [2]>([1, 1])];
80
+ tensor<int32, [4]> up_1_pad_0 = const()[name = string("up_1_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
81
+ tensor<int32, [2]> up_1_dilations_0 = const()[name = string("up_1_dilations_0"), val = tensor<int32, [2]>([1, 1])];
82
+ int32 up_1_groups_0 = const()[name = string("up_1_groups_0"), val = int32(1)];
83
+ tensor<fp16, [2688, 896, 1, 1]> var_63_to_fp16 = const()[name = string("op_63_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(13422592)))];
84
+ tensor<fp16, [2, 2688, 1, 1]> up_1_cast_fp16 = conv(dilations = up_1_dilations_0, groups = up_1_groups_0, pad = up_1_pad_0, pad_type = up_1_pad_type_0, strides = up_1_strides_0, weight = var_63_to_fp16, x = x_13_cast_fp16)[name = string("up_1_cast_fp16")];
85
+ tensor<fp16, [2, 2688, 1, 1]> gate_1_cast_fp16 = silu(x = input_5_cast_fp16)[name = string("gate_1_cast_fp16")];
86
+ tensor<fp16, [2, 2688, 1, 1]> x_15_cast_fp16 = mul(x = gate_1_cast_fp16, y = up_1_cast_fp16)[name = string("x_15_cast_fp16")];
87
+ string var_124_pad_type_0 = const()[name = string("op_124_pad_type_0"), val = string("valid")];
88
+ tensor<int32, [2]> var_124_strides_0 = const()[name = string("op_124_strides_0"), val = tensor<int32, [2]>([1, 1])];
89
+ tensor<int32, [4]> var_124_pad_0 = const()[name = string("op_124_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
90
+ tensor<int32, [2]> var_124_dilations_0 = const()[name = string("op_124_dilations_0"), val = tensor<int32, [2]>([1, 1])];
91
+ int32 var_124_groups_0 = const()[name = string("op_124_groups_0"), val = int32(1)];
92
+ tensor<fp16, [896, 2688, 1, 1]> var_64_to_fp16 = const()[name = string("op_64_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(18239552)))];
93
+ tensor<fp16, [2, 896, 1, 1]> var_124_cast_fp16 = conv(dilations = var_124_dilations_0, groups = var_124_groups_0, pad = var_124_pad_0, pad_type = var_124_pad_type_0, strides = var_124_strides_0, weight = var_64_to_fp16, x = x_15_cast_fp16)[name = string("op_124_cast_fp16")];
94
+ tensor<fp16, [2, 896, 1, 1]> var_125_cast_fp16 = mul(x = var_80_cast_fp16_2, y = var_124_cast_fp16)[name = string("op_125_cast_fp16")];
95
+ tensor<fp16, [2, 896, 1, 1]> x_19_cast_fp16 = add(x = x_5_cast_fp16, y = var_125_cast_fp16)[name = string("x_19_cast_fp16")];
96
+ int32 var_134 = const()[name = string("op_134"), val = int32(1)];
97
+ string var_144_pad_type_0 = const()[name = string("op_144_pad_type_0"), val = string("valid")];
98
+ tensor<int32, [2]> var_144_strides_0 = const()[name = string("op_144_strides_0"), val = tensor<int32, [2]>([1, 1])];
99
+ tensor<int32, [4]> var_144_pad_0 = const()[name = string("op_144_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
100
+ tensor<int32, [2]> var_144_dilations_0 = const()[name = string("op_144_dilations_0"), val = tensor<int32, [2]>([1, 1])];
101
+ int32 var_144_groups_0 = const()[name = string("op_144_groups_0"), val = int32(1)];
102
+ tensor<fp16, [2688, 896, 1, 1]> var_136_to_fp16 = const()[name = string("op_136_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(23056512)))];
103
+ tensor<fp16, [2, 2688, 1, 1]> var_144_cast_fp16 = conv(dilations = var_144_dilations_0, groups = var_144_groups_0, pad = var_144_pad_0, pad_type = var_144_pad_type_0, strides = var_144_strides_0, weight = var_136_to_fp16, x = x_3_cast_fp16)[name = string("op_144_cast_fp16")];
104
+ tensor<int32, [3]> var_145_split_sizes_0 = const()[name = string("op_145_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
105
+ int32 var_145_axis_0 = const()[name = string("op_145_axis_0"), val = int32(1)];
106
+ tensor<fp16, [2, 896, 1, 1]> var_145_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_145_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_145_cast_fp16_2 = split(axis = var_145_axis_0, split_sizes = var_145_split_sizes_0, x = var_144_cast_fp16)[name = string("op_145_cast_fp16")];
107
+ fp16 const_1_promoted_to_fp16 = const()[name = string("const_1_promoted_to_fp16"), val = fp16(-0x1p+0)];
108
+ tensor<fp16, [2, 896, 1, 1]> var_150_cast_fp16 = mul(x = x_19_cast_fp16, y = const_1_promoted_to_fp16)[name = string("op_150_cast_fp16")];
109
+ bool x_21_interleave_0 = const()[name = string("x_21_interleave_0"), val = bool(false)];
110
+ tensor<fp16, [2, 1792, 1, 1]> x_21_cast_fp16 = concat(axis = var_134, interleave = x_21_interleave_0, values = (x_19_cast_fp16, var_150_cast_fp16))[name = string("x_21_cast_fp16")];
111
+ tensor<int32, [1]> out_7_axes_0 = const()[name = string("out_7_axes_0"), val = tensor<int32, [1]>([1])];
112
+ fp16 var_160_to_fp16 = const()[name = string("op_160_to_fp16"), val = fp16(0x1.5p-17)];
113
+ tensor<fp16, [2, 1792, 1, 1]> out_7_cast_fp16 = layer_norm(axes = out_7_axes_0, epsilon = var_160_to_fp16, x = x_21_cast_fp16)[name = string("out_7_cast_fp16")];
114
+ tensor<fp16, [1, 1792, 1, 1]> layer_layers_1_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_1_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(27873472)))];
115
+ tensor<fp16, [2, 1792, 1, 1]> out_9_cast_fp16 = mul(x = out_7_cast_fp16, y = layer_layers_1_layer_norm_weight_to_fp16)[name = string("out_9_cast_fp16")];
116
+ tensor<int32, [2]> var_166_split_sizes_0 = const()[name = string("op_166_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
117
+ int32 var_166_axis_0 = const()[name = string("op_166_axis_0"), val = int32(1)];
118
+ tensor<fp16, [2, 896, 1, 1]> var_166_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_166_cast_fp16_1 = split(axis = var_166_axis_0, split_sizes = var_166_split_sizes_0, x = out_9_cast_fp16)[name = string("op_166_cast_fp16")];
119
+ fp16 var_169_promoted_to_fp16 = const()[name = string("op_169_promoted_to_fp16"), val = fp16(0x1p+0)];
120
+ tensor<fp16, [2, 896, 1, 1]> var_170_cast_fp16 = add(x = var_145_cast_fp16_1, y = var_169_promoted_to_fp16)[name = string("op_170_cast_fp16")];
121
+ tensor<fp16, [2, 896, 1, 1]> var_171_cast_fp16 = mul(x = var_166_cast_fp16_0, y = var_170_cast_fp16)[name = string("op_171_cast_fp16")];
122
+ tensor<fp16, [2, 896, 1, 1]> x_27_cast_fp16 = add(x = var_171_cast_fp16, y = var_145_cast_fp16_0)[name = string("x_27_cast_fp16")];
123
+ string input_7_pad_type_0 = const()[name = string("input_7_pad_type_0"), val = string("valid")];
124
+ tensor<int32, [2]> input_7_strides_0 = const()[name = string("input_7_strides_0"), val = tensor<int32, [2]>([1, 1])];
125
+ tensor<int32, [4]> input_7_pad_0 = const()[name = string("input_7_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
126
+ tensor<int32, [2]> input_7_dilations_0 = const()[name = string("input_7_dilations_0"), val = tensor<int32, [2]>([1, 1])];
127
+ int32 input_7_groups_0 = const()[name = string("input_7_groups_0"), val = int32(1)];
128
+ tensor<fp16, [2688, 896, 1, 1]> var_127_to_fp16 = const()[name = string("op_127_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(27877120)))];
129
+ tensor<fp16, [2, 2688, 1, 1]> input_7_cast_fp16 = conv(dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = var_127_to_fp16, x = x_27_cast_fp16)[name = string("input_7_cast_fp16")];
130
+ string up_3_pad_type_0 = const()[name = string("up_3_pad_type_0"), val = string("valid")];
131
+ tensor<int32, [2]> up_3_strides_0 = const()[name = string("up_3_strides_0"), val = tensor<int32, [2]>([1, 1])];
132
+ tensor<int32, [4]> up_3_pad_0 = const()[name = string("up_3_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
133
+ tensor<int32, [2]> up_3_dilations_0 = const()[name = string("up_3_dilations_0"), val = tensor<int32, [2]>([1, 1])];
134
+ int32 up_3_groups_0 = const()[name = string("up_3_groups_0"), val = int32(1)];
135
+ tensor<fp16, [2688, 896, 1, 1]> var_128_to_fp16 = const()[name = string("op_128_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(32694080)))];
136
+ tensor<fp16, [2, 2688, 1, 1]> up_3_cast_fp16 = conv(dilations = up_3_dilations_0, groups = up_3_groups_0, pad = up_3_pad_0, pad_type = up_3_pad_type_0, strides = up_3_strides_0, weight = var_128_to_fp16, x = x_27_cast_fp16)[name = string("up_3_cast_fp16")];
137
+ tensor<fp16, [2, 2688, 1, 1]> gate_3_cast_fp16 = silu(x = input_7_cast_fp16)[name = string("gate_3_cast_fp16")];
138
+ tensor<fp16, [2, 2688, 1, 1]> x_29_cast_fp16 = mul(x = gate_3_cast_fp16, y = up_3_cast_fp16)[name = string("x_29_cast_fp16")];
139
+ string var_189_pad_type_0 = const()[name = string("op_189_pad_type_0"), val = string("valid")];
140
+ tensor<int32, [2]> var_189_strides_0 = const()[name = string("op_189_strides_0"), val = tensor<int32, [2]>([1, 1])];
141
+ tensor<int32, [4]> var_189_pad_0 = const()[name = string("op_189_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
142
+ tensor<int32, [2]> var_189_dilations_0 = const()[name = string("op_189_dilations_0"), val = tensor<int32, [2]>([1, 1])];
143
+ int32 var_189_groups_0 = const()[name = string("op_189_groups_0"), val = int32(1)];
144
+ tensor<fp16, [896, 2688, 1, 1]> var_129_to_fp16 = const()[name = string("op_129_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(37511040)))];
145
+ tensor<fp16, [2, 896, 1, 1]> var_189_cast_fp16 = conv(dilations = var_189_dilations_0, groups = var_189_groups_0, pad = var_189_pad_0, pad_type = var_189_pad_type_0, strides = var_189_strides_0, weight = var_129_to_fp16, x = x_29_cast_fp16)[name = string("op_189_cast_fp16")];
146
+ tensor<fp16, [2, 896, 1, 1]> var_190_cast_fp16 = mul(x = var_145_cast_fp16_2, y = var_189_cast_fp16)[name = string("op_190_cast_fp16")];
147
+ tensor<fp16, [2, 896, 1, 1]> x_33_cast_fp16 = add(x = x_19_cast_fp16, y = var_190_cast_fp16)[name = string("x_33_cast_fp16")];
148
+ int32 var_199 = const()[name = string("op_199"), val = int32(1)];
149
+ string var_209_pad_type_0 = const()[name = string("op_209_pad_type_0"), val = string("valid")];
150
+ tensor<int32, [2]> var_209_strides_0 = const()[name = string("op_209_strides_0"), val = tensor<int32, [2]>([1, 1])];
151
+ tensor<int32, [4]> var_209_pad_0 = const()[name = string("op_209_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
152
+ tensor<int32, [2]> var_209_dilations_0 = const()[name = string("op_209_dilations_0"), val = tensor<int32, [2]>([1, 1])];
153
+ int32 var_209_groups_0 = const()[name = string("op_209_groups_0"), val = int32(1)];
154
+ tensor<fp16, [2688, 896, 1, 1]> var_201_to_fp16 = const()[name = string("op_201_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(42328000)))];
155
+ tensor<fp16, [2, 2688, 1, 1]> var_209_cast_fp16 = conv(dilations = var_209_dilations_0, groups = var_209_groups_0, pad = var_209_pad_0, pad_type = var_209_pad_type_0, strides = var_209_strides_0, weight = var_201_to_fp16, x = x_3_cast_fp16)[name = string("op_209_cast_fp16")];
156
+ tensor<int32, [3]> var_210_split_sizes_0 = const()[name = string("op_210_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
157
+ int32 var_210_axis_0 = const()[name = string("op_210_axis_0"), val = int32(1)];
158
+ tensor<fp16, [2, 896, 1, 1]> var_210_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_210_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_210_cast_fp16_2 = split(axis = var_210_axis_0, split_sizes = var_210_split_sizes_0, x = var_209_cast_fp16)[name = string("op_210_cast_fp16")];
159
+ fp16 const_2_promoted_to_fp16 = const()[name = string("const_2_promoted_to_fp16"), val = fp16(-0x1p+0)];
160
+ tensor<fp16, [2, 896, 1, 1]> var_215_cast_fp16 = mul(x = x_33_cast_fp16, y = const_2_promoted_to_fp16)[name = string("op_215_cast_fp16")];
161
+ bool x_35_interleave_0 = const()[name = string("x_35_interleave_0"), val = bool(false)];
162
+ tensor<fp16, [2, 1792, 1, 1]> x_35_cast_fp16 = concat(axis = var_199, interleave = x_35_interleave_0, values = (x_33_cast_fp16, var_215_cast_fp16))[name = string("x_35_cast_fp16")];
163
+ tensor<int32, [1]> out_13_axes_0 = const()[name = string("out_13_axes_0"), val = tensor<int32, [1]>([1])];
164
+ fp16 var_225_to_fp16 = const()[name = string("op_225_to_fp16"), val = fp16(0x1.5p-17)];
165
+ tensor<fp16, [2, 1792, 1, 1]> out_13_cast_fp16 = layer_norm(axes = out_13_axes_0, epsilon = var_225_to_fp16, x = x_35_cast_fp16)[name = string("out_13_cast_fp16")];
166
+ tensor<fp16, [1, 1792, 1, 1]> layer_layers_2_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_2_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(47144960)))];
167
+ tensor<fp16, [2, 1792, 1, 1]> out_15_cast_fp16 = mul(x = out_13_cast_fp16, y = layer_layers_2_layer_norm_weight_to_fp16)[name = string("out_15_cast_fp16")];
168
+ tensor<int32, [2]> var_231_split_sizes_0 = const()[name = string("op_231_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
169
+ int32 var_231_axis_0 = const()[name = string("op_231_axis_0"), val = int32(1)];
170
+ tensor<fp16, [2, 896, 1, 1]> var_231_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_231_cast_fp16_1 = split(axis = var_231_axis_0, split_sizes = var_231_split_sizes_0, x = out_15_cast_fp16)[name = string("op_231_cast_fp16")];
171
+ fp16 var_234_promoted_to_fp16 = const()[name = string("op_234_promoted_to_fp16"), val = fp16(0x1p+0)];
172
+ tensor<fp16, [2, 896, 1, 1]> var_235_cast_fp16 = add(x = var_210_cast_fp16_1, y = var_234_promoted_to_fp16)[name = string("op_235_cast_fp16")];
173
+ tensor<fp16, [2, 896, 1, 1]> var_236_cast_fp16 = mul(x = var_231_cast_fp16_0, y = var_235_cast_fp16)[name = string("op_236_cast_fp16")];
174
+ tensor<fp16, [2, 896, 1, 1]> x_41_cast_fp16 = add(x = var_236_cast_fp16, y = var_210_cast_fp16_0)[name = string("x_41_cast_fp16")];
175
+ string input_9_pad_type_0 = const()[name = string("input_9_pad_type_0"), val = string("valid")];
176
+ tensor<int32, [2]> input_9_strides_0 = const()[name = string("input_9_strides_0"), val = tensor<int32, [2]>([1, 1])];
177
+ tensor<int32, [4]> input_9_pad_0 = const()[name = string("input_9_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
178
+ tensor<int32, [2]> input_9_dilations_0 = const()[name = string("input_9_dilations_0"), val = tensor<int32, [2]>([1, 1])];
179
+ int32 input_9_groups_0 = const()[name = string("input_9_groups_0"), val = int32(1)];
180
+ tensor<fp16, [2688, 896, 1, 1]> var_192_to_fp16 = const()[name = string("op_192_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(47148608)))];
181
+ tensor<fp16, [2, 2688, 1, 1]> input_9_cast_fp16 = conv(dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = var_192_to_fp16, x = x_41_cast_fp16)[name = string("input_9_cast_fp16")];
182
+ string up_5_pad_type_0 = const()[name = string("up_5_pad_type_0"), val = string("valid")];
183
+ tensor<int32, [2]> up_5_strides_0 = const()[name = string("up_5_strides_0"), val = tensor<int32, [2]>([1, 1])];
184
+ tensor<int32, [4]> up_5_pad_0 = const()[name = string("up_5_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
185
+ tensor<int32, [2]> up_5_dilations_0 = const()[name = string("up_5_dilations_0"), val = tensor<int32, [2]>([1, 1])];
186
+ int32 up_5_groups_0 = const()[name = string("up_5_groups_0"), val = int32(1)];
187
+ tensor<fp16, [2688, 896, 1, 1]> var_193_to_fp16 = const()[name = string("op_193_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(51965568)))];
188
+ tensor<fp16, [2, 2688, 1, 1]> up_5_cast_fp16 = conv(dilations = up_5_dilations_0, groups = up_5_groups_0, pad = up_5_pad_0, pad_type = up_5_pad_type_0, strides = up_5_strides_0, weight = var_193_to_fp16, x = x_41_cast_fp16)[name = string("up_5_cast_fp16")];
189
+ tensor<fp16, [2, 2688, 1, 1]> gate_5_cast_fp16 = silu(x = input_9_cast_fp16)[name = string("gate_5_cast_fp16")];
190
+ tensor<fp16, [2, 2688, 1, 1]> x_43_cast_fp16 = mul(x = gate_5_cast_fp16, y = up_5_cast_fp16)[name = string("x_43_cast_fp16")];
191
+ string var_254_pad_type_0 = const()[name = string("op_254_pad_type_0"), val = string("valid")];
192
+ tensor<int32, [2]> var_254_strides_0 = const()[name = string("op_254_strides_0"), val = tensor<int32, [2]>([1, 1])];
193
+ tensor<int32, [4]> var_254_pad_0 = const()[name = string("op_254_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
194
+ tensor<int32, [2]> var_254_dilations_0 = const()[name = string("op_254_dilations_0"), val = tensor<int32, [2]>([1, 1])];
195
+ int32 var_254_groups_0 = const()[name = string("op_254_groups_0"), val = int32(1)];
196
+ tensor<fp16, [896, 2688, 1, 1]> var_194_to_fp16 = const()[name = string("op_194_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(56782528)))];
197
+ tensor<fp16, [2, 896, 1, 1]> var_254_cast_fp16 = conv(dilations = var_254_dilations_0, groups = var_254_groups_0, pad = var_254_pad_0, pad_type = var_254_pad_type_0, strides = var_254_strides_0, weight = var_194_to_fp16, x = x_43_cast_fp16)[name = string("op_254_cast_fp16")];
198
+ tensor<fp16, [2, 896, 1, 1]> var_255_cast_fp16 = mul(x = var_210_cast_fp16_2, y = var_254_cast_fp16)[name = string("op_255_cast_fp16")];
199
+ tensor<fp16, [2, 896, 1, 1]> x_47_cast_fp16 = add(x = x_33_cast_fp16, y = var_255_cast_fp16)[name = string("x_47_cast_fp16")];
200
+ int32 var_264 = const()[name = string("op_264"), val = int32(1)];
201
+ string var_274_pad_type_0 = const()[name = string("op_274_pad_type_0"), val = string("valid")];
202
+ tensor<int32, [2]> var_274_strides_0 = const()[name = string("op_274_strides_0"), val = tensor<int32, [2]>([1, 1])];
203
+ tensor<int32, [4]> var_274_pad_0 = const()[name = string("op_274_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
204
+ tensor<int32, [2]> var_274_dilations_0 = const()[name = string("op_274_dilations_0"), val = tensor<int32, [2]>([1, 1])];
205
+ int32 var_274_groups_0 = const()[name = string("op_274_groups_0"), val = int32(1)];
206
+ tensor<fp16, [2688, 896, 1, 1]> var_266_to_fp16 = const()[name = string("op_266_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(61599488)))];
207
+ tensor<fp16, [2, 2688, 1, 1]> var_274_cast_fp16 = conv(dilations = var_274_dilations_0, groups = var_274_groups_0, pad = var_274_pad_0, pad_type = var_274_pad_type_0, strides = var_274_strides_0, weight = var_266_to_fp16, x = x_3_cast_fp16)[name = string("op_274_cast_fp16")];
208
+ tensor<int32, [3]> var_275_split_sizes_0 = const()[name = string("op_275_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
209
+ int32 var_275_axis_0 = const()[name = string("op_275_axis_0"), val = int32(1)];
210
+ tensor<fp16, [2, 896, 1, 1]> var_275_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_275_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_275_cast_fp16_2 = split(axis = var_275_axis_0, split_sizes = var_275_split_sizes_0, x = var_274_cast_fp16)[name = string("op_275_cast_fp16")];
211
+ fp16 const_3_promoted_to_fp16 = const()[name = string("const_3_promoted_to_fp16"), val = fp16(-0x1p+0)];
212
+ tensor<fp16, [2, 896, 1, 1]> var_280_cast_fp16 = mul(x = x_47_cast_fp16, y = const_3_promoted_to_fp16)[name = string("op_280_cast_fp16")];
213
+ bool x_49_interleave_0 = const()[name = string("x_49_interleave_0"), val = bool(false)];
214
+ tensor<fp16, [2, 1792, 1, 1]> x_49_cast_fp16 = concat(axis = var_264, interleave = x_49_interleave_0, values = (x_47_cast_fp16, var_280_cast_fp16))[name = string("x_49_cast_fp16")];
215
+ tensor<int32, [1]> out_19_axes_0 = const()[name = string("out_19_axes_0"), val = tensor<int32, [1]>([1])];
216
+ fp16 var_290_to_fp16 = const()[name = string("op_290_to_fp16"), val = fp16(0x1.5p-17)];
217
+ tensor<fp16, [2, 1792, 1, 1]> out_19_cast_fp16 = layer_norm(axes = out_19_axes_0, epsilon = var_290_to_fp16, x = x_49_cast_fp16)[name = string("out_19_cast_fp16")];
218
+ tensor<fp16, [1, 1792, 1, 1]> layer_layers_3_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_3_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(66416448)))];
219
+ tensor<fp16, [2, 1792, 1, 1]> out_21_cast_fp16 = mul(x = out_19_cast_fp16, y = layer_layers_3_layer_norm_weight_to_fp16)[name = string("out_21_cast_fp16")];
220
+ tensor<int32, [2]> var_296_split_sizes_0 = const()[name = string("op_296_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
221
+ int32 var_296_axis_0 = const()[name = string("op_296_axis_0"), val = int32(1)];
222
+ tensor<fp16, [2, 896, 1, 1]> var_296_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_296_cast_fp16_1 = split(axis = var_296_axis_0, split_sizes = var_296_split_sizes_0, x = out_21_cast_fp16)[name = string("op_296_cast_fp16")];
223
+ fp16 var_299_promoted_to_fp16 = const()[name = string("op_299_promoted_to_fp16"), val = fp16(0x1p+0)];
224
+ tensor<fp16, [2, 896, 1, 1]> var_300_cast_fp16 = add(x = var_275_cast_fp16_1, y = var_299_promoted_to_fp16)[name = string("op_300_cast_fp16")];
225
+ tensor<fp16, [2, 896, 1, 1]> var_301_cast_fp16 = mul(x = var_296_cast_fp16_0, y = var_300_cast_fp16)[name = string("op_301_cast_fp16")];
226
+ tensor<fp16, [2, 896, 1, 1]> x_55_cast_fp16 = add(x = var_301_cast_fp16, y = var_275_cast_fp16_0)[name = string("x_55_cast_fp16")];
227
+ string input_pad_type_0 = const()[name = string("input_pad_type_0"), val = string("valid")];
228
+ tensor<int32, [2]> input_strides_0 = const()[name = string("input_strides_0"), val = tensor<int32, [2]>([1, 1])];
229
+ tensor<int32, [4]> input_pad_0 = const()[name = string("input_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
230
+ tensor<int32, [2]> input_dilations_0 = const()[name = string("input_dilations_0"), val = tensor<int32, [2]>([1, 1])];
231
+ int32 input_groups_0 = const()[name = string("input_groups_0"), val = int32(1)];
232
+ tensor<fp16, [2688, 896, 1, 1]> var_257_to_fp16 = const()[name = string("op_257_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(66420096)))];
233
+ tensor<fp16, [2, 2688, 1, 1]> input_cast_fp16 = conv(dilations = input_dilations_0, groups = input_groups_0, pad = input_pad_0, pad_type = input_pad_type_0, strides = input_strides_0, weight = var_257_to_fp16, x = x_55_cast_fp16)[name = string("input_cast_fp16")];
234
+ string up_pad_type_0 = const()[name = string("up_pad_type_0"), val = string("valid")];
235
+ tensor<int32, [2]> up_strides_0 = const()[name = string("up_strides_0"), val = tensor<int32, [2]>([1, 1])];
236
+ tensor<int32, [4]> up_pad_0 = const()[name = string("up_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
237
+ tensor<int32, [2]> up_dilations_0 = const()[name = string("up_dilations_0"), val = tensor<int32, [2]>([1, 1])];
238
+ int32 up_groups_0 = const()[name = string("up_groups_0"), val = int32(1)];
239
+ tensor<fp16, [2688, 896, 1, 1]> var_258_to_fp16 = const()[name = string("op_258_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(71237056)))];
240
+ tensor<fp16, [2, 2688, 1, 1]> up_cast_fp16 = conv(dilations = up_dilations_0, groups = up_groups_0, pad = up_pad_0, pad_type = up_pad_type_0, strides = up_strides_0, weight = var_258_to_fp16, x = x_55_cast_fp16)[name = string("up_cast_fp16")];
241
+ tensor<fp16, [2, 2688, 1, 1]> gate_cast_fp16 = silu(x = input_cast_fp16)[name = string("gate_cast_fp16")];
242
+ tensor<fp16, [2, 2688, 1, 1]> x_57_cast_fp16 = mul(x = gate_cast_fp16, y = up_cast_fp16)[name = string("x_57_cast_fp16")];
243
+ string var_319_pad_type_0 = const()[name = string("op_319_pad_type_0"), val = string("valid")];
244
+ tensor<int32, [2]> var_319_strides_0 = const()[name = string("op_319_strides_0"), val = tensor<int32, [2]>([1, 1])];
245
+ tensor<int32, [4]> var_319_pad_0 = const()[name = string("op_319_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
246
+ tensor<int32, [2]> var_319_dilations_0 = const()[name = string("op_319_dilations_0"), val = tensor<int32, [2]>([1, 1])];
247
+ int32 var_319_groups_0 = const()[name = string("op_319_groups_0"), val = int32(1)];
248
+ tensor<fp16, [896, 2688, 1, 1]> var_259_to_fp16 = const()[name = string("op_259_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(76054016)))];
249
+ tensor<fp16, [2, 896, 1, 1]> var_319_cast_fp16 = conv(dilations = var_319_dilations_0, groups = var_319_groups_0, pad = var_319_pad_0, pad_type = var_319_pad_type_0, strides = var_319_strides_0, weight = var_259_to_fp16, x = x_57_cast_fp16)[name = string("op_319_cast_fp16")];
250
+ tensor<fp16, [2, 896, 1, 1]> var_320_cast_fp16 = mul(x = var_275_cast_fp16_2, y = var_319_cast_fp16)[name = string("op_320_cast_fp16")];
251
+ tensor<fp16, [2, 896, 1, 1]> x_61_cast_fp16 = add(x = x_47_cast_fp16, y = var_320_cast_fp16)[name = string("x_61_cast_fp16")];
252
+ int32 var_327 = const()[name = string("op_327"), val = int32(1)];
253
+ string var_335_pad_type_0 = const()[name = string("op_335_pad_type_0"), val = string("valid")];
254
+ tensor<int32, [2]> var_335_strides_0 = const()[name = string("op_335_strides_0"), val = tensor<int32, [2]>([1, 1])];
255
+ tensor<int32, [4]> var_335_pad_0 = const()[name = string("op_335_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
256
+ tensor<int32, [2]> var_335_dilations_0 = const()[name = string("op_335_dilations_0"), val = tensor<int32, [2]>([1, 1])];
257
+ int32 var_335_groups_0 = const()[name = string("op_335_groups_0"), val = int32(1)];
258
+ tensor<fp16, [1792, 896, 1, 1]> var_329_to_fp16 = const()[name = string("op_329_to_fp16"), val = tensor<fp16, [1792, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(80870976)))];
259
+ tensor<fp16, [2, 1792, 1, 1]> var_335_cast_fp16 = conv(dilations = var_335_dilations_0, groups = var_335_groups_0, pad = var_335_pad_0, pad_type = var_335_pad_type_0, strides = var_335_strides_0, weight = var_329_to_fp16, x = x_3_cast_fp16)[name = string("op_335_cast_fp16")];
260
+ tensor<int32, [2]> var_336_split_sizes_0 = const()[name = string("op_336_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
261
+ int32 var_336_axis_0 = const()[name = string("op_336_axis_0"), val = int32(1)];
262
+ tensor<fp16, [2, 896, 1, 1]> var_336_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_336_cast_fp16_1 = split(axis = var_336_axis_0, split_sizes = var_336_split_sizes_0, x = var_335_cast_fp16)[name = string("op_336_cast_fp16")];
263
+ fp16 const_4_promoted_to_fp16 = const()[name = string("const_4_promoted_to_fp16"), val = fp16(-0x1p+0)];
264
+ tensor<fp16, [2, 896, 1, 1]> var_338_cast_fp16 = mul(x = x_61_cast_fp16, y = const_4_promoted_to_fp16)[name = string("op_338_cast_fp16")];
265
+ bool x_63_interleave_0 = const()[name = string("x_63_interleave_0"), val = bool(false)];
266
+ tensor<fp16, [2, 1792, 1, 1]> x_63_cast_fp16 = concat(axis = var_327, interleave = x_63_interleave_0, values = (x_61_cast_fp16, var_338_cast_fp16))[name = string("x_63_cast_fp16")];
267
+ tensor<int32, [1]> out_25_axes_0 = const()[name = string("out_25_axes_0"), val = tensor<int32, [1]>([1])];
268
+ fp16 var_348_to_fp16 = const()[name = string("op_348_to_fp16"), val = fp16(0x1.5p-17)];
269
+ tensor<fp16, [2, 1792, 1, 1]> out_25_cast_fp16 = layer_norm(axes = out_25_axes_0, epsilon = var_348_to_fp16, x = x_63_cast_fp16)[name = string("out_25_cast_fp16")];
270
+ tensor<int32, [2]> var_352_split_sizes_0 = const()[name = string("op_352_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
271
+ int32 var_352_axis_0 = const()[name = string("op_352_axis_0"), val = int32(1)];
272
+ tensor<fp16, [2, 896, 1, 1]> var_352_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_352_cast_fp16_1 = split(axis = var_352_axis_0, split_sizes = var_352_split_sizes_0, x = out_25_cast_fp16)[name = string("op_352_cast_fp16")];
273
+ fp16 var_355_promoted_to_fp16 = const()[name = string("op_355_promoted_to_fp16"), val = fp16(0x1p+0)];
274
+ tensor<fp16, [2, 896, 1, 1]> var_356_cast_fp16 = add(x = var_336_cast_fp16_1, y = var_355_promoted_to_fp16)[name = string("op_356_cast_fp16")];
275
+ tensor<fp16, [2, 896, 1, 1]> var_357_cast_fp16 = mul(x = var_352_cast_fp16_0, y = var_356_cast_fp16)[name = string("op_357_cast_fp16")];
276
+ tensor<fp16, [2, 896, 1, 1]> x_cast_fp16 = add(x = var_357_cast_fp16, y = var_336_cast_fp16_0)[name = string("x_cast_fp16")];
277
+ string var_363_pad_type_0 = const()[name = string("op_363_pad_type_0"), val = string("valid")];
278
+ tensor<int32, [2]> var_363_strides_0 = const()[name = string("op_363_strides_0"), val = tensor<int32, [2]>([1, 1])];
279
+ tensor<int32, [4]> var_363_pad_0 = const()[name = string("op_363_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
280
+ tensor<int32, [2]> var_363_dilations_0 = const()[name = string("op_363_dilations_0"), val = tensor<int32, [2]>([1, 1])];
281
+ int32 var_363_groups_0 = const()[name = string("op_363_groups_0"), val = int32(1)];
282
+ tensor<fp16, [64, 896, 1, 1]> var_322_to_fp16 = const()[name = string("op_322_to_fp16"), val = tensor<fp16, [64, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(84082304)))];
283
+ tensor<fp16, [2, 64, 1, 1]> predicted_noise = conv(dilations = var_363_dilations_0, groups = var_363_groups_0, pad = var_363_pad_0, pad_type = var_363_pad_type_0, strides = var_363_strides_0, weight = var_322_to_fp16, x = x_cast_fp16)[name = string("op_363_cast_fp16")];
284
+ } -> (predicted_noise);
285
+ }
diffusion_head_model.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d769d4dab4ad19b6330272e794f83af1e2551912f58ea4ab84e84c09fe5220e6
3
+ size 84197056
tts_eos_classifier.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04f65dee434405e684dd155f8a97affdc668a2ade012a1583813ccc88a1e44ba
3
+ size 243
tts_eos_classifier.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dca3bc2766c5c819eb3434e15d8a6229ee1a4243da8a91b1f339bdd69fc0da8d
3
+ size 311
tts_eos_classifier.mlmodelc/model.mil ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3405.2.1"}, {"coremlc-version", "3405.2.1"}, {"coremltools-component-milinternal", ""}, {"coremltools-version", "9.0"}})]
3
+ {
4
+ func main<ios18>(tensor<fp16, [1, 896, 1]> input_x) {
5
+ string var_14_pad_type_0 = const()[name = string("op_14_pad_type_0"), val = string("valid")];
6
+ tensor<int32, [1]> var_14_strides_0 = const()[name = string("op_14_strides_0"), val = tensor<int32, [1]>([1])];
7
+ tensor<int32, [2]> var_14_pad_0 = const()[name = string("op_14_pad_0"), val = tensor<int32, [2]>([0, 0])];
8
+ tensor<int32, [1]> var_14_dilations_0 = const()[name = string("op_14_dilations_0"), val = tensor<int32, [1]>([1])];
9
+ int32 var_14_groups_0 = const()[name = string("op_14_groups_0"), val = int32(1)];
10
+ tensor<fp16, [896, 896, 1]> var_8_to_fp16 = const()[name = string("op_8_to_fp16"), val = tensor<fp16, [896, 896, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
11
+ tensor<fp16, [896]> fc1_bias_to_fp16 = const()[name = string("fc1_bias_to_fp16"), val = tensor<fp16, [896]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1605760)))];
12
+ tensor<fp16, [1, 896, 1]> var_14_cast_fp16 = conv(bias = fc1_bias_to_fp16, dilations = var_14_dilations_0, groups = var_14_groups_0, pad = var_14_pad_0, pad_type = var_14_pad_type_0, strides = var_14_strides_0, weight = var_8_to_fp16, x = input_x)[name = string("op_14_cast_fp16")];
13
+ tensor<fp16, [1, 896, 1]> x_cast_fp16 = relu(x = var_14_cast_fp16)[name = string("x_cast_fp16")];
14
+ string var_26_pad_type_0 = const()[name = string("op_26_pad_type_0"), val = string("valid")];
15
+ tensor<int32, [1]> var_26_strides_0 = const()[name = string("op_26_strides_0"), val = tensor<int32, [1]>([1])];
16
+ tensor<int32, [2]> var_26_pad_0 = const()[name = string("op_26_pad_0"), val = tensor<int32, [2]>([0, 0])];
17
+ tensor<int32, [1]> var_26_dilations_0 = const()[name = string("op_26_dilations_0"), val = tensor<int32, [1]>([1])];
18
+ int32 var_26_groups_0 = const()[name = string("op_26_groups_0"), val = int32(1)];
19
+ tensor<fp16, [1, 896, 1]> var_20_to_fp16 = const()[name = string("op_20_to_fp16"), val = tensor<fp16, [1, 896, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1607616)))];
20
+ tensor<fp16, [1]> fc2_bias_to_fp16 = const()[name = string("fc2_bias_to_fp16"), val = tensor<fp16, [1]>([-0x1.36p-5])];
21
+ tensor<fp16, [1, 1, 1]> output = conv(bias = fc2_bias_to_fp16, dilations = var_26_dilations_0, groups = var_26_groups_0, pad = var_26_pad_0, pad_type = var_26_pad_type_0, strides = var_26_strides_0, weight = var_20_to_fp16, x = x_cast_fp16)[name = string("op_26_cast_fp16")];
22
+ } -> (output);
23
+ }
tts_eos_classifier.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dea5cf930bfc270775854b5e4fab3f5214f3c366f2b09bb0f6cc6293bffe8f9b
3
+ size 1609472
tts_input_types.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af6397a066bc69002280f54fa1bec73be40c7e8f6f671c41f8215e61a0cc0bf6
3
+ size 3712
vibe_voice_lm_model_seqlen_32.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:747917dedac470617b6eca97432d5a687327f8948b8fc8d42a1834d3e03ecf10
3
+ size 243
vibe_voice_lm_model_seqlen_32.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b99dc1e16c07edb47e0de2d9afa8227c1f487587c22524f6fa7283f1250f12f2
3
+ size 411
vibe_voice_lm_model_seqlen_32.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
vibe_voice_lm_model_seqlen_32.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce755198818c1db103eca266d8da9c9a129794f4ad325f7f0db5b2fd5dded0f3
3
+ size 119980544
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f8fef128a776b803c235e8bfb7dfe6a9215a4fd76ce36eb2292130b4ab8b20f
3
+ size 243
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3059be1bffdefb805a387df47363f3c7aa736c02ddc37570e3eae1d4bc501ef2
3
+ size 470
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41c092353dfa7572f45a72dc46fb0abcdca886fd57eaef65026d73222cacbfa8
3
+ size 597151616
vibevoice_embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:409151256820df5c3c5f816050bbd8cf86cfc075758002696a0459ae2ca23ce4
3
+ size 272269440
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd2c98a30416ec110d7722a2736b0076032da7d43596317a1ca4da37db83798b
3
+ size 243
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1353c852f5600d7562f05a51334bc13145b0ef80add2ac966e8e761d27f652ff
3
+ size 575
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab2f84b7297c9c83434cac17a10fa88fab8b41dbde027d163b542ab02fae851
3
+ size 597151616
voices/streaming_model/de-Spk0_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc4a02421dac7523c29cca7fac9327bd41ba276bbb35c781480c933b316a1a8a
3
+ size 4381274
voices/streaming_model/de-Spk1_woman.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a0d89ba9b70498ddbec0ef1bcebacf730237721b56965de0bf4a50ecedace0e
3
+ size 3337491
voices/streaming_model/en-Carter_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d06a3c578fb7d3285fd82004ff7de76ef9806b8a23033180ab1ae42fdd7c8b0c
3
+ size 2685462
voices/streaming_model/en-Davis_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4c322bfd7c95cfd37e790af2c3c1a467b51ac11c71dee28a137dba34d3dad80
3
+ size 1560939
voices/streaming_model/en-Emma_woman.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5555dadabb0e316c14236f8d30c9133337875c95c8ab0921713ef2a74c1116d6
3
+ size 2118241
voices/streaming_model/en-Frank_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2da2d1fe940ca0beba5d099ef0d0eacb671a677e7fc30657f4f7eb62cda7cca0
3
+ size 2133945
voices/streaming_model/en-Grace_woman.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fc683ca87553d3a8d583a58f27ccc25c4ecd87847c1e8512bc1237d6b5f2cb0
3
+ size 1761680
voices/streaming_model/en-Mike_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfa43cd27ce326b1ef6abd1fd72a25eefad15b1d8fefebf6b120572e6786398b
3
+ size 1266668
voices/streaming_model/fr-Spk0_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64916486f0e7caefe421f042ecf6c74fe5fb9c6c98a7081f26a05e74d82aa2e7
3
+ size 2753069
voices/streaming_model/fr-Spk1_woman.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e587c944f6a7cd44eb7ea311f99e614154bedb47c505574dc8685d60eb6f0c92
3
+ size 2663208
voices/streaming_model/in-Samuel_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbdb9168868f2aa90ba8197b95804c274cdf9347922b07ecd08b517132c7d428
3
+ size 2398021
voices/streaming_model/it-Spk0_woman.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b84ebbff0cb296751b6bbbb147f2b30a365c2e6e2fcd02e24d5400fd4c322172
3
+ size 1612206
voices/streaming_model/it-Spk1_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f2c27e97d68778d60f51658f9de701c93b5003530bc2b19d3dcc3cffa94c67e
3
+ size 1798452
voices/streaming_model/jp-Spk0_man.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b55a4392e9f2e171af6a12e88c9646535e7b22fb2edefb90fcfd5eea9cb0b67
3
+ size 2923013
voices/streaming_model/jp-Spk1_woman.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4501729bac7b99978000940380d473a68a1432610f805a153deae76f332fbfba
3
+ size 2908957
voices/streaming_model/kr-Spk0_woman.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09208fda6c5a33c59e7eef7c171580406a4c584d31512d7270badf7578a921b4
3
+ size 2600985