Add VibeVoice CoreML models
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- .gitignore +1 -0
- acoustic_connector.mlmodelc/analytics/coremldata.bin +3 -0
- acoustic_connector.mlmodelc/coremldata.bin +3 -0
- acoustic_connector.mlmodelc/model.mil +39 -0
- acoustic_connector.mlmodelc/weights/weight.bin +3 -0
- acoustic_connector.mlpackage/Data/com.apple.CoreML/model.mlmodel +3 -0
- acoustic_connector.mlpackage/Data/com.apple.CoreML/weights/weight.bin +3 -0
- acoustic_connector.mlpackage/Manifest.json +18 -0
- decoder_coreml_12_ne.mlmodelc/analytics/coremldata.bin +3 -0
- decoder_coreml_12_ne.mlmodelc/coremldata.bin +3 -0
- decoder_coreml_12_ne.mlmodelc/model.mil +0 -0
- decoder_coreml_12_ne.mlmodelc/weights/weight.bin +3 -0
- diffusion_head_model.mlmodelc/analytics/coremldata.bin +3 -0
- diffusion_head_model.mlmodelc/coremldata.bin +3 -0
- diffusion_head_model.mlmodelc/model.mil +285 -0
- diffusion_head_model.mlmodelc/weights/weight.bin +3 -0
- tts_eos_classifier.mlmodelc/analytics/coremldata.bin +3 -0
- tts_eos_classifier.mlmodelc/coremldata.bin +3 -0
- tts_eos_classifier.mlmodelc/model.mil +23 -0
- tts_eos_classifier.mlmodelc/weights/weight.bin +3 -0
- tts_input_types.npy +3 -0
- vibe_voice_lm_model_seqlen_32.mlmodelc/analytics/coremldata.bin +3 -0
- vibe_voice_lm_model_seqlen_32.mlmodelc/coremldata.bin +3 -0
- vibe_voice_lm_model_seqlen_32.mlmodelc/model.mil +0 -0
- vibe_voice_lm_model_seqlen_32.mlmodelc/weights/weight.bin +3 -0
- vibe_voice_tts_lm_model_seqlen_8.mlmodelc/analytics/coremldata.bin +3 -0
- vibe_voice_tts_lm_model_seqlen_8.mlmodelc/coremldata.bin +3 -0
- vibe_voice_tts_lm_model_seqlen_8.mlmodelc/model.mil +0 -0
- vibe_voice_tts_lm_model_seqlen_8.mlmodelc/weights/weight.bin +3 -0
- vibevoice_embeddings.npy +3 -0
- vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/analytics/coremldata.bin +3 -0
- vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/coremldata.bin +3 -0
- vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/model.mil +0 -0
- vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/weights/weight.bin +3 -0
- voices/streaming_model/de-Spk0_man.npz +3 -0
- voices/streaming_model/de-Spk1_woman.npz +3 -0
- voices/streaming_model/en-Carter_man.npz +3 -0
- voices/streaming_model/en-Davis_man.npz +3 -0
- voices/streaming_model/en-Emma_woman.npz +3 -0
- voices/streaming_model/en-Frank_man.npz +3 -0
- voices/streaming_model/en-Grace_woman.npz +3 -0
- voices/streaming_model/en-Mike_man.npz +3 -0
- voices/streaming_model/fr-Spk0_man.npz +3 -0
- voices/streaming_model/fr-Spk1_woman.npz +3 -0
- voices/streaming_model/in-Samuel_man.npz +3 -0
- voices/streaming_model/it-Spk0_woman.npz +3 -0
- voices/streaming_model/it-Spk1_man.npz +3 -0
- voices/streaming_model/jp-Spk0_man.npz +3 -0
- voices/streaming_model/jp-Spk1_woman.npz +3 -0
- voices/streaming_model/kr-Spk0_woman.npz +3 -0
.gitignore
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
*.DS_Store
|
acoustic_connector.mlmodelc/analytics/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2249a24e4950fe95fee340914d53f3bab173945d674fcaaff1fe3ef8e8b8b91d
|
| 3 |
+
size 243
|
acoustic_connector.mlmodelc/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e8bc76edab8d50f1a45b48c600a7417bbc08feed346a25be77a291660dfa9eb
|
| 3 |
+
size 311
|
acoustic_connector.mlmodelc/model.mil
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
program(1.3)
|
| 2 |
+
[buildInfo = dict<string, string>({{"coremlc-component-MIL", "3405.2.1"}, {"coremlc-version", "3405.2.1"}, {"coremltools-component-milinternal", ""}, {"coremltools-version", "9.0"}})]
|
| 3 |
+
{
|
| 4 |
+
func main<ios18>(tensor<fp16, [1, 64, 1]> input_x) {
|
| 5 |
+
string x_1_pad_type_0 = const()[name = string("x_1_pad_type_0"), val = string("valid")];
|
| 6 |
+
tensor<int32, [1]> x_1_strides_0 = const()[name = string("x_1_strides_0"), val = tensor<int32, [1]>([1])];
|
| 7 |
+
tensor<int32, [2]> x_1_pad_0 = const()[name = string("x_1_pad_0"), val = tensor<int32, [2]>([0, 0])];
|
| 8 |
+
tensor<int32, [1]> x_1_dilations_0 = const()[name = string("x_1_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 9 |
+
int32 x_1_groups_0 = const()[name = string("x_1_groups_0"), val = int32(1)];
|
| 10 |
+
tensor<fp16, [896, 64, 1]> var_9_to_fp16 = const()[name = string("op_9_to_fp16"), val = tensor<fp16, [896, 64, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
|
| 11 |
+
tensor<fp16, [896]> fc1_bias_to_fp16 = const()[name = string("fc1_bias_to_fp16"), val = tensor<fp16, [896]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(114816)))];
|
| 12 |
+
tensor<fp16, [1, 896, 1]> x_1_cast_fp16 = conv(bias = fc1_bias_to_fp16, dilations = x_1_dilations_0, groups = x_1_groups_0, pad = x_1_pad_0, pad_type = x_1_pad_type_0, strides = x_1_strides_0, weight = var_9_to_fp16, x = input_x)[name = string("x_1_cast_fp16")];
|
| 13 |
+
int32 var_20 = const()[name = string("op_20"), val = int32(1)];
|
| 14 |
+
tensor<int32, [1]> x_3_axes_0 = const()[name = string("x_3_axes_0"), val = tensor<int32, [1]>([-2])];
|
| 15 |
+
tensor<fp16, [1, 896, 1, 1]> x_3_cast_fp16 = expand_dims(axes = x_3_axes_0, x = x_1_cast_fp16)[name = string("x_3_cast_fp16")];
|
| 16 |
+
fp16 const_0_promoted_to_fp16 = const()[name = string("const_0_promoted_to_fp16"), val = fp16(-0x1p+0)];
|
| 17 |
+
tensor<fp16, [1, 896, 1, 1]> var_25_cast_fp16 = mul(x = x_3_cast_fp16, y = const_0_promoted_to_fp16)[name = string("op_25_cast_fp16")];
|
| 18 |
+
bool x_5_interleave_0 = const()[name = string("x_5_interleave_0"), val = bool(false)];
|
| 19 |
+
tensor<fp16, [1, 1792, 1, 1]> x_5_cast_fp16 = concat(axis = var_20, interleave = x_5_interleave_0, values = (x_3_cast_fp16, var_25_cast_fp16))[name = string("x_5_cast_fp16")];
|
| 20 |
+
tensor<int32, [1]> out_1_axes_0 = const()[name = string("out_1_axes_0"), val = tensor<int32, [1]>([1])];
|
| 21 |
+
fp16 var_35_to_fp16 = const()[name = string("op_35_to_fp16"), val = fp16(0x1.1p-20)];
|
| 22 |
+
tensor<fp16, [1, 1792, 1, 1]> out_1_cast_fp16 = layer_norm(axes = out_1_axes_0, epsilon = var_35_to_fp16, x = x_5_cast_fp16)[name = string("out_1_cast_fp16")];
|
| 23 |
+
tensor<fp16, [1, 1792, 1, 1]> norm_weight_to_fp16 = const()[name = string("norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(116672)))];
|
| 24 |
+
tensor<fp16, [1, 1792, 1, 1]> out_3_cast_fp16 = mul(x = out_1_cast_fp16, y = norm_weight_to_fp16)[name = string("out_3_cast_fp16")];
|
| 25 |
+
tensor<int32, [2]> var_41_split_sizes_0 = const()[name = string("op_41_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
|
| 26 |
+
int32 var_41_axis_0 = const()[name = string("op_41_axis_0"), val = int32(1)];
|
| 27 |
+
tensor<fp16, [1, 896, 1, 1]> var_41_cast_fp16_0, tensor<fp16, [1, 896, 1, 1]> var_41_cast_fp16_1 = split(axis = var_41_axis_0, split_sizes = var_41_split_sizes_0, x = out_3_cast_fp16)[name = string("op_41_cast_fp16")];
|
| 28 |
+
tensor<int32, [1]> x_axes_0 = const()[name = string("x_axes_0"), val = tensor<int32, [1]>([-2])];
|
| 29 |
+
tensor<fp16, [1, 896, 1]> x_cast_fp16 = squeeze(axes = x_axes_0, x = var_41_cast_fp16_0)[name = string("x_cast_fp16")];
|
| 30 |
+
string var_55_pad_type_0 = const()[name = string("op_55_pad_type_0"), val = string("valid")];
|
| 31 |
+
tensor<int32, [1]> var_55_strides_0 = const()[name = string("op_55_strides_0"), val = tensor<int32, [1]>([1])];
|
| 32 |
+
tensor<int32, [2]> var_55_pad_0 = const()[name = string("op_55_pad_0"), val = tensor<int32, [2]>([0, 0])];
|
| 33 |
+
tensor<int32, [1]> var_55_dilations_0 = const()[name = string("op_55_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 34 |
+
int32 var_55_groups_0 = const()[name = string("op_55_groups_0"), val = int32(1)];
|
| 35 |
+
tensor<fp16, [896, 896, 1]> var_49_to_fp16 = const()[name = string("op_49_to_fp16"), val = tensor<fp16, [896, 896, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(120320)))];
|
| 36 |
+
tensor<fp16, [896]> fc2_bias_to_fp16 = const()[name = string("fc2_bias_to_fp16"), val = tensor<fp16, [896]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1726016)))];
|
| 37 |
+
tensor<fp16, [1, 896, 1]> output = conv(bias = fc2_bias_to_fp16, dilations = var_55_dilations_0, groups = var_55_groups_0, pad = var_55_pad_0, pad_type = var_55_pad_type_0, strides = var_55_strides_0, weight = var_49_to_fp16, x = x_cast_fp16)[name = string("op_55_cast_fp16")];
|
| 38 |
+
} -> (output);
|
| 39 |
+
}
|
acoustic_connector.mlmodelc/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c32f4b059cd82c4eac2af379ff3686ca657f13c8a5b96809719351ab6f1ebb35
|
| 3 |
+
size 1727872
|
acoustic_connector.mlpackage/Data/com.apple.CoreML/model.mlmodel
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:037d6874f5deca07a926cf619475199d686a8ca143ec6252a215a65ac56e8957
|
| 3 |
+
size 5151
|
acoustic_connector.mlpackage/Data/com.apple.CoreML/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c32f4b059cd82c4eac2af379ff3686ca657f13c8a5b96809719351ab6f1ebb35
|
| 3 |
+
size 1727872
|
acoustic_connector.mlpackage/Manifest.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"fileFormatVersion": "1.0.0",
|
| 3 |
+
"itemInfoEntries": {
|
| 4 |
+
"966eb60d-af76-4b59-ab8e-302dfe8a26ac": {
|
| 5 |
+
"author": "com.apple.CoreML",
|
| 6 |
+
"description": "CoreML Model Weights",
|
| 7 |
+
"name": "weights",
|
| 8 |
+
"path": "com.apple.CoreML/weights"
|
| 9 |
+
},
|
| 10 |
+
"c3a8450e-db98-4818-b49d-61a61abf029a": {
|
| 11 |
+
"author": "com.apple.CoreML",
|
| 12 |
+
"description": "CoreML Model Specification",
|
| 13 |
+
"name": "model.mlmodel",
|
| 14 |
+
"path": "com.apple.CoreML/model.mlmodel"
|
| 15 |
+
}
|
| 16 |
+
},
|
| 17 |
+
"rootModelIdentifier": "c3a8450e-db98-4818-b49d-61a61abf029a"
|
| 18 |
+
}
|
decoder_coreml_12_ne.mlmodelc/analytics/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b968c0e7f610fbc5fac13cf6044682ba49b62b2f24021430ab81379dd5246080
|
| 3 |
+
size 243
|
decoder_coreml_12_ne.mlmodelc/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f76e909cf4859b460963ef38d863a85005db09d3def7305a5ac17ce56cc718ca
|
| 3 |
+
size 738
|
decoder_coreml_12_ne.mlmodelc/model.mil
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
decoder_coreml_12_ne.mlmodelc/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2db7f6ae03e24acf862b5f21c73a8ce5c1932d3a75423766141457a21db2a74f
|
| 3 |
+
size 687406272
|
diffusion_head_model.mlmodelc/analytics/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:97a61f48b53134e0a5bbefb076fbe166269691a90a00faef7841288f725d0caf
|
| 3 |
+
size 243
|
diffusion_head_model.mlmodelc/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ab6d23550b62e579a3d527e42a8f40f955ad7c173ffb3f0a7e11ea8e04b2f151
|
| 3 |
+
size 443
|
diffusion_head_model.mlmodelc/model.mil
ADDED
|
@@ -0,0 +1,285 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
program(1.3)
|
| 2 |
+
[buildInfo = dict<string, string>({{"coremlc-component-MIL", "3405.2.1"}, {"coremlc-version", "3405.2.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})]
|
| 3 |
+
{
|
| 4 |
+
func main<ios18>(tensor<fp16, [2, 896, 1, 1]> condition, tensor<fp16, [2, 64, 1, 1]> noisy_images, tensor<fp16, [256, 1, 1]> timesteps) {
|
| 5 |
+
string x_5_pad_type_0 = const()[name = string("x_5_pad_type_0"), val = string("valid")];
|
| 6 |
+
tensor<int32, [2]> x_5_strides_0 = const()[name = string("x_5_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 7 |
+
tensor<int32, [4]> x_5_pad_0 = const()[name = string("x_5_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 8 |
+
tensor<int32, [2]> x_5_dilations_0 = const()[name = string("x_5_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 9 |
+
int32 x_5_groups_0 = const()[name = string("x_5_groups_0"), val = int32(1)];
|
| 10 |
+
tensor<fp16, [896, 64, 1, 1]> var_21_to_fp16 = const()[name = string("op_21_to_fp16"), val = tensor<fp16, [896, 64, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
|
| 11 |
+
tensor<fp16, [2, 896, 1, 1]> x_5_cast_fp16 = conv(dilations = x_5_dilations_0, groups = x_5_groups_0, pad = x_5_pad_0, pad_type = x_5_pad_type_0, strides = x_5_strides_0, weight = var_21_to_fp16, x = noisy_images)[name = string("x_5_cast_fp16")];
|
| 12 |
+
tensor<int32, [1]> var_34_axes_0 = const()[name = string("op_34_axes_0"), val = tensor<int32, [1]>([0])];
|
| 13 |
+
tensor<fp16, [1, 256, 1, 1]> var_34_cast_fp16 = expand_dims(axes = var_34_axes_0, x = timesteps)[name = string("op_34_cast_fp16")];
|
| 14 |
+
string var_39_pad_type_0 = const()[name = string("op_39_pad_type_0"), val = string("valid")];
|
| 15 |
+
tensor<int32, [2]> var_39_strides_0 = const()[name = string("op_39_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 16 |
+
tensor<int32, [4]> var_39_pad_0 = const()[name = string("op_39_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 17 |
+
tensor<int32, [2]> var_39_dilations_0 = const()[name = string("op_39_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 18 |
+
int32 var_39_groups_0 = const()[name = string("op_39_groups_0"), val = int32(1)];
|
| 19 |
+
tensor<fp16, [896, 256, 1, 1]> var_29_to_fp16 = const()[name = string("op_29_to_fp16"), val = tensor<fp16, [896, 256, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(114816)))];
|
| 20 |
+
tensor<fp16, [1, 896, 1, 1]> var_39_cast_fp16 = conv(dilations = var_39_dilations_0, groups = var_39_groups_0, pad = var_39_pad_0, pad_type = var_39_pad_type_0, strides = var_39_strides_0, weight = var_29_to_fp16, x = var_34_cast_fp16)[name = string("op_39_cast_fp16")];
|
| 21 |
+
tensor<int32, [1]> input_1_axes_0 = const()[name = string("input_1_axes_0"), val = tensor<int32, [1]>([0])];
|
| 22 |
+
tensor<fp16, [896, 1, 1]> input_1_cast_fp16 = squeeze(axes = input_1_axes_0, x = var_39_cast_fp16)[name = string("input_1_cast_fp16")];
|
| 23 |
+
tensor<fp16, [896, 1, 1]> x_1_cast_fp16 = silu(x = input_1_cast_fp16)[name = string("x_1_cast_fp16")];
|
| 24 |
+
tensor<int32, [1]> var_42_axes_0 = const()[name = string("op_42_axes_0"), val = tensor<int32, [1]>([0])];
|
| 25 |
+
tensor<fp16, [1, 896, 1, 1]> var_42_cast_fp16 = expand_dims(axes = var_42_axes_0, x = x_1_cast_fp16)[name = string("op_42_cast_fp16")];
|
| 26 |
+
string var_47_pad_type_0 = const()[name = string("op_47_pad_type_0"), val = string("valid")];
|
| 27 |
+
tensor<int32, [2]> var_47_strides_0 = const()[name = string("op_47_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 28 |
+
tensor<int32, [4]> var_47_pad_0 = const()[name = string("op_47_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 29 |
+
tensor<int32, [2]> var_47_dilations_0 = const()[name = string("op_47_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 30 |
+
int32 var_47_groups_0 = const()[name = string("op_47_groups_0"), val = int32(1)];
|
| 31 |
+
tensor<fp16, [896, 896, 1, 1]> var_27_to_fp16 = const()[name = string("op_27_to_fp16"), val = tensor<fp16, [896, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(573632)))];
|
| 32 |
+
tensor<fp16, [1, 896, 1, 1]> var_47_cast_fp16 = conv(dilations = var_47_dilations_0, groups = var_47_groups_0, pad = var_47_pad_0, pad_type = var_47_pad_type_0, strides = var_47_strides_0, weight = var_27_to_fp16, x = var_42_cast_fp16)[name = string("op_47_cast_fp16")];
|
| 33 |
+
tensor<int32, [1]> t_axes_0 = const()[name = string("t_axes_0"), val = tensor<int32, [1]>([0])];
|
| 34 |
+
tensor<fp16, [896, 1, 1]> t_cast_fp16 = squeeze(axes = t_axes_0, x = var_47_cast_fp16)[name = string("t_cast_fp16")];
|
| 35 |
+
string condition_pad_type_0 = const()[name = string("condition_pad_type_0"), val = string("valid")];
|
| 36 |
+
tensor<int32, [2]> condition_strides_0 = const()[name = string("condition_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 37 |
+
tensor<int32, [4]> condition_pad_0 = const()[name = string("condition_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 38 |
+
tensor<int32, [2]> condition_dilations_0 = const()[name = string("condition_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 39 |
+
int32 condition_groups_0 = const()[name = string("condition_groups_0"), val = int32(1)];
|
| 40 |
+
tensor<fp16, [896, 896, 1, 1]> var_54_to_fp16 = const()[name = string("op_54_to_fp16"), val = tensor<fp16, [896, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2179328)))];
|
| 41 |
+
tensor<fp16, [2, 896, 1, 1]> condition_cast_fp16 = conv(dilations = condition_dilations_0, groups = condition_groups_0, pad = condition_pad_0, pad_type = condition_pad_type_0, strides = condition_strides_0, weight = var_54_to_fp16, x = condition)[name = string("condition_cast_fp16")];
|
| 42 |
+
tensor<fp16, [2, 896, 1, 1]> input_3_cast_fp16 = add(x = condition_cast_fp16, y = t_cast_fp16)[name = string("input_3_cast_fp16")];
|
| 43 |
+
int32 var_69 = const()[name = string("op_69"), val = int32(1)];
|
| 44 |
+
tensor<fp16, [2, 896, 1, 1]> x_3_cast_fp16 = silu(x = input_3_cast_fp16)[name = string("x_3_cast_fp16")];
|
| 45 |
+
string var_79_pad_type_0 = const()[name = string("op_79_pad_type_0"), val = string("valid")];
|
| 46 |
+
tensor<int32, [2]> var_79_strides_0 = const()[name = string("op_79_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 47 |
+
tensor<int32, [4]> var_79_pad_0 = const()[name = string("op_79_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 48 |
+
tensor<int32, [2]> var_79_dilations_0 = const()[name = string("op_79_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 49 |
+
int32 var_79_groups_0 = const()[name = string("op_79_groups_0"), val = int32(1)];
|
| 50 |
+
tensor<fp16, [2688, 896, 1, 1]> var_71_to_fp16 = const()[name = string("op_71_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3785024)))];
|
| 51 |
+
tensor<fp16, [2, 2688, 1, 1]> var_79_cast_fp16 = conv(dilations = var_79_dilations_0, groups = var_79_groups_0, pad = var_79_pad_0, pad_type = var_79_pad_type_0, strides = var_79_strides_0, weight = var_71_to_fp16, x = x_3_cast_fp16)[name = string("op_79_cast_fp16")];
|
| 52 |
+
tensor<int32, [3]> var_80_split_sizes_0 = const()[name = string("op_80_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
|
| 53 |
+
int32 var_80_axis_0 = const()[name = string("op_80_axis_0"), val = int32(1)];
|
| 54 |
+
tensor<fp16, [2, 896, 1, 1]> var_80_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_80_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_80_cast_fp16_2 = split(axis = var_80_axis_0, split_sizes = var_80_split_sizes_0, x = var_79_cast_fp16)[name = string("op_80_cast_fp16")];
|
| 55 |
+
fp16 const_0_promoted_to_fp16 = const()[name = string("const_0_promoted_to_fp16"), val = fp16(-0x1p+0)];
|
| 56 |
+
tensor<fp16, [2, 896, 1, 1]> var_85_cast_fp16 = mul(x = x_5_cast_fp16, y = const_0_promoted_to_fp16)[name = string("op_85_cast_fp16")];
|
| 57 |
+
bool x_7_interleave_0 = const()[name = string("x_7_interleave_0"), val = bool(false)];
|
| 58 |
+
tensor<fp16, [2, 1792, 1, 1]> x_7_cast_fp16 = concat(axis = var_69, interleave = x_7_interleave_0, values = (x_5_cast_fp16, var_85_cast_fp16))[name = string("x_7_cast_fp16")];
|
| 59 |
+
tensor<int32, [1]> out_1_axes_0 = const()[name = string("out_1_axes_0"), val = tensor<int32, [1]>([1])];
|
| 60 |
+
fp16 var_95_to_fp16 = const()[name = string("op_95_to_fp16"), val = fp16(0x1.5p-17)];
|
| 61 |
+
tensor<fp16, [2, 1792, 1, 1]> out_1_cast_fp16 = layer_norm(axes = out_1_axes_0, epsilon = var_95_to_fp16, x = x_7_cast_fp16)[name = string("out_1_cast_fp16")];
|
| 62 |
+
tensor<fp16, [1, 1792, 1, 1]> layer_layers_0_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_0_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(8601984)))];
|
| 63 |
+
tensor<fp16, [2, 1792, 1, 1]> out_3_cast_fp16 = mul(x = out_1_cast_fp16, y = layer_layers_0_layer_norm_weight_to_fp16)[name = string("out_3_cast_fp16")];
|
| 64 |
+
tensor<int32, [2]> var_101_split_sizes_0 = const()[name = string("op_101_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
|
| 65 |
+
int32 var_101_axis_0 = const()[name = string("op_101_axis_0"), val = int32(1)];
|
| 66 |
+
tensor<fp16, [2, 896, 1, 1]> var_101_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_101_cast_fp16_1 = split(axis = var_101_axis_0, split_sizes = var_101_split_sizes_0, x = out_3_cast_fp16)[name = string("op_101_cast_fp16")];
|
| 67 |
+
fp16 var_104_promoted_to_fp16 = const()[name = string("op_104_promoted_to_fp16"), val = fp16(0x1p+0)];
|
| 68 |
+
tensor<fp16, [2, 896, 1, 1]> var_105_cast_fp16 = add(x = var_80_cast_fp16_1, y = var_104_promoted_to_fp16)[name = string("op_105_cast_fp16")];
|
| 69 |
+
tensor<fp16, [2, 896, 1, 1]> var_106_cast_fp16 = mul(x = var_101_cast_fp16_0, y = var_105_cast_fp16)[name = string("op_106_cast_fp16")];
|
| 70 |
+
tensor<fp16, [2, 896, 1, 1]> x_13_cast_fp16 = add(x = var_106_cast_fp16, y = var_80_cast_fp16_0)[name = string("x_13_cast_fp16")];
|
| 71 |
+
string input_5_pad_type_0 = const()[name = string("input_5_pad_type_0"), val = string("valid")];
|
| 72 |
+
tensor<int32, [2]> input_5_strides_0 = const()[name = string("input_5_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 73 |
+
tensor<int32, [4]> input_5_pad_0 = const()[name = string("input_5_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 74 |
+
tensor<int32, [2]> input_5_dilations_0 = const()[name = string("input_5_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 75 |
+
int32 input_5_groups_0 = const()[name = string("input_5_groups_0"), val = int32(1)];
|
| 76 |
+
tensor<fp16, [2688, 896, 1, 1]> var_62_to_fp16 = const()[name = string("op_62_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(8605632)))];
|
| 77 |
+
tensor<fp16, [2, 2688, 1, 1]> input_5_cast_fp16 = conv(dilations = input_5_dilations_0, groups = input_5_groups_0, pad = input_5_pad_0, pad_type = input_5_pad_type_0, strides = input_5_strides_0, weight = var_62_to_fp16, x = x_13_cast_fp16)[name = string("input_5_cast_fp16")];
|
| 78 |
+
string up_1_pad_type_0 = const()[name = string("up_1_pad_type_0"), val = string("valid")];
|
| 79 |
+
tensor<int32, [2]> up_1_strides_0 = const()[name = string("up_1_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 80 |
+
tensor<int32, [4]> up_1_pad_0 = const()[name = string("up_1_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 81 |
+
tensor<int32, [2]> up_1_dilations_0 = const()[name = string("up_1_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 82 |
+
int32 up_1_groups_0 = const()[name = string("up_1_groups_0"), val = int32(1)];
|
| 83 |
+
tensor<fp16, [2688, 896, 1, 1]> var_63_to_fp16 = const()[name = string("op_63_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(13422592)))];
|
| 84 |
+
tensor<fp16, [2, 2688, 1, 1]> up_1_cast_fp16 = conv(dilations = up_1_dilations_0, groups = up_1_groups_0, pad = up_1_pad_0, pad_type = up_1_pad_type_0, strides = up_1_strides_0, weight = var_63_to_fp16, x = x_13_cast_fp16)[name = string("up_1_cast_fp16")];
|
| 85 |
+
tensor<fp16, [2, 2688, 1, 1]> gate_1_cast_fp16 = silu(x = input_5_cast_fp16)[name = string("gate_1_cast_fp16")];
|
| 86 |
+
tensor<fp16, [2, 2688, 1, 1]> x_15_cast_fp16 = mul(x = gate_1_cast_fp16, y = up_1_cast_fp16)[name = string("x_15_cast_fp16")];
|
| 87 |
+
string var_124_pad_type_0 = const()[name = string("op_124_pad_type_0"), val = string("valid")];
|
| 88 |
+
tensor<int32, [2]> var_124_strides_0 = const()[name = string("op_124_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 89 |
+
tensor<int32, [4]> var_124_pad_0 = const()[name = string("op_124_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 90 |
+
tensor<int32, [2]> var_124_dilations_0 = const()[name = string("op_124_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 91 |
+
int32 var_124_groups_0 = const()[name = string("op_124_groups_0"), val = int32(1)];
|
| 92 |
+
tensor<fp16, [896, 2688, 1, 1]> var_64_to_fp16 = const()[name = string("op_64_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(18239552)))];
|
| 93 |
+
tensor<fp16, [2, 896, 1, 1]> var_124_cast_fp16 = conv(dilations = var_124_dilations_0, groups = var_124_groups_0, pad = var_124_pad_0, pad_type = var_124_pad_type_0, strides = var_124_strides_0, weight = var_64_to_fp16, x = x_15_cast_fp16)[name = string("op_124_cast_fp16")];
|
| 94 |
+
tensor<fp16, [2, 896, 1, 1]> var_125_cast_fp16 = mul(x = var_80_cast_fp16_2, y = var_124_cast_fp16)[name = string("op_125_cast_fp16")];
|
| 95 |
+
tensor<fp16, [2, 896, 1, 1]> x_19_cast_fp16 = add(x = x_5_cast_fp16, y = var_125_cast_fp16)[name = string("x_19_cast_fp16")];
|
| 96 |
+
int32 var_134 = const()[name = string("op_134"), val = int32(1)];
|
| 97 |
+
string var_144_pad_type_0 = const()[name = string("op_144_pad_type_0"), val = string("valid")];
|
| 98 |
+
tensor<int32, [2]> var_144_strides_0 = const()[name = string("op_144_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 99 |
+
tensor<int32, [4]> var_144_pad_0 = const()[name = string("op_144_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 100 |
+
tensor<int32, [2]> var_144_dilations_0 = const()[name = string("op_144_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 101 |
+
int32 var_144_groups_0 = const()[name = string("op_144_groups_0"), val = int32(1)];
|
| 102 |
+
tensor<fp16, [2688, 896, 1, 1]> var_136_to_fp16 = const()[name = string("op_136_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(23056512)))];
|
| 103 |
+
tensor<fp16, [2, 2688, 1, 1]> var_144_cast_fp16 = conv(dilations = var_144_dilations_0, groups = var_144_groups_0, pad = var_144_pad_0, pad_type = var_144_pad_type_0, strides = var_144_strides_0, weight = var_136_to_fp16, x = x_3_cast_fp16)[name = string("op_144_cast_fp16")];
|
| 104 |
+
tensor<int32, [3]> var_145_split_sizes_0 = const()[name = string("op_145_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
|
| 105 |
+
int32 var_145_axis_0 = const()[name = string("op_145_axis_0"), val = int32(1)];
|
| 106 |
+
tensor<fp16, [2, 896, 1, 1]> var_145_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_145_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_145_cast_fp16_2 = split(axis = var_145_axis_0, split_sizes = var_145_split_sizes_0, x = var_144_cast_fp16)[name = string("op_145_cast_fp16")];
|
| 107 |
+
fp16 const_1_promoted_to_fp16 = const()[name = string("const_1_promoted_to_fp16"), val = fp16(-0x1p+0)];
|
| 108 |
+
tensor<fp16, [2, 896, 1, 1]> var_150_cast_fp16 = mul(x = x_19_cast_fp16, y = const_1_promoted_to_fp16)[name = string("op_150_cast_fp16")];
|
| 109 |
+
bool x_21_interleave_0 = const()[name = string("x_21_interleave_0"), val = bool(false)];
|
| 110 |
+
tensor<fp16, [2, 1792, 1, 1]> x_21_cast_fp16 = concat(axis = var_134, interleave = x_21_interleave_0, values = (x_19_cast_fp16, var_150_cast_fp16))[name = string("x_21_cast_fp16")];
|
| 111 |
+
tensor<int32, [1]> out_7_axes_0 = const()[name = string("out_7_axes_0"), val = tensor<int32, [1]>([1])];
|
| 112 |
+
fp16 var_160_to_fp16 = const()[name = string("op_160_to_fp16"), val = fp16(0x1.5p-17)];
|
| 113 |
+
tensor<fp16, [2, 1792, 1, 1]> out_7_cast_fp16 = layer_norm(axes = out_7_axes_0, epsilon = var_160_to_fp16, x = x_21_cast_fp16)[name = string("out_7_cast_fp16")];
|
| 114 |
+
tensor<fp16, [1, 1792, 1, 1]> layer_layers_1_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_1_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(27873472)))];
|
| 115 |
+
tensor<fp16, [2, 1792, 1, 1]> out_9_cast_fp16 = mul(x = out_7_cast_fp16, y = layer_layers_1_layer_norm_weight_to_fp16)[name = string("out_9_cast_fp16")];
|
| 116 |
+
tensor<int32, [2]> var_166_split_sizes_0 = const()[name = string("op_166_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
|
| 117 |
+
int32 var_166_axis_0 = const()[name = string("op_166_axis_0"), val = int32(1)];
|
| 118 |
+
tensor<fp16, [2, 896, 1, 1]> var_166_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_166_cast_fp16_1 = split(axis = var_166_axis_0, split_sizes = var_166_split_sizes_0, x = out_9_cast_fp16)[name = string("op_166_cast_fp16")];
|
| 119 |
+
fp16 var_169_promoted_to_fp16 = const()[name = string("op_169_promoted_to_fp16"), val = fp16(0x1p+0)];
|
| 120 |
+
tensor<fp16, [2, 896, 1, 1]> var_170_cast_fp16 = add(x = var_145_cast_fp16_1, y = var_169_promoted_to_fp16)[name = string("op_170_cast_fp16")];
|
| 121 |
+
tensor<fp16, [2, 896, 1, 1]> var_171_cast_fp16 = mul(x = var_166_cast_fp16_0, y = var_170_cast_fp16)[name = string("op_171_cast_fp16")];
|
| 122 |
+
tensor<fp16, [2, 896, 1, 1]> x_27_cast_fp16 = add(x = var_171_cast_fp16, y = var_145_cast_fp16_0)[name = string("x_27_cast_fp16")];
|
| 123 |
+
string input_7_pad_type_0 = const()[name = string("input_7_pad_type_0"), val = string("valid")];
|
| 124 |
+
tensor<int32, [2]> input_7_strides_0 = const()[name = string("input_7_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 125 |
+
tensor<int32, [4]> input_7_pad_0 = const()[name = string("input_7_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 126 |
+
tensor<int32, [2]> input_7_dilations_0 = const()[name = string("input_7_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 127 |
+
int32 input_7_groups_0 = const()[name = string("input_7_groups_0"), val = int32(1)];
|
| 128 |
+
tensor<fp16, [2688, 896, 1, 1]> var_127_to_fp16 = const()[name = string("op_127_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(27877120)))];
|
| 129 |
+
tensor<fp16, [2, 2688, 1, 1]> input_7_cast_fp16 = conv(dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = var_127_to_fp16, x = x_27_cast_fp16)[name = string("input_7_cast_fp16")];
|
| 130 |
+
string up_3_pad_type_0 = const()[name = string("up_3_pad_type_0"), val = string("valid")];
|
| 131 |
+
tensor<int32, [2]> up_3_strides_0 = const()[name = string("up_3_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 132 |
+
tensor<int32, [4]> up_3_pad_0 = const()[name = string("up_3_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 133 |
+
tensor<int32, [2]> up_3_dilations_0 = const()[name = string("up_3_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 134 |
+
int32 up_3_groups_0 = const()[name = string("up_3_groups_0"), val = int32(1)];
|
| 135 |
+
tensor<fp16, [2688, 896, 1, 1]> var_128_to_fp16 = const()[name = string("op_128_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(32694080)))];
|
| 136 |
+
tensor<fp16, [2, 2688, 1, 1]> up_3_cast_fp16 = conv(dilations = up_3_dilations_0, groups = up_3_groups_0, pad = up_3_pad_0, pad_type = up_3_pad_type_0, strides = up_3_strides_0, weight = var_128_to_fp16, x = x_27_cast_fp16)[name = string("up_3_cast_fp16")];
|
| 137 |
+
tensor<fp16, [2, 2688, 1, 1]> gate_3_cast_fp16 = silu(x = input_7_cast_fp16)[name = string("gate_3_cast_fp16")];
|
| 138 |
+
tensor<fp16, [2, 2688, 1, 1]> x_29_cast_fp16 = mul(x = gate_3_cast_fp16, y = up_3_cast_fp16)[name = string("x_29_cast_fp16")];
|
| 139 |
+
string var_189_pad_type_0 = const()[name = string("op_189_pad_type_0"), val = string("valid")];
|
| 140 |
+
tensor<int32, [2]> var_189_strides_0 = const()[name = string("op_189_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 141 |
+
tensor<int32, [4]> var_189_pad_0 = const()[name = string("op_189_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 142 |
+
tensor<int32, [2]> var_189_dilations_0 = const()[name = string("op_189_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 143 |
+
int32 var_189_groups_0 = const()[name = string("op_189_groups_0"), val = int32(1)];
|
| 144 |
+
tensor<fp16, [896, 2688, 1, 1]> var_129_to_fp16 = const()[name = string("op_129_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(37511040)))];
|
| 145 |
+
tensor<fp16, [2, 896, 1, 1]> var_189_cast_fp16 = conv(dilations = var_189_dilations_0, groups = var_189_groups_0, pad = var_189_pad_0, pad_type = var_189_pad_type_0, strides = var_189_strides_0, weight = var_129_to_fp16, x = x_29_cast_fp16)[name = string("op_189_cast_fp16")];
|
| 146 |
+
tensor<fp16, [2, 896, 1, 1]> var_190_cast_fp16 = mul(x = var_145_cast_fp16_2, y = var_189_cast_fp16)[name = string("op_190_cast_fp16")];
|
| 147 |
+
tensor<fp16, [2, 896, 1, 1]> x_33_cast_fp16 = add(x = x_19_cast_fp16, y = var_190_cast_fp16)[name = string("x_33_cast_fp16")];
|
| 148 |
+
int32 var_199 = const()[name = string("op_199"), val = int32(1)];
|
| 149 |
+
string var_209_pad_type_0 = const()[name = string("op_209_pad_type_0"), val = string("valid")];
|
| 150 |
+
tensor<int32, [2]> var_209_strides_0 = const()[name = string("op_209_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 151 |
+
tensor<int32, [4]> var_209_pad_0 = const()[name = string("op_209_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 152 |
+
tensor<int32, [2]> var_209_dilations_0 = const()[name = string("op_209_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 153 |
+
int32 var_209_groups_0 = const()[name = string("op_209_groups_0"), val = int32(1)];
|
| 154 |
+
tensor<fp16, [2688, 896, 1, 1]> var_201_to_fp16 = const()[name = string("op_201_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(42328000)))];
|
| 155 |
+
tensor<fp16, [2, 2688, 1, 1]> var_209_cast_fp16 = conv(dilations = var_209_dilations_0, groups = var_209_groups_0, pad = var_209_pad_0, pad_type = var_209_pad_type_0, strides = var_209_strides_0, weight = var_201_to_fp16, x = x_3_cast_fp16)[name = string("op_209_cast_fp16")];
|
| 156 |
+
tensor<int32, [3]> var_210_split_sizes_0 = const()[name = string("op_210_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
|
| 157 |
+
int32 var_210_axis_0 = const()[name = string("op_210_axis_0"), val = int32(1)];
|
| 158 |
+
tensor<fp16, [2, 896, 1, 1]> var_210_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_210_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_210_cast_fp16_2 = split(axis = var_210_axis_0, split_sizes = var_210_split_sizes_0, x = var_209_cast_fp16)[name = string("op_210_cast_fp16")];
|
| 159 |
+
fp16 const_2_promoted_to_fp16 = const()[name = string("const_2_promoted_to_fp16"), val = fp16(-0x1p+0)];
|
| 160 |
+
tensor<fp16, [2, 896, 1, 1]> var_215_cast_fp16 = mul(x = x_33_cast_fp16, y = const_2_promoted_to_fp16)[name = string("op_215_cast_fp16")];
|
| 161 |
+
bool x_35_interleave_0 = const()[name = string("x_35_interleave_0"), val = bool(false)];
|
| 162 |
+
tensor<fp16, [2, 1792, 1, 1]> x_35_cast_fp16 = concat(axis = var_199, interleave = x_35_interleave_0, values = (x_33_cast_fp16, var_215_cast_fp16))[name = string("x_35_cast_fp16")];
|
| 163 |
+
tensor<int32, [1]> out_13_axes_0 = const()[name = string("out_13_axes_0"), val = tensor<int32, [1]>([1])];
|
| 164 |
+
fp16 var_225_to_fp16 = const()[name = string("op_225_to_fp16"), val = fp16(0x1.5p-17)];
|
| 165 |
+
tensor<fp16, [2, 1792, 1, 1]> out_13_cast_fp16 = layer_norm(axes = out_13_axes_0, epsilon = var_225_to_fp16, x = x_35_cast_fp16)[name = string("out_13_cast_fp16")];
|
| 166 |
+
tensor<fp16, [1, 1792, 1, 1]> layer_layers_2_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_2_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(47144960)))];
|
| 167 |
+
tensor<fp16, [2, 1792, 1, 1]> out_15_cast_fp16 = mul(x = out_13_cast_fp16, y = layer_layers_2_layer_norm_weight_to_fp16)[name = string("out_15_cast_fp16")];
|
| 168 |
+
tensor<int32, [2]> var_231_split_sizes_0 = const()[name = string("op_231_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
|
| 169 |
+
int32 var_231_axis_0 = const()[name = string("op_231_axis_0"), val = int32(1)];
|
| 170 |
+
tensor<fp16, [2, 896, 1, 1]> var_231_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_231_cast_fp16_1 = split(axis = var_231_axis_0, split_sizes = var_231_split_sizes_0, x = out_15_cast_fp16)[name = string("op_231_cast_fp16")];
|
| 171 |
+
fp16 var_234_promoted_to_fp16 = const()[name = string("op_234_promoted_to_fp16"), val = fp16(0x1p+0)];
|
| 172 |
+
tensor<fp16, [2, 896, 1, 1]> var_235_cast_fp16 = add(x = var_210_cast_fp16_1, y = var_234_promoted_to_fp16)[name = string("op_235_cast_fp16")];
|
| 173 |
+
tensor<fp16, [2, 896, 1, 1]> var_236_cast_fp16 = mul(x = var_231_cast_fp16_0, y = var_235_cast_fp16)[name = string("op_236_cast_fp16")];
|
| 174 |
+
tensor<fp16, [2, 896, 1, 1]> x_41_cast_fp16 = add(x = var_236_cast_fp16, y = var_210_cast_fp16_0)[name = string("x_41_cast_fp16")];
|
| 175 |
+
string input_9_pad_type_0 = const()[name = string("input_9_pad_type_0"), val = string("valid")];
|
| 176 |
+
tensor<int32, [2]> input_9_strides_0 = const()[name = string("input_9_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 177 |
+
tensor<int32, [4]> input_9_pad_0 = const()[name = string("input_9_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 178 |
+
tensor<int32, [2]> input_9_dilations_0 = const()[name = string("input_9_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 179 |
+
int32 input_9_groups_0 = const()[name = string("input_9_groups_0"), val = int32(1)];
|
| 180 |
+
tensor<fp16, [2688, 896, 1, 1]> var_192_to_fp16 = const()[name = string("op_192_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(47148608)))];
|
| 181 |
+
tensor<fp16, [2, 2688, 1, 1]> input_9_cast_fp16 = conv(dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = var_192_to_fp16, x = x_41_cast_fp16)[name = string("input_9_cast_fp16")];
|
| 182 |
+
string up_5_pad_type_0 = const()[name = string("up_5_pad_type_0"), val = string("valid")];
|
| 183 |
+
tensor<int32, [2]> up_5_strides_0 = const()[name = string("up_5_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 184 |
+
tensor<int32, [4]> up_5_pad_0 = const()[name = string("up_5_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 185 |
+
tensor<int32, [2]> up_5_dilations_0 = const()[name = string("up_5_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 186 |
+
int32 up_5_groups_0 = const()[name = string("up_5_groups_0"), val = int32(1)];
|
| 187 |
+
tensor<fp16, [2688, 896, 1, 1]> var_193_to_fp16 = const()[name = string("op_193_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(51965568)))];
|
| 188 |
+
tensor<fp16, [2, 2688, 1, 1]> up_5_cast_fp16 = conv(dilations = up_5_dilations_0, groups = up_5_groups_0, pad = up_5_pad_0, pad_type = up_5_pad_type_0, strides = up_5_strides_0, weight = var_193_to_fp16, x = x_41_cast_fp16)[name = string("up_5_cast_fp16")];
|
| 189 |
+
tensor<fp16, [2, 2688, 1, 1]> gate_5_cast_fp16 = silu(x = input_9_cast_fp16)[name = string("gate_5_cast_fp16")];
|
| 190 |
+
tensor<fp16, [2, 2688, 1, 1]> x_43_cast_fp16 = mul(x = gate_5_cast_fp16, y = up_5_cast_fp16)[name = string("x_43_cast_fp16")];
|
| 191 |
+
string var_254_pad_type_0 = const()[name = string("op_254_pad_type_0"), val = string("valid")];
|
| 192 |
+
tensor<int32, [2]> var_254_strides_0 = const()[name = string("op_254_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 193 |
+
tensor<int32, [4]> var_254_pad_0 = const()[name = string("op_254_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 194 |
+
tensor<int32, [2]> var_254_dilations_0 = const()[name = string("op_254_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 195 |
+
int32 var_254_groups_0 = const()[name = string("op_254_groups_0"), val = int32(1)];
|
| 196 |
+
tensor<fp16, [896, 2688, 1, 1]> var_194_to_fp16 = const()[name = string("op_194_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(56782528)))];
|
| 197 |
+
tensor<fp16, [2, 896, 1, 1]> var_254_cast_fp16 = conv(dilations = var_254_dilations_0, groups = var_254_groups_0, pad = var_254_pad_0, pad_type = var_254_pad_type_0, strides = var_254_strides_0, weight = var_194_to_fp16, x = x_43_cast_fp16)[name = string("op_254_cast_fp16")];
|
| 198 |
+
tensor<fp16, [2, 896, 1, 1]> var_255_cast_fp16 = mul(x = var_210_cast_fp16_2, y = var_254_cast_fp16)[name = string("op_255_cast_fp16")];
|
| 199 |
+
tensor<fp16, [2, 896, 1, 1]> x_47_cast_fp16 = add(x = x_33_cast_fp16, y = var_255_cast_fp16)[name = string("x_47_cast_fp16")];
|
| 200 |
+
int32 var_264 = const()[name = string("op_264"), val = int32(1)];
|
| 201 |
+
string var_274_pad_type_0 = const()[name = string("op_274_pad_type_0"), val = string("valid")];
|
| 202 |
+
tensor<int32, [2]> var_274_strides_0 = const()[name = string("op_274_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 203 |
+
tensor<int32, [4]> var_274_pad_0 = const()[name = string("op_274_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 204 |
+
tensor<int32, [2]> var_274_dilations_0 = const()[name = string("op_274_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 205 |
+
int32 var_274_groups_0 = const()[name = string("op_274_groups_0"), val = int32(1)];
|
| 206 |
+
tensor<fp16, [2688, 896, 1, 1]> var_266_to_fp16 = const()[name = string("op_266_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(61599488)))];
|
| 207 |
+
tensor<fp16, [2, 2688, 1, 1]> var_274_cast_fp16 = conv(dilations = var_274_dilations_0, groups = var_274_groups_0, pad = var_274_pad_0, pad_type = var_274_pad_type_0, strides = var_274_strides_0, weight = var_266_to_fp16, x = x_3_cast_fp16)[name = string("op_274_cast_fp16")];
|
| 208 |
+
tensor<int32, [3]> var_275_split_sizes_0 = const()[name = string("op_275_split_sizes_0"), val = tensor<int32, [3]>([896, 896, 896])];
|
| 209 |
+
int32 var_275_axis_0 = const()[name = string("op_275_axis_0"), val = int32(1)];
|
| 210 |
+
tensor<fp16, [2, 896, 1, 1]> var_275_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_275_cast_fp16_1, tensor<fp16, [2, 896, 1, 1]> var_275_cast_fp16_2 = split(axis = var_275_axis_0, split_sizes = var_275_split_sizes_0, x = var_274_cast_fp16)[name = string("op_275_cast_fp16")];
|
| 211 |
+
fp16 const_3_promoted_to_fp16 = const()[name = string("const_3_promoted_to_fp16"), val = fp16(-0x1p+0)];
|
| 212 |
+
tensor<fp16, [2, 896, 1, 1]> var_280_cast_fp16 = mul(x = x_47_cast_fp16, y = const_3_promoted_to_fp16)[name = string("op_280_cast_fp16")];
|
| 213 |
+
bool x_49_interleave_0 = const()[name = string("x_49_interleave_0"), val = bool(false)];
|
| 214 |
+
tensor<fp16, [2, 1792, 1, 1]> x_49_cast_fp16 = concat(axis = var_264, interleave = x_49_interleave_0, values = (x_47_cast_fp16, var_280_cast_fp16))[name = string("x_49_cast_fp16")];
|
| 215 |
+
tensor<int32, [1]> out_19_axes_0 = const()[name = string("out_19_axes_0"), val = tensor<int32, [1]>([1])];
|
| 216 |
+
fp16 var_290_to_fp16 = const()[name = string("op_290_to_fp16"), val = fp16(0x1.5p-17)];
|
| 217 |
+
tensor<fp16, [2, 1792, 1, 1]> out_19_cast_fp16 = layer_norm(axes = out_19_axes_0, epsilon = var_290_to_fp16, x = x_49_cast_fp16)[name = string("out_19_cast_fp16")];
|
| 218 |
+
tensor<fp16, [1, 1792, 1, 1]> layer_layers_3_layer_norm_weight_to_fp16 = const()[name = string("layer_layers_3_layer_norm_weight_to_fp16"), val = tensor<fp16, [1, 1792, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(66416448)))];
|
| 219 |
+
tensor<fp16, [2, 1792, 1, 1]> out_21_cast_fp16 = mul(x = out_19_cast_fp16, y = layer_layers_3_layer_norm_weight_to_fp16)[name = string("out_21_cast_fp16")];
|
| 220 |
+
tensor<int32, [2]> var_296_split_sizes_0 = const()[name = string("op_296_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
|
| 221 |
+
int32 var_296_axis_0 = const()[name = string("op_296_axis_0"), val = int32(1)];
|
| 222 |
+
tensor<fp16, [2, 896, 1, 1]> var_296_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_296_cast_fp16_1 = split(axis = var_296_axis_0, split_sizes = var_296_split_sizes_0, x = out_21_cast_fp16)[name = string("op_296_cast_fp16")];
|
| 223 |
+
fp16 var_299_promoted_to_fp16 = const()[name = string("op_299_promoted_to_fp16"), val = fp16(0x1p+0)];
|
| 224 |
+
tensor<fp16, [2, 896, 1, 1]> var_300_cast_fp16 = add(x = var_275_cast_fp16_1, y = var_299_promoted_to_fp16)[name = string("op_300_cast_fp16")];
|
| 225 |
+
tensor<fp16, [2, 896, 1, 1]> var_301_cast_fp16 = mul(x = var_296_cast_fp16_0, y = var_300_cast_fp16)[name = string("op_301_cast_fp16")];
|
| 226 |
+
tensor<fp16, [2, 896, 1, 1]> x_55_cast_fp16 = add(x = var_301_cast_fp16, y = var_275_cast_fp16_0)[name = string("x_55_cast_fp16")];
|
| 227 |
+
string input_pad_type_0 = const()[name = string("input_pad_type_0"), val = string("valid")];
|
| 228 |
+
tensor<int32, [2]> input_strides_0 = const()[name = string("input_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 229 |
+
tensor<int32, [4]> input_pad_0 = const()[name = string("input_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 230 |
+
tensor<int32, [2]> input_dilations_0 = const()[name = string("input_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 231 |
+
int32 input_groups_0 = const()[name = string("input_groups_0"), val = int32(1)];
|
| 232 |
+
tensor<fp16, [2688, 896, 1, 1]> var_257_to_fp16 = const()[name = string("op_257_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(66420096)))];
|
| 233 |
+
tensor<fp16, [2, 2688, 1, 1]> input_cast_fp16 = conv(dilations = input_dilations_0, groups = input_groups_0, pad = input_pad_0, pad_type = input_pad_type_0, strides = input_strides_0, weight = var_257_to_fp16, x = x_55_cast_fp16)[name = string("input_cast_fp16")];
|
| 234 |
+
string up_pad_type_0 = const()[name = string("up_pad_type_0"), val = string("valid")];
|
| 235 |
+
tensor<int32, [2]> up_strides_0 = const()[name = string("up_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 236 |
+
tensor<int32, [4]> up_pad_0 = const()[name = string("up_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 237 |
+
tensor<int32, [2]> up_dilations_0 = const()[name = string("up_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 238 |
+
int32 up_groups_0 = const()[name = string("up_groups_0"), val = int32(1)];
|
| 239 |
+
tensor<fp16, [2688, 896, 1, 1]> var_258_to_fp16 = const()[name = string("op_258_to_fp16"), val = tensor<fp16, [2688, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(71237056)))];
|
| 240 |
+
tensor<fp16, [2, 2688, 1, 1]> up_cast_fp16 = conv(dilations = up_dilations_0, groups = up_groups_0, pad = up_pad_0, pad_type = up_pad_type_0, strides = up_strides_0, weight = var_258_to_fp16, x = x_55_cast_fp16)[name = string("up_cast_fp16")];
|
| 241 |
+
tensor<fp16, [2, 2688, 1, 1]> gate_cast_fp16 = silu(x = input_cast_fp16)[name = string("gate_cast_fp16")];
|
| 242 |
+
tensor<fp16, [2, 2688, 1, 1]> x_57_cast_fp16 = mul(x = gate_cast_fp16, y = up_cast_fp16)[name = string("x_57_cast_fp16")];
|
| 243 |
+
string var_319_pad_type_0 = const()[name = string("op_319_pad_type_0"), val = string("valid")];
|
| 244 |
+
tensor<int32, [2]> var_319_strides_0 = const()[name = string("op_319_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 245 |
+
tensor<int32, [4]> var_319_pad_0 = const()[name = string("op_319_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 246 |
+
tensor<int32, [2]> var_319_dilations_0 = const()[name = string("op_319_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 247 |
+
int32 var_319_groups_0 = const()[name = string("op_319_groups_0"), val = int32(1)];
|
| 248 |
+
tensor<fp16, [896, 2688, 1, 1]> var_259_to_fp16 = const()[name = string("op_259_to_fp16"), val = tensor<fp16, [896, 2688, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(76054016)))];
|
| 249 |
+
tensor<fp16, [2, 896, 1, 1]> var_319_cast_fp16 = conv(dilations = var_319_dilations_0, groups = var_319_groups_0, pad = var_319_pad_0, pad_type = var_319_pad_type_0, strides = var_319_strides_0, weight = var_259_to_fp16, x = x_57_cast_fp16)[name = string("op_319_cast_fp16")];
|
| 250 |
+
tensor<fp16, [2, 896, 1, 1]> var_320_cast_fp16 = mul(x = var_275_cast_fp16_2, y = var_319_cast_fp16)[name = string("op_320_cast_fp16")];
|
| 251 |
+
tensor<fp16, [2, 896, 1, 1]> x_61_cast_fp16 = add(x = x_47_cast_fp16, y = var_320_cast_fp16)[name = string("x_61_cast_fp16")];
|
| 252 |
+
int32 var_327 = const()[name = string("op_327"), val = int32(1)];
|
| 253 |
+
string var_335_pad_type_0 = const()[name = string("op_335_pad_type_0"), val = string("valid")];
|
| 254 |
+
tensor<int32, [2]> var_335_strides_0 = const()[name = string("op_335_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 255 |
+
tensor<int32, [4]> var_335_pad_0 = const()[name = string("op_335_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 256 |
+
tensor<int32, [2]> var_335_dilations_0 = const()[name = string("op_335_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 257 |
+
int32 var_335_groups_0 = const()[name = string("op_335_groups_0"), val = int32(1)];
|
| 258 |
+
tensor<fp16, [1792, 896, 1, 1]> var_329_to_fp16 = const()[name = string("op_329_to_fp16"), val = tensor<fp16, [1792, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(80870976)))];
|
| 259 |
+
tensor<fp16, [2, 1792, 1, 1]> var_335_cast_fp16 = conv(dilations = var_335_dilations_0, groups = var_335_groups_0, pad = var_335_pad_0, pad_type = var_335_pad_type_0, strides = var_335_strides_0, weight = var_329_to_fp16, x = x_3_cast_fp16)[name = string("op_335_cast_fp16")];
|
| 260 |
+
tensor<int32, [2]> var_336_split_sizes_0 = const()[name = string("op_336_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
|
| 261 |
+
int32 var_336_axis_0 = const()[name = string("op_336_axis_0"), val = int32(1)];
|
| 262 |
+
tensor<fp16, [2, 896, 1, 1]> var_336_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_336_cast_fp16_1 = split(axis = var_336_axis_0, split_sizes = var_336_split_sizes_0, x = var_335_cast_fp16)[name = string("op_336_cast_fp16")];
|
| 263 |
+
fp16 const_4_promoted_to_fp16 = const()[name = string("const_4_promoted_to_fp16"), val = fp16(-0x1p+0)];
|
| 264 |
+
tensor<fp16, [2, 896, 1, 1]> var_338_cast_fp16 = mul(x = x_61_cast_fp16, y = const_4_promoted_to_fp16)[name = string("op_338_cast_fp16")];
|
| 265 |
+
bool x_63_interleave_0 = const()[name = string("x_63_interleave_0"), val = bool(false)];
|
| 266 |
+
tensor<fp16, [2, 1792, 1, 1]> x_63_cast_fp16 = concat(axis = var_327, interleave = x_63_interleave_0, values = (x_61_cast_fp16, var_338_cast_fp16))[name = string("x_63_cast_fp16")];
|
| 267 |
+
tensor<int32, [1]> out_25_axes_0 = const()[name = string("out_25_axes_0"), val = tensor<int32, [1]>([1])];
|
| 268 |
+
fp16 var_348_to_fp16 = const()[name = string("op_348_to_fp16"), val = fp16(0x1.5p-17)];
|
| 269 |
+
tensor<fp16, [2, 1792, 1, 1]> out_25_cast_fp16 = layer_norm(axes = out_25_axes_0, epsilon = var_348_to_fp16, x = x_63_cast_fp16)[name = string("out_25_cast_fp16")];
|
| 270 |
+
tensor<int32, [2]> var_352_split_sizes_0 = const()[name = string("op_352_split_sizes_0"), val = tensor<int32, [2]>([896, 896])];
|
| 271 |
+
int32 var_352_axis_0 = const()[name = string("op_352_axis_0"), val = int32(1)];
|
| 272 |
+
tensor<fp16, [2, 896, 1, 1]> var_352_cast_fp16_0, tensor<fp16, [2, 896, 1, 1]> var_352_cast_fp16_1 = split(axis = var_352_axis_0, split_sizes = var_352_split_sizes_0, x = out_25_cast_fp16)[name = string("op_352_cast_fp16")];
|
| 273 |
+
fp16 var_355_promoted_to_fp16 = const()[name = string("op_355_promoted_to_fp16"), val = fp16(0x1p+0)];
|
| 274 |
+
tensor<fp16, [2, 896, 1, 1]> var_356_cast_fp16 = add(x = var_336_cast_fp16_1, y = var_355_promoted_to_fp16)[name = string("op_356_cast_fp16")];
|
| 275 |
+
tensor<fp16, [2, 896, 1, 1]> var_357_cast_fp16 = mul(x = var_352_cast_fp16_0, y = var_356_cast_fp16)[name = string("op_357_cast_fp16")];
|
| 276 |
+
tensor<fp16, [2, 896, 1, 1]> x_cast_fp16 = add(x = var_357_cast_fp16, y = var_336_cast_fp16_0)[name = string("x_cast_fp16")];
|
| 277 |
+
string var_363_pad_type_0 = const()[name = string("op_363_pad_type_0"), val = string("valid")];
|
| 278 |
+
tensor<int32, [2]> var_363_strides_0 = const()[name = string("op_363_strides_0"), val = tensor<int32, [2]>([1, 1])];
|
| 279 |
+
tensor<int32, [4]> var_363_pad_0 = const()[name = string("op_363_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
|
| 280 |
+
tensor<int32, [2]> var_363_dilations_0 = const()[name = string("op_363_dilations_0"), val = tensor<int32, [2]>([1, 1])];
|
| 281 |
+
int32 var_363_groups_0 = const()[name = string("op_363_groups_0"), val = int32(1)];
|
| 282 |
+
tensor<fp16, [64, 896, 1, 1]> var_322_to_fp16 = const()[name = string("op_322_to_fp16"), val = tensor<fp16, [64, 896, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(84082304)))];
|
| 283 |
+
tensor<fp16, [2, 64, 1, 1]> predicted_noise = conv(dilations = var_363_dilations_0, groups = var_363_groups_0, pad = var_363_pad_0, pad_type = var_363_pad_type_0, strides = var_363_strides_0, weight = var_322_to_fp16, x = x_cast_fp16)[name = string("op_363_cast_fp16")];
|
| 284 |
+
} -> (predicted_noise);
|
| 285 |
+
}
|
diffusion_head_model.mlmodelc/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d769d4dab4ad19b6330272e794f83af1e2551912f58ea4ab84e84c09fe5220e6
|
| 3 |
+
size 84197056
|
tts_eos_classifier.mlmodelc/analytics/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:04f65dee434405e684dd155f8a97affdc668a2ade012a1583813ccc88a1e44ba
|
| 3 |
+
size 243
|
tts_eos_classifier.mlmodelc/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dca3bc2766c5c819eb3434e15d8a6229ee1a4243da8a91b1f339bdd69fc0da8d
|
| 3 |
+
size 311
|
tts_eos_classifier.mlmodelc/model.mil
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
program(1.3)
|
| 2 |
+
[buildInfo = dict<string, string>({{"coremlc-component-MIL", "3405.2.1"}, {"coremlc-version", "3405.2.1"}, {"coremltools-component-milinternal", ""}, {"coremltools-version", "9.0"}})]
|
| 3 |
+
{
|
| 4 |
+
func main<ios18>(tensor<fp16, [1, 896, 1]> input_x) {
|
| 5 |
+
string var_14_pad_type_0 = const()[name = string("op_14_pad_type_0"), val = string("valid")];
|
| 6 |
+
tensor<int32, [1]> var_14_strides_0 = const()[name = string("op_14_strides_0"), val = tensor<int32, [1]>([1])];
|
| 7 |
+
tensor<int32, [2]> var_14_pad_0 = const()[name = string("op_14_pad_0"), val = tensor<int32, [2]>([0, 0])];
|
| 8 |
+
tensor<int32, [1]> var_14_dilations_0 = const()[name = string("op_14_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 9 |
+
int32 var_14_groups_0 = const()[name = string("op_14_groups_0"), val = int32(1)];
|
| 10 |
+
tensor<fp16, [896, 896, 1]> var_8_to_fp16 = const()[name = string("op_8_to_fp16"), val = tensor<fp16, [896, 896, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
|
| 11 |
+
tensor<fp16, [896]> fc1_bias_to_fp16 = const()[name = string("fc1_bias_to_fp16"), val = tensor<fp16, [896]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1605760)))];
|
| 12 |
+
tensor<fp16, [1, 896, 1]> var_14_cast_fp16 = conv(bias = fc1_bias_to_fp16, dilations = var_14_dilations_0, groups = var_14_groups_0, pad = var_14_pad_0, pad_type = var_14_pad_type_0, strides = var_14_strides_0, weight = var_8_to_fp16, x = input_x)[name = string("op_14_cast_fp16")];
|
| 13 |
+
tensor<fp16, [1, 896, 1]> x_cast_fp16 = relu(x = var_14_cast_fp16)[name = string("x_cast_fp16")];
|
| 14 |
+
string var_26_pad_type_0 = const()[name = string("op_26_pad_type_0"), val = string("valid")];
|
| 15 |
+
tensor<int32, [1]> var_26_strides_0 = const()[name = string("op_26_strides_0"), val = tensor<int32, [1]>([1])];
|
| 16 |
+
tensor<int32, [2]> var_26_pad_0 = const()[name = string("op_26_pad_0"), val = tensor<int32, [2]>([0, 0])];
|
| 17 |
+
tensor<int32, [1]> var_26_dilations_0 = const()[name = string("op_26_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 18 |
+
int32 var_26_groups_0 = const()[name = string("op_26_groups_0"), val = int32(1)];
|
| 19 |
+
tensor<fp16, [1, 896, 1]> var_20_to_fp16 = const()[name = string("op_20_to_fp16"), val = tensor<fp16, [1, 896, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1607616)))];
|
| 20 |
+
tensor<fp16, [1]> fc2_bias_to_fp16 = const()[name = string("fc2_bias_to_fp16"), val = tensor<fp16, [1]>([-0x1.36p-5])];
|
| 21 |
+
tensor<fp16, [1, 1, 1]> output = conv(bias = fc2_bias_to_fp16, dilations = var_26_dilations_0, groups = var_26_groups_0, pad = var_26_pad_0, pad_type = var_26_pad_type_0, strides = var_26_strides_0, weight = var_20_to_fp16, x = x_cast_fp16)[name = string("op_26_cast_fp16")];
|
| 22 |
+
} -> (output);
|
| 23 |
+
}
|
tts_eos_classifier.mlmodelc/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dea5cf930bfc270775854b5e4fab3f5214f3c366f2b09bb0f6cc6293bffe8f9b
|
| 3 |
+
size 1609472
|
tts_input_types.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:af6397a066bc69002280f54fa1bec73be40c7e8f6f671c41f8215e61a0cc0bf6
|
| 3 |
+
size 3712
|
vibe_voice_lm_model_seqlen_32.mlmodelc/analytics/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:747917dedac470617b6eca97432d5a687327f8948b8fc8d42a1834d3e03ecf10
|
| 3 |
+
size 243
|
vibe_voice_lm_model_seqlen_32.mlmodelc/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b99dc1e16c07edb47e0de2d9afa8227c1f487587c22524f6fa7283f1250f12f2
|
| 3 |
+
size 411
|
vibe_voice_lm_model_seqlen_32.mlmodelc/model.mil
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
vibe_voice_lm_model_seqlen_32.mlmodelc/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ce755198818c1db103eca266d8da9c9a129794f4ad325f7f0db5b2fd5dded0f3
|
| 3 |
+
size 119980544
|
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/analytics/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f8fef128a776b803c235e8bfb7dfe6a9215a4fd76ce36eb2292130b4ab8b20f
|
| 3 |
+
size 243
|
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3059be1bffdefb805a387df47363f3c7aa736c02ddc37570e3eae1d4bc501ef2
|
| 3 |
+
size 470
|
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/model.mil
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
vibe_voice_tts_lm_model_seqlen_8.mlmodelc/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41c092353dfa7572f45a72dc46fb0abcdca886fd57eaef65026d73222cacbfa8
|
| 3 |
+
size 597151616
|
vibevoice_embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:409151256820df5c3c5f816050bbd8cf86cfc075758002696a0459ae2ca23ce4
|
| 3 |
+
size 272269440
|
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/analytics/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cd2c98a30416ec110d7722a2736b0076032da7d43596317a1ca4da37db83798b
|
| 3 |
+
size 243
|
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/coremldata.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1353c852f5600d7562f05a51334bc13145b0ef80add2ac966e8e761d27f652ff
|
| 3 |
+
size 575
|
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/model.mil
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
vibevoice_tts_lm_model_fused_seqlen_8.mlmodelc/weights/weight.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ab2f84b7297c9c83434cac17a10fa88fab8b41dbde027d163b542ab02fae851
|
| 3 |
+
size 597151616
|
voices/streaming_model/de-Spk0_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bc4a02421dac7523c29cca7fac9327bd41ba276bbb35c781480c933b316a1a8a
|
| 3 |
+
size 4381274
|
voices/streaming_model/de-Spk1_woman.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2a0d89ba9b70498ddbec0ef1bcebacf730237721b56965de0bf4a50ecedace0e
|
| 3 |
+
size 3337491
|
voices/streaming_model/en-Carter_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d06a3c578fb7d3285fd82004ff7de76ef9806b8a23033180ab1ae42fdd7c8b0c
|
| 3 |
+
size 2685462
|
voices/streaming_model/en-Davis_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4c322bfd7c95cfd37e790af2c3c1a467b51ac11c71dee28a137dba34d3dad80
|
| 3 |
+
size 1560939
|
voices/streaming_model/en-Emma_woman.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5555dadabb0e316c14236f8d30c9133337875c95c8ab0921713ef2a74c1116d6
|
| 3 |
+
size 2118241
|
voices/streaming_model/en-Frank_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2da2d1fe940ca0beba5d099ef0d0eacb671a677e7fc30657f4f7eb62cda7cca0
|
| 3 |
+
size 2133945
|
voices/streaming_model/en-Grace_woman.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9fc683ca87553d3a8d583a58f27ccc25c4ecd87847c1e8512bc1237d6b5f2cb0
|
| 3 |
+
size 1761680
|
voices/streaming_model/en-Mike_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dfa43cd27ce326b1ef6abd1fd72a25eefad15b1d8fefebf6b120572e6786398b
|
| 3 |
+
size 1266668
|
voices/streaming_model/fr-Spk0_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:64916486f0e7caefe421f042ecf6c74fe5fb9c6c98a7081f26a05e74d82aa2e7
|
| 3 |
+
size 2753069
|
voices/streaming_model/fr-Spk1_woman.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e587c944f6a7cd44eb7ea311f99e614154bedb47c505574dc8685d60eb6f0c92
|
| 3 |
+
size 2663208
|
voices/streaming_model/in-Samuel_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bbdb9168868f2aa90ba8197b95804c274cdf9347922b07ecd08b517132c7d428
|
| 3 |
+
size 2398021
|
voices/streaming_model/it-Spk0_woman.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b84ebbff0cb296751b6bbbb147f2b30a365c2e6e2fcd02e24d5400fd4c322172
|
| 3 |
+
size 1612206
|
voices/streaming_model/it-Spk1_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f2c27e97d68778d60f51658f9de701c93b5003530bc2b19d3dcc3cffa94c67e
|
| 3 |
+
size 1798452
|
voices/streaming_model/jp-Spk0_man.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b55a4392e9f2e171af6a12e88c9646535e7b22fb2edefb90fcfd5eea9cb0b67
|
| 3 |
+
size 2923013
|
voices/streaming_model/jp-Spk1_woman.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4501729bac7b99978000940380d473a68a1432610f805a153deae76f332fbfba
|
| 3 |
+
size 2908957
|
voices/streaming_model/kr-Spk0_woman.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:09208fda6c5a33c59e7eef7c171580406a4c584d31512d7270badf7578a921b4
|
| 3 |
+
size 2600985
|