Delete silero_vad_stateful_4bit.mlmodelc
Browse files
silero_vad_stateful_4bit.mlmodelc/analytics/coremldata.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:67d19ac70e21c7f01639678563b5cf6eb89bf4b15cbb7280b84dc204f30bfcbc
|
| 3 |
-
size 243
|
|
|
|
|
|
|
|
|
|
|
|
silero_vad_stateful_4bit.mlmodelc/coremldata.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:1d3c298161f5be74328c6c0253a12800e7a50b9a0b3c9342744eb78b3e4aef0a
|
| 3 |
-
size 788
|
|
|
|
|
|
|
|
|
|
|
|
silero_vad_stateful_4bit.mlmodelc/metadata.json
DELETED
|
@@ -1,124 +0,0 @@
|
|
| 1 |
-
[
|
| 2 |
-
{
|
| 3 |
-
"shortDescription" : "Production Silero VAD with LSTM states and full architecture",
|
| 4 |
-
"metadataOutputVersion" : "3.0",
|
| 5 |
-
"outputSchema" : [
|
| 6 |
-
{
|
| 7 |
-
"hasShapeFlexibility" : "0",
|
| 8 |
-
"isOptional" : "0",
|
| 9 |
-
"dataType" : "Float32",
|
| 10 |
-
"formattedType" : "MultiArray (Float32 1 × 1)",
|
| 11 |
-
"shortDescription" : "Voice activity probability (0.0-1.0)",
|
| 12 |
-
"shape" : "[1, 1]",
|
| 13 |
-
"name" : "vad_probability",
|
| 14 |
-
"type" : "MultiArray"
|
| 15 |
-
},
|
| 16 |
-
{
|
| 17 |
-
"hasShapeFlexibility" : "0",
|
| 18 |
-
"isOptional" : "0",
|
| 19 |
-
"dataType" : "Float32",
|
| 20 |
-
"formattedType" : "MultiArray (Float32 1 × 1 × 128)",
|
| 21 |
-
"shortDescription" : "Updated LSTM hidden state for next inference",
|
| 22 |
-
"shape" : "[1, 1, 128]",
|
| 23 |
-
"name" : "h_state_out",
|
| 24 |
-
"type" : "MultiArray"
|
| 25 |
-
},
|
| 26 |
-
{
|
| 27 |
-
"hasShapeFlexibility" : "0",
|
| 28 |
-
"isOptional" : "0",
|
| 29 |
-
"dataType" : "Float32",
|
| 30 |
-
"formattedType" : "MultiArray (Float32 1 × 1 × 128)",
|
| 31 |
-
"shortDescription" : "Updated LSTM cell state for next inference",
|
| 32 |
-
"shape" : "[1, 1, 128]",
|
| 33 |
-
"name" : "c_state_out",
|
| 34 |
-
"type" : "MultiArray"
|
| 35 |
-
}
|
| 36 |
-
],
|
| 37 |
-
"version" : "3.0",
|
| 38 |
-
"modelParameters" : [
|
| 39 |
-
|
| 40 |
-
],
|
| 41 |
-
"author" : "Trained Stateful Silero VAD",
|
| 42 |
-
"specificationVersion" : 8,
|
| 43 |
-
"storagePrecision" : "Mixed (Float16, Palettized (4 bits))",
|
| 44 |
-
"mlProgramOperationTypeHistogram" : {
|
| 45 |
-
"Ios17.square" : 6,
|
| 46 |
-
"Ios17.reshape" : 8,
|
| 47 |
-
"Ios16.reduceMean" : 5,
|
| 48 |
-
"Ios17.matmul" : 6,
|
| 49 |
-
"Ios17.transpose" : 9,
|
| 50 |
-
"Ios17.expandDims" : 2,
|
| 51 |
-
"Ios17.add" : 3,
|
| 52 |
-
"Ios16.sigmoid" : 5,
|
| 53 |
-
"Ios17.sliceByIndex" : 3,
|
| 54 |
-
"Ios17.squeeze" : 3,
|
| 55 |
-
"Ios17.gather" : 6,
|
| 56 |
-
"Ios17.layerNorm" : 1,
|
| 57 |
-
"Ios17.sqrt" : 3,
|
| 58 |
-
"Ios17.conv" : 5,
|
| 59 |
-
"Ios16.constexprLutToDense" : 8,
|
| 60 |
-
"Ios16.relu" : 9,
|
| 61 |
-
"Ios17.lstm" : 1,
|
| 62 |
-
"Ios17.linear" : 8,
|
| 63 |
-
"Stack" : 1,
|
| 64 |
-
"Ios17.concat" : 4,
|
| 65 |
-
"Ios17.cast" : 6,
|
| 66 |
-
"Ios17.mul" : 11
|
| 67 |
-
},
|
| 68 |
-
"computePrecision" : "Mixed (Float16, Float32, Int32, UInt16)",
|
| 69 |
-
"stateSchema" : [
|
| 70 |
-
|
| 71 |
-
],
|
| 72 |
-
"isUpdatable" : "0",
|
| 73 |
-
"availability" : {
|
| 74 |
-
"macOS" : "14.0",
|
| 75 |
-
"tvOS" : "17.0",
|
| 76 |
-
"visionOS" : "1.0",
|
| 77 |
-
"watchOS" : "10.0",
|
| 78 |
-
"iOS" : "17.0",
|
| 79 |
-
"macCatalyst" : "17.0"
|
| 80 |
-
},
|
| 81 |
-
"modelType" : {
|
| 82 |
-
"name" : "MLModelType_mlProgram"
|
| 83 |
-
},
|
| 84 |
-
"inputSchema" : [
|
| 85 |
-
{
|
| 86 |
-
"hasShapeFlexibility" : "0",
|
| 87 |
-
"isOptional" : "0",
|
| 88 |
-
"dataType" : "Float32",
|
| 89 |
-
"formattedType" : "MultiArray (Float32 1 × 512)",
|
| 90 |
-
"shortDescription" : "32ms audio chunk (512 samples at 16kHz)",
|
| 91 |
-
"shape" : "[1, 512]",
|
| 92 |
-
"name" : "audio_chunk",
|
| 93 |
-
"type" : "MultiArray"
|
| 94 |
-
},
|
| 95 |
-
{
|
| 96 |
-
"hasShapeFlexibility" : "0",
|
| 97 |
-
"isOptional" : "0",
|
| 98 |
-
"dataType" : "Float32",
|
| 99 |
-
"formattedType" : "MultiArray (Float32 1 × 1 × 128)",
|
| 100 |
-
"shortDescription" : "LSTM hidden state from previous inference",
|
| 101 |
-
"shape" : "[1, 1, 128]",
|
| 102 |
-
"name" : "h_state_in",
|
| 103 |
-
"type" : "MultiArray"
|
| 104 |
-
},
|
| 105 |
-
{
|
| 106 |
-
"hasShapeFlexibility" : "0",
|
| 107 |
-
"isOptional" : "0",
|
| 108 |
-
"dataType" : "Float32",
|
| 109 |
-
"formattedType" : "MultiArray (Float32 1 × 1 × 128)",
|
| 110 |
-
"shortDescription" : "LSTM cell state from previous inference",
|
| 111 |
-
"shape" : "[1, 1, 128]",
|
| 112 |
-
"name" : "c_state_in",
|
| 113 |
-
"type" : "MultiArray"
|
| 114 |
-
}
|
| 115 |
-
],
|
| 116 |
-
"userDefinedMetadata" : {
|
| 117 |
-
"com.github.apple.coremltools.version" : "8.3.0",
|
| 118 |
-
"com.github.apple.coremltools.source_dialect" : "TorchScript",
|
| 119 |
-
"com.github.apple.coremltools.source" : "torch==2.5.0"
|
| 120 |
-
},
|
| 121 |
-
"generatedClassName" : "silero_vad_improved_stateful_4bit",
|
| 122 |
-
"method" : "predict"
|
| 123 |
-
}
|
| 124 |
-
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
silero_vad_stateful_4bit.mlmodelc/model.mil
DELETED
|
@@ -1,266 +0,0 @@
|
|
| 1 |
-
program(1.0)
|
| 2 |
-
[buildInfo = dict<tensor<string, []>, tensor<string, []>>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}})]
|
| 3 |
-
{
|
| 4 |
-
func main<ios17>(tensor<fp32, [1, 512]> audio_chunk, tensor<fp32, [1, 1, 128]> c_state_in, tensor<fp32, [1, 1, 128]> h_state_in) {
|
| 5 |
-
tensor<int32, [2]> var_20_begin_0 = const()[name = tensor<string, []>("op_20_begin_0"), val = tensor<int32, [2]>([0, 0])];
|
| 6 |
-
tensor<int32, [2]> var_20_end_0 = const()[name = tensor<string, []>("op_20_end_0"), val = tensor<int32, [2]>([1, 256])];
|
| 7 |
-
tensor<bool, [2]> var_20_end_mask_0 = const()[name = tensor<string, []>("op_20_end_mask_0"), val = tensor<bool, [2]>([true, false])];
|
| 8 |
-
tensor<string, []> audio_chunk_to_fp16_dtype_0 = const()[name = tensor<string, []>("audio_chunk_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
|
| 9 |
-
tensor<fp16, [1, 512]> audio_chunk_to_fp16 = cast(dtype = audio_chunk_to_fp16_dtype_0, x = audio_chunk)[name = tensor<string, []>("cast_5")];
|
| 10 |
-
tensor<fp16, [1, 256]> var_20_cast_fp16 = slice_by_index(begin = var_20_begin_0, end = var_20_end_0, end_mask = var_20_end_mask_0, x = audio_chunk_to_fp16)[name = tensor<string, []>("op_20_cast_fp16")];
|
| 11 |
-
tensor<fp16, [256]> window_to_fp16 = const()[name = tensor<string, []>("window_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(64)))];
|
| 12 |
-
tensor<fp16, [1, 256]> frame_1_cast_fp16 = mul(x = var_20_cast_fp16, y = window_to_fp16)[name = tensor<string, []>("frame_1_cast_fp16")];
|
| 13 |
-
tensor<bool, []> matmul_1_transpose_x_1 = const()[name = tensor<string, []>("matmul_1_transpose_x_1"), val = tensor<bool, []>(false)];
|
| 14 |
-
tensor<bool, []> matmul_1_transpose_y_1 = const()[name = tensor<string, []>("matmul_1_transpose_y_1"), val = tensor<bool, []>(true)];
|
| 15 |
-
tensor<fp16, [256, 256]> cos_0_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [32768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(640))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(33472))), name = tensor<string, []>("cos_0_to_fp16_palettized"), shape = tensor<uint32, [2]>([256, 256])];
|
| 16 |
-
tensor<fp16, [256, 1]> matmul_1_cast_fp16 = matmul(transpose_x = matmul_1_transpose_x_1, transpose_y = matmul_1_transpose_y_1, x = cos_0_to_fp16_palettized, y = frame_1_cast_fp16)[name = tensor<string, []>("matmul_1_cast_fp16")];
|
| 17 |
-
tensor<bool, []> matmul_3_transpose_x_1 = const()[name = tensor<string, []>("matmul_3_transpose_x_1"), val = tensor<bool, []>(false)];
|
| 18 |
-
tensor<bool, []> matmul_3_transpose_y_1 = const()[name = tensor<string, []>("matmul_3_transpose_y_1"), val = tensor<bool, []>(true)];
|
| 19 |
-
tensor<fp16, [256, 256]> sin_1_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [32768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(33600))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(66432))), name = tensor<string, []>("sin_1_to_fp16_palettized"), shape = tensor<uint32, [2]>([256, 256])];
|
| 20 |
-
tensor<fp16, [256, 1]> matmul_3_cast_fp16 = matmul(transpose_x = matmul_3_transpose_x_1, transpose_y = matmul_3_transpose_y_1, x = sin_1_to_fp16_palettized, y = frame_1_cast_fp16)[name = tensor<string, []>("matmul_3_cast_fp16")];
|
| 21 |
-
tensor<fp16, []> mul_2_y_0_to_fp16 = const()[name = tensor<string, []>("mul_2_y_0_to_fp16"), val = tensor<fp16, []>(-0x1p+0)];
|
| 22 |
-
tensor<fp16, [256, 1]> mul_2_cast_fp16 = mul(x = matmul_3_cast_fp16, y = mul_2_y_0_to_fp16)[name = tensor<string, []>("mul_2_cast_fp16")];
|
| 23 |
-
tensor<int32, [2]> transpose_2_perm_0 = const()[name = tensor<string, []>("transpose_2_perm_0"), val = tensor<int32, [2]>([-1, 0])];
|
| 24 |
-
tensor<int32, [2]> transpose_3_perm_0 = const()[name = tensor<string, []>("transpose_3_perm_0"), val = tensor<int32, [2]>([-1, 0])];
|
| 25 |
-
tensor<int32, []> gather_0_axis_0 = const()[name = tensor<string, []>("gather_0_axis_0"), val = tensor<int32, []>(-1)];
|
| 26 |
-
tensor<int32, []> gather_0_batch_dims_0 = const()[name = tensor<string, []>("gather_0_batch_dims_0"), val = tensor<int32, []>(0)];
|
| 27 |
-
tensor<bool, []> gather_0_validate_indices_0 = const()[name = tensor<string, []>("gather_0_validate_indices_0"), val = tensor<bool, []>(false)];
|
| 28 |
-
tensor<uint16, [129]> range_1d_2_to_uint16 = const()[name = tensor<string, []>("range_1d_2_to_uint16"), val = tensor<uint16, [129]>([0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128])];
|
| 29 |
-
tensor<fp16, [1, 256]> transpose_2_cast_fp16 = transpose(perm = transpose_2_perm_0, x = matmul_1_cast_fp16)[name = tensor<string, []>("transpose_8")];
|
| 30 |
-
tensor<fp16, [1, 129]> gather_0_cast_fp16_cast_uint16 = gather(axis = gather_0_axis_0, batch_dims = gather_0_batch_dims_0, indices = range_1d_2_to_uint16, validate_indices = gather_0_validate_indices_0, x = transpose_2_cast_fp16)[name = tensor<string, []>("gather_0_cast_fp16_cast_uint16")];
|
| 31 |
-
tensor<int32, []> gather_1_axis_0 = const()[name = tensor<string, []>("gather_1_axis_0"), val = tensor<int32, []>(-1)];
|
| 32 |
-
tensor<int32, []> gather_1_batch_dims_0 = const()[name = tensor<string, []>("gather_1_batch_dims_0"), val = tensor<int32, []>(0)];
|
| 33 |
-
tensor<bool, []> gather_1_validate_indices_0 = const()[name = tensor<string, []>("gather_1_validate_indices_0"), val = tensor<bool, []>(false)];
|
| 34 |
-
tensor<fp16, [1, 256]> transpose_3_cast_fp16 = transpose(perm = transpose_3_perm_0, x = mul_2_cast_fp16)[name = tensor<string, []>("transpose_7")];
|
| 35 |
-
tensor<fp16, [1, 129]> gather_1_cast_fp16_cast_uint16 = gather(axis = gather_1_axis_0, batch_dims = gather_1_batch_dims_0, indices = range_1d_2_to_uint16, validate_indices = gather_1_validate_indices_0, x = transpose_3_cast_fp16)[name = tensor<string, []>("gather_1_cast_fp16_cast_uint16")];
|
| 36 |
-
tensor<fp16, [1, 129]> square_0_cast_fp16 = square(x = gather_0_cast_fp16_cast_uint16)[name = tensor<string, []>("square_0_cast_fp16")];
|
| 37 |
-
tensor<fp16, [1, 129]> square_1_cast_fp16 = square(x = gather_1_cast_fp16_cast_uint16)[name = tensor<string, []>("square_1_cast_fp16")];
|
| 38 |
-
tensor<fp16, [1, 129]> add_2_cast_fp16 = add(x = square_0_cast_fp16, y = square_1_cast_fp16)[name = tensor<string, []>("add_2_cast_fp16")];
|
| 39 |
-
tensor<fp16, [1, 129]> sqrt_0_cast_fp16 = sqrt(x = add_2_cast_fp16)[name = tensor<string, []>("sqrt_0_cast_fp16")];
|
| 40 |
-
tensor<int32, [2]> var_25_begin_0 = const()[name = tensor<string, []>("op_25_begin_0"), val = tensor<int32, [2]>([0, 128])];
|
| 41 |
-
tensor<int32, [2]> var_25_end_0 = const()[name = tensor<string, []>("op_25_end_0"), val = tensor<int32, [2]>([1, 384])];
|
| 42 |
-
tensor<bool, [2]> var_25_end_mask_0 = const()[name = tensor<string, []>("op_25_end_mask_0"), val = tensor<bool, [2]>([true, false])];
|
| 43 |
-
tensor<fp16, [1, 256]> var_25_cast_fp16 = slice_by_index(begin = var_25_begin_0, end = var_25_end_0, end_mask = var_25_end_mask_0, x = audio_chunk_to_fp16)[name = tensor<string, []>("op_25_cast_fp16")];
|
| 44 |
-
tensor<fp16, [1, 256]> frame_3_cast_fp16 = mul(x = var_25_cast_fp16, y = window_to_fp16)[name = tensor<string, []>("frame_3_cast_fp16")];
|
| 45 |
-
tensor<bool, []> matmul_6_transpose_x_1 = const()[name = tensor<string, []>("matmul_6_transpose_x_1"), val = tensor<bool, []>(false)];
|
| 46 |
-
tensor<bool, []> matmul_6_transpose_y_1 = const()[name = tensor<string, []>("matmul_6_transpose_y_1"), val = tensor<bool, []>(true)];
|
| 47 |
-
tensor<fp16, [256, 1]> matmul_6_cast_fp16 = matmul(transpose_x = matmul_6_transpose_x_1, transpose_y = matmul_6_transpose_y_1, x = cos_0_to_fp16_palettized, y = frame_3_cast_fp16)[name = tensor<string, []>("matmul_6_cast_fp16")];
|
| 48 |
-
tensor<bool, []> matmul_8_transpose_x_1 = const()[name = tensor<string, []>("matmul_8_transpose_x_1"), val = tensor<bool, []>(false)];
|
| 49 |
-
tensor<bool, []> matmul_8_transpose_y_1 = const()[name = tensor<string, []>("matmul_8_transpose_y_1"), val = tensor<bool, []>(true)];
|
| 50 |
-
tensor<fp16, [256, 1]> matmul_8_cast_fp16 = matmul(transpose_x = matmul_8_transpose_x_1, transpose_y = matmul_8_transpose_y_1, x = sin_1_to_fp16_palettized, y = frame_3_cast_fp16)[name = tensor<string, []>("matmul_8_cast_fp16")];
|
| 51 |
-
tensor<fp16, []> mul_4_y_0_to_fp16 = const()[name = tensor<string, []>("mul_4_y_0_to_fp16"), val = tensor<fp16, []>(-0x1p+0)];
|
| 52 |
-
tensor<fp16, [256, 1]> mul_4_cast_fp16 = mul(x = matmul_8_cast_fp16, y = mul_4_y_0_to_fp16)[name = tensor<string, []>("mul_4_cast_fp16")];
|
| 53 |
-
tensor<int32, [2]> transpose_6_perm_0 = const()[name = tensor<string, []>("transpose_6_perm_0"), val = tensor<int32, [2]>([-1, 0])];
|
| 54 |
-
tensor<int32, [2]> transpose_7_perm_0 = const()[name = tensor<string, []>("transpose_7_perm_0"), val = tensor<int32, [2]>([-1, 0])];
|
| 55 |
-
tensor<int32, []> gather_2_axis_0 = const()[name = tensor<string, []>("gather_2_axis_0"), val = tensor<int32, []>(-1)];
|
| 56 |
-
tensor<int32, []> gather_2_batch_dims_0 = const()[name = tensor<string, []>("gather_2_batch_dims_0"), val = tensor<int32, []>(0)];
|
| 57 |
-
tensor<bool, []> gather_2_validate_indices_0 = const()[name = tensor<string, []>("gather_2_validate_indices_0"), val = tensor<bool, []>(false)];
|
| 58 |
-
tensor<fp16, [1, 256]> transpose_6_cast_fp16 = transpose(perm = transpose_6_perm_0, x = matmul_6_cast_fp16)[name = tensor<string, []>("transpose_6")];
|
| 59 |
-
tensor<fp16, [1, 129]> gather_2_cast_fp16_cast_uint16 = gather(axis = gather_2_axis_0, batch_dims = gather_2_batch_dims_0, indices = range_1d_2_to_uint16, validate_indices = gather_2_validate_indices_0, x = transpose_6_cast_fp16)[name = tensor<string, []>("gather_2_cast_fp16_cast_uint16")];
|
| 60 |
-
tensor<int32, []> gather_3_axis_0 = const()[name = tensor<string, []>("gather_3_axis_0"), val = tensor<int32, []>(-1)];
|
| 61 |
-
tensor<int32, []> gather_3_batch_dims_0 = const()[name = tensor<string, []>("gather_3_batch_dims_0"), val = tensor<int32, []>(0)];
|
| 62 |
-
tensor<bool, []> gather_3_validate_indices_0 = const()[name = tensor<string, []>("gather_3_validate_indices_0"), val = tensor<bool, []>(false)];
|
| 63 |
-
tensor<fp16, [1, 256]> transpose_7_cast_fp16 = transpose(perm = transpose_7_perm_0, x = mul_4_cast_fp16)[name = tensor<string, []>("transpose_5")];
|
| 64 |
-
tensor<fp16, [1, 129]> gather_3_cast_fp16_cast_uint16 = gather(axis = gather_3_axis_0, batch_dims = gather_3_batch_dims_0, indices = range_1d_2_to_uint16, validate_indices = gather_3_validate_indices_0, x = transpose_7_cast_fp16)[name = tensor<string, []>("gather_3_cast_fp16_cast_uint16")];
|
| 65 |
-
tensor<fp16, [1, 129]> square_2_cast_fp16 = square(x = gather_2_cast_fp16_cast_uint16)[name = tensor<string, []>("square_2_cast_fp16")];
|
| 66 |
-
tensor<fp16, [1, 129]> square_3_cast_fp16 = square(x = gather_3_cast_fp16_cast_uint16)[name = tensor<string, []>("square_3_cast_fp16")];
|
| 67 |
-
tensor<fp16, [1, 129]> add_4_cast_fp16 = add(x = square_2_cast_fp16, y = square_3_cast_fp16)[name = tensor<string, []>("add_4_cast_fp16")];
|
| 68 |
-
tensor<fp16, [1, 129]> sqrt_1_cast_fp16 = sqrt(x = add_4_cast_fp16)[name = tensor<string, []>("sqrt_1_cast_fp16")];
|
| 69 |
-
tensor<int32, [2]> var_30_begin_0 = const()[name = tensor<string, []>("op_30_begin_0"), val = tensor<int32, [2]>([0, 256])];
|
| 70 |
-
tensor<int32, [2]> var_30_end_0 = const()[name = tensor<string, []>("op_30_end_0"), val = tensor<int32, [2]>([1, 1])];
|
| 71 |
-
tensor<bool, [2]> var_30_end_mask_0 = const()[name = tensor<string, []>("op_30_end_mask_0"), val = tensor<bool, [2]>([true, true])];
|
| 72 |
-
tensor<fp16, [1, 256]> var_30_cast_fp16 = slice_by_index(begin = var_30_begin_0, end = var_30_end_0, end_mask = var_30_end_mask_0, x = audio_chunk_to_fp16)[name = tensor<string, []>("op_30_cast_fp16")];
|
| 73 |
-
tensor<fp16, [1, 256]> frame_cast_fp16 = mul(x = var_30_cast_fp16, y = window_to_fp16)[name = tensor<string, []>("frame_cast_fp16")];
|
| 74 |
-
tensor<bool, []> matmul_11_transpose_x_1 = const()[name = tensor<string, []>("matmul_11_transpose_x_1"), val = tensor<bool, []>(false)];
|
| 75 |
-
tensor<bool, []> matmul_11_transpose_y_1 = const()[name = tensor<string, []>("matmul_11_transpose_y_1"), val = tensor<bool, []>(true)];
|
| 76 |
-
tensor<fp16, [256, 1]> matmul_11_cast_fp16 = matmul(transpose_x = matmul_11_transpose_x_1, transpose_y = matmul_11_transpose_y_1, x = cos_0_to_fp16_palettized, y = frame_cast_fp16)[name = tensor<string, []>("matmul_11_cast_fp16")];
|
| 77 |
-
tensor<bool, []> matmul_13_transpose_x_1 = const()[name = tensor<string, []>("matmul_13_transpose_x_1"), val = tensor<bool, []>(false)];
|
| 78 |
-
tensor<bool, []> matmul_13_transpose_y_1 = const()[name = tensor<string, []>("matmul_13_transpose_y_1"), val = tensor<bool, []>(true)];
|
| 79 |
-
tensor<fp16, [256, 1]> matmul_13_cast_fp16 = matmul(transpose_x = matmul_13_transpose_x_1, transpose_y = matmul_13_transpose_y_1, x = sin_1_to_fp16_palettized, y = frame_cast_fp16)[name = tensor<string, []>("matmul_13_cast_fp16")];
|
| 80 |
-
tensor<fp16, []> mul_6_y_0_to_fp16 = const()[name = tensor<string, []>("mul_6_y_0_to_fp16"), val = tensor<fp16, []>(-0x1p+0)];
|
| 81 |
-
tensor<fp16, [256, 1]> mul_6_cast_fp16 = mul(x = matmul_13_cast_fp16, y = mul_6_y_0_to_fp16)[name = tensor<string, []>("mul_6_cast_fp16")];
|
| 82 |
-
tensor<int32, [2]> transpose_10_perm_0 = const()[name = tensor<string, []>("transpose_10_perm_0"), val = tensor<int32, [2]>([-1, 0])];
|
| 83 |
-
tensor<int32, [2]> transpose_11_perm_0 = const()[name = tensor<string, []>("transpose_11_perm_0"), val = tensor<int32, [2]>([-1, 0])];
|
| 84 |
-
tensor<int32, []> gather_4_axis_0 = const()[name = tensor<string, []>("gather_4_axis_0"), val = tensor<int32, []>(-1)];
|
| 85 |
-
tensor<int32, []> gather_4_batch_dims_0 = const()[name = tensor<string, []>("gather_4_batch_dims_0"), val = tensor<int32, []>(0)];
|
| 86 |
-
tensor<bool, []> gather_4_validate_indices_0 = const()[name = tensor<string, []>("gather_4_validate_indices_0"), val = tensor<bool, []>(false)];
|
| 87 |
-
tensor<fp16, [1, 256]> transpose_10_cast_fp16 = transpose(perm = transpose_10_perm_0, x = matmul_11_cast_fp16)[name = tensor<string, []>("transpose_4")];
|
| 88 |
-
tensor<fp16, [1, 129]> gather_4_cast_fp16_cast_uint16 = gather(axis = gather_4_axis_0, batch_dims = gather_4_batch_dims_0, indices = range_1d_2_to_uint16, validate_indices = gather_4_validate_indices_0, x = transpose_10_cast_fp16)[name = tensor<string, []>("gather_4_cast_fp16_cast_uint16")];
|
| 89 |
-
tensor<int32, []> gather_5_axis_0 = const()[name = tensor<string, []>("gather_5_axis_0"), val = tensor<int32, []>(-1)];
|
| 90 |
-
tensor<int32, []> gather_5_batch_dims_0 = const()[name = tensor<string, []>("gather_5_batch_dims_0"), val = tensor<int32, []>(0)];
|
| 91 |
-
tensor<bool, []> gather_5_validate_indices_0 = const()[name = tensor<string, []>("gather_5_validate_indices_0"), val = tensor<bool, []>(false)];
|
| 92 |
-
tensor<fp16, [1, 256]> transpose_11_cast_fp16 = transpose(perm = transpose_11_perm_0, x = mul_6_cast_fp16)[name = tensor<string, []>("transpose_3")];
|
| 93 |
-
tensor<fp16, [1, 129]> gather_5_cast_fp16_cast_uint16 = gather(axis = gather_5_axis_0, batch_dims = gather_5_batch_dims_0, indices = range_1d_2_to_uint16, validate_indices = gather_5_validate_indices_0, x = transpose_11_cast_fp16)[name = tensor<string, []>("gather_5_cast_fp16_cast_uint16")];
|
| 94 |
-
tensor<fp16, [1, 129]> square_4_cast_fp16 = square(x = gather_4_cast_fp16_cast_uint16)[name = tensor<string, []>("square_4_cast_fp16")];
|
| 95 |
-
tensor<fp16, [1, 129]> square_5_cast_fp16 = square(x = gather_5_cast_fp16_cast_uint16)[name = tensor<string, []>("square_5_cast_fp16")];
|
| 96 |
-
tensor<fp16, [1, 129]> add_6_cast_fp16 = add(x = square_4_cast_fp16, y = square_5_cast_fp16)[name = tensor<string, []>("add_6_cast_fp16")];
|
| 97 |
-
tensor<fp16, [1, 129]> sqrt_2_cast_fp16 = sqrt(x = add_6_cast_fp16)[name = tensor<string, []>("sqrt_2_cast_fp16")];
|
| 98 |
-
tensor<int32, []> input_1_axis_0 = const()[name = tensor<string, []>("input_1_axis_0"), val = tensor<int32, []>(2)];
|
| 99 |
-
tensor<fp16, [1, 129, 3]> input_1_cast_fp16 = stack(axis = input_1_axis_0, values = (sqrt_0_cast_fp16, sqrt_1_cast_fp16, sqrt_2_cast_fp16))[name = tensor<string, []>("input_1_cast_fp16")];
|
| 100 |
-
tensor<string, []> x_1_pad_type_0 = const()[name = tensor<string, []>("x_1_pad_type_0"), val = tensor<string, []>("custom")];
|
| 101 |
-
tensor<int32, [2]> x_1_pad_0 = const()[name = tensor<string, []>("x_1_pad_0"), val = tensor<int32, [2]>([1, 1])];
|
| 102 |
-
tensor<int32, [1]> x_1_strides_0 = const()[name = tensor<string, []>("x_1_strides_0"), val = tensor<int32, [1]>([1])];
|
| 103 |
-
tensor<int32, [1]> x_1_dilations_0 = const()[name = tensor<string, []>("x_1_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 104 |
-
tensor<int32, []> x_1_groups_0 = const()[name = tensor<string, []>("x_1_groups_0"), val = tensor<int32, []>(1)];
|
| 105 |
-
tensor<fp16, [128, 129, 3]> encoder_encoder_0_conv_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [24768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(66560))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(91392))), name = tensor<string, []>("encoder_encoder_0_conv_weight_to_fp16_palettized"), shape = tensor<uint32, [3]>([128, 129, 3])];
|
| 106 |
-
tensor<fp16, [128]> encoder_encoder_0_conv_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_0_conv_bias_to_fp16"), val = tensor<fp16, [128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(91520)))];
|
| 107 |
-
tensor<fp16, [1, 128, 3]> x_1_cast_fp16 = conv(bias = encoder_encoder_0_conv_bias_to_fp16, dilations = x_1_dilations_0, groups = x_1_groups_0, pad = x_1_pad_0, pad_type = x_1_pad_type_0, strides = x_1_strides_0, weight = encoder_encoder_0_conv_weight_to_fp16_palettized, x = input_1_cast_fp16)[name = tensor<string, []>("x_1_cast_fp16")];
|
| 108 |
-
tensor<int32, [1]> reduce_mean_0_axes_0 = const()[name = tensor<string, []>("reduce_mean_0_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 109 |
-
tensor<bool, []> reduce_mean_0_keep_dims_0 = const()[name = tensor<string, []>("reduce_mean_0_keep_dims_0"), val = tensor<bool, []>(true)];
|
| 110 |
-
tensor<fp16, [1, 128, 1]> reduce_mean_0_cast_fp16 = reduce_mean(axes = reduce_mean_0_axes_0, keep_dims = reduce_mean_0_keep_dims_0, x = x_1_cast_fp16)[name = tensor<string, []>("reduce_mean_0_cast_fp16")];
|
| 111 |
-
tensor<int32, []> concat_0_axis_0 = const()[name = tensor<string, []>("concat_0_axis_0"), val = tensor<int32, []>(-1)];
|
| 112 |
-
tensor<bool, []> concat_0_interleave_0 = const()[name = tensor<string, []>("concat_0_interleave_0"), val = tensor<bool, []>(false)];
|
| 113 |
-
tensor<fp16, [1, 128, 1]> concat_0_cast_fp16 = concat(axis = concat_0_axis_0, interleave = concat_0_interleave_0, values = reduce_mean_0_cast_fp16)[name = tensor<string, []>("concat_0_cast_fp16")];
|
| 114 |
-
tensor<int32, [2]> var_65 = const()[name = tensor<string, []>("op_65"), val = tensor<int32, [2]>([1, 128])];
|
| 115 |
-
tensor<fp16, [1, 128]> input_3_cast_fp16 = reshape(shape = var_65, x = concat_0_cast_fp16)[name = tensor<string, []>("input_3_cast_fp16")];
|
| 116 |
-
tensor<fp16, [16, 128]> encoder_encoder_0_se_fc1_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_0_se_fc1_weight_to_fp16"), val = tensor<fp16, [16, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(91840)))];
|
| 117 |
-
tensor<fp16, [16]> encoder_encoder_0_se_fc1_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_0_se_fc1_bias_to_fp16"), val = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(96000)))];
|
| 118 |
-
tensor<fp16, [1, 16]> linear_0_cast_fp16 = linear(bias = encoder_encoder_0_se_fc1_bias_to_fp16, weight = encoder_encoder_0_se_fc1_weight_to_fp16, x = input_3_cast_fp16)[name = tensor<string, []>("linear_0_cast_fp16")];
|
| 119 |
-
tensor<fp16, [1, 16]> input_7_cast_fp16 = relu(x = linear_0_cast_fp16)[name = tensor<string, []>("input_7_cast_fp16")];
|
| 120 |
-
tensor<fp16, [128, 16]> encoder_encoder_0_se_fc2_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_0_se_fc2_weight_to_fp16"), val = tensor<fp16, [128, 16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(96128)))];
|
| 121 |
-
tensor<fp16, [128]> encoder_encoder_0_se_fc2_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_0_se_fc2_bias_to_fp16"), val = tensor<fp16, [128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(100288)))];
|
| 122 |
-
tensor<fp16, [1, 128]> linear_1_cast_fp16 = linear(bias = encoder_encoder_0_se_fc2_bias_to_fp16, weight = encoder_encoder_0_se_fc2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor<string, []>("linear_1_cast_fp16")];
|
| 123 |
-
tensor<fp16, [1, 128]> y_1_cast_fp16 = sigmoid(x = linear_1_cast_fp16)[name = tensor<string, []>("y_1_cast_fp16")];
|
| 124 |
-
tensor<int32, [3]> var_75 = const()[name = tensor<string, []>("op_75"), val = tensor<int32, [3]>([1, 128, 1])];
|
| 125 |
-
tensor<fp16, [1, 128, 1]> y_3_cast_fp16 = reshape(shape = var_75, x = y_1_cast_fp16)[name = tensor<string, []>("y_3_cast_fp16")];
|
| 126 |
-
tensor<fp16, [1, 128, 3]> input_11_cast_fp16 = mul(x = x_1_cast_fp16, y = y_3_cast_fp16)[name = tensor<string, []>("input_11_cast_fp16")];
|
| 127 |
-
tensor<fp16, [1, 128, 3]> input_13_cast_fp16 = relu(x = input_11_cast_fp16)[name = tensor<string, []>("input_13_cast_fp16")];
|
| 128 |
-
tensor<string, []> x_3_pad_type_0 = const()[name = tensor<string, []>("x_3_pad_type_0"), val = tensor<string, []>("custom")];
|
| 129 |
-
tensor<int32, [2]> x_3_pad_0 = const()[name = tensor<string, []>("x_3_pad_0"), val = tensor<int32, [2]>([1, 1])];
|
| 130 |
-
tensor<int32, [1]> x_3_strides_0 = const()[name = tensor<string, []>("x_3_strides_0"), val = tensor<int32, [1]>([1])];
|
| 131 |
-
tensor<int32, [1]> x_3_dilations_0 = const()[name = tensor<string, []>("x_3_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 132 |
-
tensor<int32, []> x_3_groups_0 = const()[name = tensor<string, []>("x_3_groups_0"), val = tensor<int32, []>(1)];
|
| 133 |
-
tensor<fp16, [64, 128, 3]> encoder_encoder_1_conv_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [12288]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(100608))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(112960))), name = tensor<string, []>("encoder_encoder_1_conv_weight_to_fp16_palettized"), shape = tensor<uint32, [3]>([64, 128, 3])];
|
| 134 |
-
tensor<fp16, [64]> encoder_encoder_1_conv_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_1_conv_bias_to_fp16"), val = tensor<fp16, [64]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(113088)))];
|
| 135 |
-
tensor<fp16, [1, 64, 3]> x_3_cast_fp16 = conv(bias = encoder_encoder_1_conv_bias_to_fp16, dilations = x_3_dilations_0, groups = x_3_groups_0, pad = x_3_pad_0, pad_type = x_3_pad_type_0, strides = x_3_strides_0, weight = encoder_encoder_1_conv_weight_to_fp16_palettized, x = input_13_cast_fp16)[name = tensor<string, []>("x_3_cast_fp16")];
|
| 136 |
-
tensor<int32, [1]> reduce_mean_1_axes_0 = const()[name = tensor<string, []>("reduce_mean_1_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 137 |
-
tensor<bool, []> reduce_mean_1_keep_dims_0 = const()[name = tensor<string, []>("reduce_mean_1_keep_dims_0"), val = tensor<bool, []>(true)];
|
| 138 |
-
tensor<fp16, [1, 64, 1]> reduce_mean_1_cast_fp16 = reduce_mean(axes = reduce_mean_1_axes_0, keep_dims = reduce_mean_1_keep_dims_0, x = x_3_cast_fp16)[name = tensor<string, []>("reduce_mean_1_cast_fp16")];
|
| 139 |
-
tensor<int32, []> concat_1_axis_0 = const()[name = tensor<string, []>("concat_1_axis_0"), val = tensor<int32, []>(-1)];
|
| 140 |
-
tensor<bool, []> concat_1_interleave_0 = const()[name = tensor<string, []>("concat_1_interleave_0"), val = tensor<bool, []>(false)];
|
| 141 |
-
tensor<fp16, [1, 64, 1]> concat_1_cast_fp16 = concat(axis = concat_1_axis_0, interleave = concat_1_interleave_0, values = reduce_mean_1_cast_fp16)[name = tensor<string, []>("concat_1_cast_fp16")];
|
| 142 |
-
tensor<int32, [2]> var_94 = const()[name = tensor<string, []>("op_94"), val = tensor<int32, [2]>([1, 64])];
|
| 143 |
-
tensor<fp16, [1, 64]> input_15_cast_fp16 = reshape(shape = var_94, x = concat_1_cast_fp16)[name = tensor<string, []>("input_15_cast_fp16")];
|
| 144 |
-
tensor<fp16, [8, 64]> encoder_encoder_1_se_fc1_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_1_se_fc1_weight_to_fp16"), val = tensor<fp16, [8, 64]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(113280)))];
|
| 145 |
-
tensor<fp16, [8]> encoder_encoder_1_se_fc1_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_1_se_fc1_bias_to_fp16"), val = tensor<fp16, [8]>([0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0])];
|
| 146 |
-
tensor<fp16, [1, 8]> linear_2_cast_fp16 = linear(bias = encoder_encoder_1_se_fc1_bias_to_fp16, weight = encoder_encoder_1_se_fc1_weight_to_fp16, x = input_15_cast_fp16)[name = tensor<string, []>("linear_2_cast_fp16")];
|
| 147 |
-
tensor<fp16, [1, 8]> input_19_cast_fp16 = relu(x = linear_2_cast_fp16)[name = tensor<string, []>("input_19_cast_fp16")];
|
| 148 |
-
tensor<fp16, [64, 8]> encoder_encoder_1_se_fc2_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_1_se_fc2_weight_to_fp16"), val = tensor<fp16, [64, 8]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(114368)))];
|
| 149 |
-
tensor<fp16, [64]> encoder_encoder_1_se_fc2_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_1_se_fc2_bias_to_fp16"), val = tensor<fp16, [64]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(115456)))];
|
| 150 |
-
tensor<fp16, [1, 64]> linear_3_cast_fp16 = linear(bias = encoder_encoder_1_se_fc2_bias_to_fp16, weight = encoder_encoder_1_se_fc2_weight_to_fp16, x = input_19_cast_fp16)[name = tensor<string, []>("linear_3_cast_fp16")];
|
| 151 |
-
tensor<fp16, [1, 64]> y_5_cast_fp16 = sigmoid(x = linear_3_cast_fp16)[name = tensor<string, []>("y_5_cast_fp16")];
|
| 152 |
-
tensor<int32, [3]> var_104 = const()[name = tensor<string, []>("op_104"), val = tensor<int32, [3]>([1, 64, 1])];
|
| 153 |
-
tensor<fp16, [1, 64, 1]> y_7_cast_fp16 = reshape(shape = var_104, x = y_5_cast_fp16)[name = tensor<string, []>("y_7_cast_fp16")];
|
| 154 |
-
tensor<fp16, [1, 64, 3]> input_23_cast_fp16 = mul(x = x_3_cast_fp16, y = y_7_cast_fp16)[name = tensor<string, []>("input_23_cast_fp16")];
|
| 155 |
-
tensor<fp16, [1, 64, 3]> input_25_cast_fp16 = relu(x = input_23_cast_fp16)[name = tensor<string, []>("input_25_cast_fp16")];
|
| 156 |
-
tensor<string, []> x_5_pad_type_0 = const()[name = tensor<string, []>("x_5_pad_type_0"), val = tensor<string, []>("custom")];
|
| 157 |
-
tensor<int32, [2]> x_5_pad_0 = const()[name = tensor<string, []>("x_5_pad_0"), val = tensor<int32, [2]>([1, 1])];
|
| 158 |
-
tensor<int32, [1]> x_5_strides_0 = const()[name = tensor<string, []>("x_5_strides_0"), val = tensor<int32, [1]>([1])];
|
| 159 |
-
tensor<int32, [1]> x_5_dilations_0 = const()[name = tensor<string, []>("x_5_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 160 |
-
tensor<int32, []> x_5_groups_0 = const()[name = tensor<string, []>("x_5_groups_0"), val = tensor<int32, []>(1)];
|
| 161 |
-
tensor<fp16, [64, 64, 3]> encoder_encoder_2_conv_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [6144]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(115648))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(121856))), name = tensor<string, []>("encoder_encoder_2_conv_weight_to_fp16_palettized"), shape = tensor<uint32, [3]>([64, 64, 3])];
|
| 162 |
-
tensor<fp16, [64]> encoder_encoder_2_conv_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_2_conv_bias_to_fp16"), val = tensor<fp16, [64]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(121984)))];
|
| 163 |
-
tensor<fp16, [1, 64, 3]> x_5_cast_fp16 = conv(bias = encoder_encoder_2_conv_bias_to_fp16, dilations = x_5_dilations_0, groups = x_5_groups_0, pad = x_5_pad_0, pad_type = x_5_pad_type_0, strides = x_5_strides_0, weight = encoder_encoder_2_conv_weight_to_fp16_palettized, x = input_25_cast_fp16)[name = tensor<string, []>("x_5_cast_fp16")];
|
| 164 |
-
tensor<int32, [1]> reduce_mean_2_axes_0 = const()[name = tensor<string, []>("reduce_mean_2_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 165 |
-
tensor<bool, []> reduce_mean_2_keep_dims_0 = const()[name = tensor<string, []>("reduce_mean_2_keep_dims_0"), val = tensor<bool, []>(true)];
|
| 166 |
-
tensor<fp16, [1, 64, 1]> reduce_mean_2_cast_fp16 = reduce_mean(axes = reduce_mean_2_axes_0, keep_dims = reduce_mean_2_keep_dims_0, x = x_5_cast_fp16)[name = tensor<string, []>("reduce_mean_2_cast_fp16")];
|
| 167 |
-
tensor<int32, []> concat_2_axis_0 = const()[name = tensor<string, []>("concat_2_axis_0"), val = tensor<int32, []>(-1)];
|
| 168 |
-
tensor<bool, []> concat_2_interleave_0 = const()[name = tensor<string, []>("concat_2_interleave_0"), val = tensor<bool, []>(false)];
|
| 169 |
-
tensor<fp16, [1, 64, 1]> concat_2_cast_fp16 = concat(axis = concat_2_axis_0, interleave = concat_2_interleave_0, values = reduce_mean_2_cast_fp16)[name = tensor<string, []>("concat_2_cast_fp16")];
|
| 170 |
-
tensor<int32, [2]> var_123 = const()[name = tensor<string, []>("op_123"), val = tensor<int32, [2]>([1, 64])];
|
| 171 |
-
tensor<fp16, [1, 64]> input_27_cast_fp16 = reshape(shape = var_123, x = concat_2_cast_fp16)[name = tensor<string, []>("input_27_cast_fp16")];
|
| 172 |
-
tensor<fp16, [8, 64]> encoder_encoder_2_se_fc1_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_2_se_fc1_weight_to_fp16"), val = tensor<fp16, [8, 64]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(122176)))];
|
| 173 |
-
tensor<fp16, [8]> encoder_encoder_2_se_fc1_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_2_se_fc1_bias_to_fp16"), val = tensor<fp16, [8]>([0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0, 0x0p+0])];
|
| 174 |
-
tensor<fp16, [1, 8]> linear_4_cast_fp16 = linear(bias = encoder_encoder_2_se_fc1_bias_to_fp16, weight = encoder_encoder_2_se_fc1_weight_to_fp16, x = input_27_cast_fp16)[name = tensor<string, []>("linear_4_cast_fp16")];
|
| 175 |
-
tensor<fp16, [1, 8]> input_31_cast_fp16 = relu(x = linear_4_cast_fp16)[name = tensor<string, []>("input_31_cast_fp16")];
|
| 176 |
-
tensor<fp16, [64, 8]> encoder_encoder_2_se_fc2_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_2_se_fc2_weight_to_fp16"), val = tensor<fp16, [64, 8]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(123264)))];
|
| 177 |
-
tensor<fp16, [64]> encoder_encoder_2_se_fc2_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_2_se_fc2_bias_to_fp16"), val = tensor<fp16, [64]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(124352)))];
|
| 178 |
-
tensor<fp16, [1, 64]> linear_5_cast_fp16 = linear(bias = encoder_encoder_2_se_fc2_bias_to_fp16, weight = encoder_encoder_2_se_fc2_weight_to_fp16, x = input_31_cast_fp16)[name = tensor<string, []>("linear_5_cast_fp16")];
|
| 179 |
-
tensor<fp16, [1, 64]> y_9_cast_fp16 = sigmoid(x = linear_5_cast_fp16)[name = tensor<string, []>("y_9_cast_fp16")];
|
| 180 |
-
tensor<int32, [3]> var_133 = const()[name = tensor<string, []>("op_133"), val = tensor<int32, [3]>([1, 64, 1])];
|
| 181 |
-
tensor<fp16, [1, 64, 1]> y_11_cast_fp16 = reshape(shape = var_133, x = y_9_cast_fp16)[name = tensor<string, []>("y_11_cast_fp16")];
|
| 182 |
-
tensor<fp16, [1, 64, 3]> input_35_cast_fp16 = mul(x = x_5_cast_fp16, y = y_11_cast_fp16)[name = tensor<string, []>("input_35_cast_fp16")];
|
| 183 |
-
tensor<fp16, [1, 64, 3]> input_37_cast_fp16 = relu(x = input_35_cast_fp16)[name = tensor<string, []>("input_37_cast_fp16")];
|
| 184 |
-
tensor<string, []> x_7_pad_type_0 = const()[name = tensor<string, []>("x_7_pad_type_0"), val = tensor<string, []>("custom")];
|
| 185 |
-
tensor<int32, [2]> x_7_pad_0 = const()[name = tensor<string, []>("x_7_pad_0"), val = tensor<int32, [2]>([1, 1])];
|
| 186 |
-
tensor<int32, [1]> x_7_strides_0 = const()[name = tensor<string, []>("x_7_strides_0"), val = tensor<int32, [1]>([1])];
|
| 187 |
-
tensor<int32, [1]> x_7_dilations_0 = const()[name = tensor<string, []>("x_7_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 188 |
-
tensor<int32, []> x_7_groups_0 = const()[name = tensor<string, []>("x_7_groups_0"), val = tensor<int32, []>(1)];
|
| 189 |
-
tensor<fp16, [128, 64, 3]> encoder_encoder_3_conv_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [12288]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(124544))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(136896))), name = tensor<string, []>("encoder_encoder_3_conv_weight_to_fp16_palettized"), shape = tensor<uint32, [3]>([128, 64, 3])];
|
| 190 |
-
tensor<fp16, [128]> encoder_encoder_3_conv_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_3_conv_bias_to_fp16"), val = tensor<fp16, [128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(137024)))];
|
| 191 |
-
tensor<fp16, [1, 128, 3]> x_7_cast_fp16 = conv(bias = encoder_encoder_3_conv_bias_to_fp16, dilations = x_7_dilations_0, groups = x_7_groups_0, pad = x_7_pad_0, pad_type = x_7_pad_type_0, strides = x_7_strides_0, weight = encoder_encoder_3_conv_weight_to_fp16_palettized, x = input_37_cast_fp16)[name = tensor<string, []>("x_7_cast_fp16")];
|
| 192 |
-
tensor<int32, [1]> reduce_mean_3_axes_0 = const()[name = tensor<string, []>("reduce_mean_3_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 193 |
-
tensor<bool, []> reduce_mean_3_keep_dims_0 = const()[name = tensor<string, []>("reduce_mean_3_keep_dims_0"), val = tensor<bool, []>(true)];
|
| 194 |
-
tensor<fp16, [1, 128, 1]> reduce_mean_3_cast_fp16 = reduce_mean(axes = reduce_mean_3_axes_0, keep_dims = reduce_mean_3_keep_dims_0, x = x_7_cast_fp16)[name = tensor<string, []>("reduce_mean_3_cast_fp16")];
|
| 195 |
-
tensor<int32, []> concat_3_axis_0 = const()[name = tensor<string, []>("concat_3_axis_0"), val = tensor<int32, []>(-1)];
|
| 196 |
-
tensor<bool, []> concat_3_interleave_0 = const()[name = tensor<string, []>("concat_3_interleave_0"), val = tensor<bool, []>(false)];
|
| 197 |
-
tensor<fp16, [1, 128, 1]> concat_3_cast_fp16 = concat(axis = concat_3_axis_0, interleave = concat_3_interleave_0, values = reduce_mean_3_cast_fp16)[name = tensor<string, []>("concat_3_cast_fp16")];
|
| 198 |
-
tensor<int32, [2]> var_152 = const()[name = tensor<string, []>("op_152"), val = tensor<int32, [2]>([1, 128])];
|
| 199 |
-
tensor<fp16, [1, 128]> input_39_cast_fp16 = reshape(shape = var_152, x = concat_3_cast_fp16)[name = tensor<string, []>("input_39_cast_fp16")];
|
| 200 |
-
tensor<fp16, [16, 128]> encoder_encoder_3_se_fc1_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_3_se_fc1_weight_to_fp16"), val = tensor<fp16, [16, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(137344)))];
|
| 201 |
-
tensor<fp16, [16]> encoder_encoder_3_se_fc1_bias_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_3_se_fc1_bias_to_fp16"), val = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(141504)))];
|
| 202 |
-
tensor<fp16, [1, 16]> linear_6_cast_fp16 = linear(bias = encoder_encoder_3_se_fc1_bias_to_fp16, weight = encoder_encoder_3_se_fc1_weight_to_fp16, x = input_39_cast_fp16)[name = tensor<string, []>("linear_6_cast_fp16")];
|
| 203 |
-
tensor<fp16, [1, 16]> input_43_cast_fp16 = relu(x = linear_6_cast_fp16)[name = tensor<string, []>("input_43_cast_fp16")];
|
| 204 |
-
tensor<fp16, [128, 16]> encoder_encoder_3_se_fc2_weight_to_fp16 = const()[name = tensor<string, []>("encoder_encoder_3_se_fc2_weight_to_fp16"), val = tensor<fp16, [128, 16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(141632)))];
|
| 205 |
-
tensor<fp16, [1, 128]> linear_7_cast_fp16 = linear(bias = encoder_encoder_0_se_fc2_bias_to_fp16, weight = encoder_encoder_3_se_fc2_weight_to_fp16, x = input_43_cast_fp16)[name = tensor<string, []>("linear_7_cast_fp16")];
|
| 206 |
-
tensor<fp16, [1, 128]> y_13_cast_fp16 = sigmoid(x = linear_7_cast_fp16)[name = tensor<string, []>("y_13_cast_fp16")];
|
| 207 |
-
tensor<int32, [3]> var_162 = const()[name = tensor<string, []>("op_162"), val = tensor<int32, [3]>([1, 128, 1])];
|
| 208 |
-
tensor<fp16, [1, 128, 1]> y_cast_fp16 = reshape(shape = var_162, x = y_13_cast_fp16)[name = tensor<string, []>("y_cast_fp16")];
|
| 209 |
-
tensor<fp16, [1, 128, 3]> input_47_cast_fp16 = mul(x = x_7_cast_fp16, y = y_cast_fp16)[name = tensor<string, []>("input_47_cast_fp16")];
|
| 210 |
-
tensor<fp16, [1, 128, 3]> x_9_cast_fp16 = relu(x = input_47_cast_fp16)[name = tensor<string, []>("x_9_cast_fp16")];
|
| 211 |
-
tensor<int32, [1]> x_11_axes_0 = const()[name = tensor<string, []>("x_11_axes_0"), val = tensor<int32, [1]>([2])];
|
| 212 |
-
tensor<bool, []> x_11_keep_dims_0 = const()[name = tensor<string, []>("x_11_keep_dims_0"), val = tensor<bool, []>(true)];
|
| 213 |
-
tensor<fp16, [1, 128, 1]> x_11_cast_fp16 = reduce_mean(axes = x_11_axes_0, keep_dims = x_11_keep_dims_0, x = x_9_cast_fp16)[name = tensor<string, []>("x_11_cast_fp16")];
|
| 214 |
-
tensor<int32, [3]> transpose_12_perm_0 = const()[name = tensor<string, []>("transpose_12_perm_0"), val = tensor<int32, [3]>([2, 0, 1])];
|
| 215 |
-
tensor<int32, [1]> input_49_batch_first_lstm_h0_squeeze_axes_0 = const()[name = tensor<string, []>("input_49_batch_first_lstm_h0_squeeze_axes_0"), val = tensor<int32, [1]>([0])];
|
| 216 |
-
tensor<string, []> h_state_in_to_fp16_dtype_0 = const()[name = tensor<string, []>("h_state_in_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
|
| 217 |
-
tensor<fp16, [1, 1, 128]> h_state_in_to_fp16 = cast(dtype = h_state_in_to_fp16_dtype_0, x = h_state_in)[name = tensor<string, []>("cast_4")];
|
| 218 |
-
tensor<fp16, [1, 128]> input_49_batch_first_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_49_batch_first_lstm_h0_squeeze_axes_0, x = h_state_in_to_fp16)[name = tensor<string, []>("input_49_batch_first_lstm_h0_squeeze_cast_fp16")];
|
| 219 |
-
tensor<int32, [1]> input_49_batch_first_lstm_c0_squeeze_axes_0 = const()[name = tensor<string, []>("input_49_batch_first_lstm_c0_squeeze_axes_0"), val = tensor<int32, [1]>([0])];
|
| 220 |
-
tensor<string, []> c_state_in_to_fp16_dtype_0 = const()[name = tensor<string, []>("c_state_in_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
|
| 221 |
-
tensor<fp16, [1, 1, 128]> c_state_in_to_fp16 = cast(dtype = c_state_in_to_fp16_dtype_0, x = c_state_in)[name = tensor<string, []>("cast_3")];
|
| 222 |
-
tensor<fp16, [1, 128]> input_49_batch_first_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_49_batch_first_lstm_c0_squeeze_axes_0, x = c_state_in_to_fp16)[name = tensor<string, []>("input_49_batch_first_lstm_c0_squeeze_cast_fp16")];
|
| 223 |
-
tensor<string, []> input_49_batch_first_direction_0 = const()[name = tensor<string, []>("input_49_batch_first_direction_0"), val = tensor<string, []>("forward")];
|
| 224 |
-
tensor<bool, []> input_49_batch_first_output_sequence_0 = const()[name = tensor<string, []>("input_49_batch_first_output_sequence_0"), val = tensor<bool, []>(true)];
|
| 225 |
-
tensor<string, []> input_49_batch_first_recurrent_activation_0 = const()[name = tensor<string, []>("input_49_batch_first_recurrent_activation_0"), val = tensor<string, []>("sigmoid")];
|
| 226 |
-
tensor<string, []> input_49_batch_first_cell_activation_0 = const()[name = tensor<string, []>("input_49_batch_first_cell_activation_0"), val = tensor<string, []>("tanh")];
|
| 227 |
-
tensor<string, []> input_49_batch_first_activation_0 = const()[name = tensor<string, []>("input_49_batch_first_activation_0"), val = tensor<string, []>("tanh")];
|
| 228 |
-
tensor<fp16, [512, 128]> concat_5_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [32768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(145792))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(178624))), name = tensor<string, []>("concat_5_to_fp16_palettized"), shape = tensor<uint32, [2]>([512, 128])];
|
| 229 |
-
tensor<fp16, [512, 128]> concat_6_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor<uint8, [32768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(178752))), lut = tensor<fp16, [16]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(211584))), name = tensor<string, []>("concat_6_to_fp16_palettized"), shape = tensor<uint32, [2]>([512, 128])];
|
| 230 |
-
tensor<fp16, [512]> concat_4_to_fp16 = const()[name = tensor<string, []>("concat_4_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(211712)))];
|
| 231 |
-
tensor<fp16, [1, 1, 128]> transpose_12_cast_fp16 = transpose(perm = transpose_12_perm_0, x = x_11_cast_fp16)[name = tensor<string, []>("transpose_2")];
|
| 232 |
-
tensor<fp16, [1, 1, 128]> input_49_batch_first_cast_fp16_0, tensor<fp16, [1, 128]> input_49_batch_first_cast_fp16_1, tensor<fp16, [1, 128]> input_49_batch_first_cast_fp16_2 = lstm(activation = input_49_batch_first_activation_0, bias = concat_4_to_fp16, cell_activation = input_49_batch_first_cell_activation_0, direction = input_49_batch_first_direction_0, initial_c = input_49_batch_first_lstm_c0_squeeze_cast_fp16, initial_h = input_49_batch_first_lstm_h0_squeeze_cast_fp16, output_sequence = input_49_batch_first_output_sequence_0, recurrent_activation = input_49_batch_first_recurrent_activation_0, weight_hh = concat_6_to_fp16_palettized, weight_ih = concat_5_to_fp16_palettized, x = transpose_12_cast_fp16)[name = tensor<string, []>("input_49_batch_first_cast_fp16")];
|
| 233 |
-
tensor<int32, [3]> input_49_perm_0 = const()[name = tensor<string, []>("input_49_perm_0"), val = tensor<int32, [3]>([1, 0, 2])];
|
| 234 |
-
tensor<int32, [1]> var_189_axes_0 = const()[name = tensor<string, []>("op_189_axes_0"), val = tensor<int32, [1]>([0])];
|
| 235 |
-
tensor<fp16, [1, 1, 128]> var_189_cast_fp16 = expand_dims(axes = var_189_axes_0, x = input_49_batch_first_cast_fp16_1)[name = tensor<string, []>("op_189_cast_fp16")];
|
| 236 |
-
tensor<string, []> var_189_cast_fp16_to_fp32_dtype_0 = const()[name = tensor<string, []>("op_189_cast_fp16_to_fp32_dtype_0"), val = tensor<string, []>("fp32")];
|
| 237 |
-
tensor<int32, [1]> var_190_axes_0 = const()[name = tensor<string, []>("op_190_axes_0"), val = tensor<int32, [1]>([0])];
|
| 238 |
-
tensor<fp16, [1, 1, 128]> var_190_cast_fp16 = expand_dims(axes = var_190_axes_0, x = input_49_batch_first_cast_fp16_2)[name = tensor<string, []>("op_190_cast_fp16")];
|
| 239 |
-
tensor<string, []> var_190_cast_fp16_to_fp32_dtype_0 = const()[name = tensor<string, []>("op_190_cast_fp16_to_fp32_dtype_0"), val = tensor<string, []>("fp32")];
|
| 240 |
-
tensor<int32, [1]> var_198_axes_0 = const()[name = tensor<string, []>("op_198_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 241 |
-
tensor<fp16, [128]> decoder_layer_norm_weight_to_fp16 = const()[name = tensor<string, []>("decoder_layer_norm_weight_to_fp16"), val = tensor<fp16, [128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(212800)))];
|
| 242 |
-
tensor<fp16, []> var_172_to_fp16 = const()[name = tensor<string, []>("op_172_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
|
| 243 |
-
tensor<fp16, [1, 1, 128]> input_49_cast_fp16 = transpose(perm = input_49_perm_0, x = input_49_batch_first_cast_fp16_0)[name = tensor<string, []>("transpose_1")];
|
| 244 |
-
tensor<fp16, [1, 1, 128]> var_198_cast_fp16 = layer_norm(axes = var_198_axes_0, beta = encoder_encoder_0_se_fc2_bias_to_fp16, epsilon = var_172_to_fp16, gamma = decoder_layer_norm_weight_to_fp16, x = input_49_cast_fp16)[name = tensor<string, []>("op_198_cast_fp16")];
|
| 245 |
-
tensor<fp16, []> var_199_to_fp16 = const()[name = tensor<string, []>("op_199_to_fp16"), val = tensor<fp16, []>(0x1.334p-3)];
|
| 246 |
-
tensor<fp16, [1, 1, 128]> x_cast_fp16 = mul(x = var_198_cast_fp16, y = var_199_to_fp16)[name = tensor<string, []>("x_cast_fp16")];
|
| 247 |
-
tensor<int32, [3]> input_51_perm_0 = const()[name = tensor<string, []>("input_51_perm_0"), val = tensor<int32, [3]>([0, 2, 1])];
|
| 248 |
-
tensor<fp16, [1, 128, 1]> input_51_cast_fp16 = transpose(perm = input_51_perm_0, x = x_cast_fp16)[name = tensor<string, []>("transpose_0")];
|
| 249 |
-
tensor<fp16, [1, 128, 1]> input_55_cast_fp16 = relu(x = input_51_cast_fp16)[name = tensor<string, []>("input_55_cast_fp16")];
|
| 250 |
-
tensor<string, []> input_pad_type_0 = const()[name = tensor<string, []>("input_pad_type_0"), val = tensor<string, []>("valid")];
|
| 251 |
-
tensor<int32, [1]> input_strides_0 = const()[name = tensor<string, []>("input_strides_0"), val = tensor<int32, [1]>([1])];
|
| 252 |
-
tensor<int32, [2]> input_pad_0 = const()[name = tensor<string, []>("input_pad_0"), val = tensor<int32, [2]>([0, 0])];
|
| 253 |
-
tensor<int32, [1]> input_dilations_0 = const()[name = tensor<string, []>("input_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 254 |
-
tensor<int32, []> input_groups_0 = const()[name = tensor<string, []>("input_groups_0"), val = tensor<int32, []>(1)];
|
| 255 |
-
tensor<fp16, [1, 128, 1]> decoder_conv_weight_to_fp16 = const()[name = tensor<string, []>("decoder_conv_weight_to_fp16"), val = tensor<fp16, [1, 128, 1]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(213120)))];
|
| 256 |
-
tensor<fp16, [1]> decoder_conv_bias_to_fp16 = const()[name = tensor<string, []>("decoder_conv_bias_to_fp16"), val = tensor<fp16, [1]>([-0x1.808p-6])];
|
| 257 |
-
tensor<fp16, [1, 1, 1]> input_cast_fp16 = conv(bias = decoder_conv_bias_to_fp16, dilations = input_dilations_0, groups = input_groups_0, pad = input_pad_0, pad_type = input_pad_type_0, strides = input_strides_0, weight = decoder_conv_weight_to_fp16, x = input_55_cast_fp16)[name = tensor<string, []>("input_cast_fp16")];
|
| 258 |
-
tensor<fp16, [1, 1, 1]> var_211_cast_fp16 = sigmoid(x = input_cast_fp16)[name = tensor<string, []>("op_211_cast_fp16")];
|
| 259 |
-
tensor<int32, [1]> var_212_axes_0 = const()[name = tensor<string, []>("op_212_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 260 |
-
tensor<fp16, [1, 1]> var_212_cast_fp16 = squeeze(axes = var_212_axes_0, x = var_211_cast_fp16)[name = tensor<string, []>("op_212_cast_fp16")];
|
| 261 |
-
tensor<string, []> var_212_cast_fp16_to_fp32_dtype_0 = const()[name = tensor<string, []>("op_212_cast_fp16_to_fp32_dtype_0"), val = tensor<string, []>("fp32")];
|
| 262 |
-
tensor<fp32, [1, 1, 128]> h_state_out = cast(dtype = var_189_cast_fp16_to_fp32_dtype_0, x = var_189_cast_fp16)[name = tensor<string, []>("cast_0")];
|
| 263 |
-
tensor<fp32, [1, 1, 128]> c_state_out = cast(dtype = var_190_cast_fp16_to_fp32_dtype_0, x = var_190_cast_fp16)[name = tensor<string, []>("cast_1")];
|
| 264 |
-
tensor<fp32, [1, 1]> vad_probability = cast(dtype = var_212_cast_fp16_to_fp32_dtype_0, x = var_212_cast_fp16)[name = tensor<string, []>("cast_2")];
|
| 265 |
-
} -> (vad_probability, h_state_out, c_state_out);
|
| 266 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
silero_vad_stateful_4bit.mlmodelc/weights/weight.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:2b948126693c78abcaf0a1dbaff2e7970b2481f5887c4c63cab1f73ded6a0537
|
| 3 |
-
size 213440
|
|
|
|
|
|
|
|
|
|
|
|