diff --git "a/conformer_streaming.mlmodelc/model.mil" "b/conformer_streaming.mlmodelc/model.mil" new file mode 100644--- /dev/null +++ "b/conformer_streaming.mlmodelc/model.mil" @@ -0,0 +1,3015 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor cache_last_channel, tensor cache_last_channel_len, tensor cache_last_time, tensor pre_encoded, tensor pre_encoded_length) { + tensor var_30 = const()[name = tensor("op_30"), val = tensor(-1)]; + tensor var_44 = const()[name = tensor("op_44"), val = tensor(1)]; + tensor var_83_begin_0 = const()[name = tensor("op_83_begin_0"), val = tensor([0, 2, 0])]; + tensor var_83_end_0 = const()[name = tensor("op_83_end_0"), val = tensor([1, 18, 512])]; + tensor var_83_end_mask_0 = const()[name = tensor("op_83_end_mask_0"), val = tensor([true, true, true])]; + tensor pre_encoded_to_fp16_dtype_0 = const()[name = tensor("pre_encoded_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor pre_encoded_to_fp16 = cast(dtype = pre_encoded_to_fp16_dtype_0, x = pre_encoded)[name = tensor("cast_178")]; + tensor var_83_cast_fp16 = slice_by_index(begin = var_83_begin_0, end = var_83_end_0, end_mask = var_83_end_mask_0, x = pre_encoded_to_fp16)[name = tensor("op_83_cast_fp16")]; + tensor var_85 = const()[name = tensor("op_85"), val = tensor(2)]; + tensor var_86 = sub(x = pre_encoded_length, y = var_85)[name = tensor("op_86")]; + tensor var_86_promoted_to_fp16_dtype_0 = const()[name = tensor("op_86_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_39_promoted_to_fp16 = const()[name = tensor("op_39_promoted_to_fp16"), val = tensor(0x0p+0)]; + tensor const_0_to_fp16 = const()[name = tensor("const_0_to_fp16"), val = tensor(inf)]; + tensor var_86_to_fp16 = cast(dtype = var_86_promoted_to_fp16_dtype_0, x = var_86)[name = tensor("cast_177")]; + tensor clip_0_cast_fp16 = clip(alpha = var_39_promoted_to_fp16, beta = const_0_to_fp16, x = var_86_to_fp16)[name = tensor("clip_0_cast_fp16")]; + tensor cache_keep_size = const()[name = tensor("cache_keep_size"), val = tensor([14])]; + tensor var_102_promoted_to_fp16 = const()[name = tensor("op_102_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor padding_length_cast_fp16 = add(x = clip_0_cast_fp16, y = var_102_promoted_to_fp16)[name = tensor("padding_length_cast_fp16")]; + tensor const_2 = const()[name = tensor("const_2"), val = tensor(-1)]; + tensor var_104 = mul(x = cache_last_channel_len, y = const_2)[name = tensor("op_104")]; + tensor var_105 = const()[name = tensor("op_105"), val = tensor(70)]; + tensor offset = add(x = var_104, y = var_105)[name = tensor("offset")]; + tensor var_145_axes_0 = const()[name = tensor("op_145_axes_0"), val = tensor([-1])]; + tensor var_145_cast_fp16 = expand_dims(axes = var_145_axes_0, x = padding_length_cast_fp16)[name = tensor("op_145_cast_fp16")]; + tensor var_144_promoted_to_fp16 = const()[name = tensor("op_144_promoted_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor pad_mask_1_cast_fp16 = less(x = var_144_promoted_to_fp16, y = var_145_cast_fp16)[name = tensor("pad_mask_1_cast_fp16")]; + tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85]])]; + tensor var_151_axes_0 = const()[name = tensor("op_151_axes_0"), val = tensor([-1])]; + tensor var_151 = expand_dims(axes = var_151_axes_0, x = offset)[name = tensor("op_151")]; + tensor pad_mask_off = greater_equal(x = expand_dims_1, y = var_151)[name = tensor("pad_mask_off")]; + tensor pad_mask_3 = logical_and(x = pad_mask_off, y = pad_mask_1_cast_fp16)[name = tensor("pad_mask_3")]; + tensor var_154_axes_0 = const()[name = tensor("op_154_axes_0"), val = tensor([1])]; + tensor var_154 = expand_dims(axes = var_154_axes_0, x = pad_mask_3)[name = tensor("op_154")]; + tensor var_155 = const()[name = tensor("op_155"), val = tensor([1, 86, 1])]; + tensor pad_mask_for_att_mask_1 = tile(reps = var_155, x = var_154)[name = tensor("pad_mask_for_att_mask_1")]; + tensor var_157_perm_0 = const()[name = tensor("op_157_perm_0"), val = tensor([0, 2, 1])]; + tensor var_157 = transpose(perm = var_157_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_239")]; + tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_157)[name = tensor("pad_mask_for_att_mask")]; + tensor const_10 = const()[name = tensor("const_10"), val = tensor([[[true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false], [false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false], [false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false], [false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false], [false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false], [false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; + tensor att_mask_9 = logical_and(x = pad_mask_for_att_mask, y = const_10)[name = tensor("att_mask_9")]; + tensor att_mask = logical_not(x = att_mask_9)[name = tensor("att_mask")]; + tensor pad_mask_5 = logical_not(x = pad_mask_3)[name = tensor("pad_mask_5")]; + tensor pad_mask_begin_0 = const()[name = tensor("pad_mask_begin_0"), val = tensor([0, 70])]; + tensor pad_mask_end_0 = const()[name = tensor("pad_mask_end_0"), val = tensor([1, 86])]; + tensor pad_mask_end_mask_0 = const()[name = tensor("pad_mask_end_mask_0"), val = tensor([true, true])]; + tensor pad_mask = slice_by_index(begin = pad_mask_begin_0, end = pad_mask_end_0, end_mask = pad_mask_end_mask_0, x = pad_mask_5)[name = tensor("pad_mask")]; + tensor mask_1_begin_0 = const()[name = tensor("mask_1_begin_0"), val = tensor([0, 70, 0])]; + tensor mask_1_end_0 = const()[name = tensor("mask_1_end_0"), val = tensor([1, 86, 86])]; + tensor mask_1_end_mask_0 = const()[name = tensor("mask_1_end_mask_0"), val = tensor([true, true, true])]; + tensor mask_1 = slice_by_index(begin = mask_1_begin_0, end = mask_1_end_0, end_mask = mask_1_end_mask_0, x = att_mask)[name = tensor("mask_1")]; + tensor cache_1_begin_0 = const()[name = tensor("cache_1_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_1_end_0 = const()[name = tensor("cache_1_end_0"), val = tensor([1, 1, 70, 512])]; + tensor cache_1_end_mask_0 = const()[name = tensor("cache_1_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_1_squeeze_mask_0 = const()[name = tensor("cache_1_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_channel_to_fp16_dtype_0 = const()[name = tensor("cache_last_channel_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_channel_to_fp16 = cast(dtype = cache_last_channel_to_fp16_dtype_0, x = cache_last_channel)[name = tensor("cast_176")]; + tensor cache_1_cast_fp16 = slice_by_index(begin = cache_1_begin_0, end = cache_1_end_0, end_mask = cache_1_end_mask_0, squeeze_mask = cache_1_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_1_cast_fp16")]; + tensor cache_3_begin_0 = const()[name = tensor("cache_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_3_end_0 = const()[name = tensor("cache_3_end_0"), val = tensor([1, 1, 512, 8])]; + tensor cache_3_end_mask_0 = const()[name = tensor("cache_3_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_3_squeeze_mask_0 = const()[name = tensor("cache_3_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_time_to_fp16_dtype_0 = const()[name = tensor("cache_last_time_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_time_to_fp16 = cast(dtype = cache_last_time_to_fp16_dtype_0, x = cache_last_time)[name = tensor("cast_175")]; + tensor cache_3_cast_fp16 = slice_by_index(begin = cache_3_begin_0, end = cache_3_end_0, end_mask = cache_3_end_mask_0, squeeze_mask = cache_3_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_3_cast_fp16")]; + tensor input_3_axes_0 = const()[name = tensor("input_3_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320)))]; + tensor encoder_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1408)))]; + tensor var_21_to_fp16 = const()[name = tensor("op_21_to_fp16"), val = tensor(0x1.5p-17)]; + tensor input_3_cast_fp16 = layer_norm(axes = input_3_axes_0, beta = encoder_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_0_norm_feed_forward1_weight_to_fp16, x = var_83_cast_fp16)[name = tensor("input_3_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2496)))]; + tensor linear_0_bias_0_to_fp16 = const()[name = tensor("linear_0_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2099712)))]; + tensor linear_0_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor input_7_cast_fp16 = silu(x = linear_0_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2103872)))]; + tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4201088)))]; + tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_194_to_fp16 = const()[name = tensor("op_194_to_fp16"), val = tensor(0x1p-1)]; + tensor var_195_cast_fp16 = mul(x = linear_1_cast_fp16, y = var_194_to_fp16)[name = tensor("op_195_cast_fp16")]; + tensor input_13_cast_fp16 = add(x = var_83_cast_fp16, y = var_195_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor key_1_axes_0 = const()[name = tensor("key_1_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4202176)))]; + tensor encoder_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4203264)))]; + tensor key_1_cast_fp16 = layer_norm(axes = key_1_axes_0, beta = encoder_layers_0_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_0_norm_self_att_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("key_1_cast_fp16")]; + tensor input_15_interleave_0 = const()[name = tensor("input_15_interleave_0"), val = tensor(false)]; + tensor input_15_cast_fp16 = concat(axis = var_44, interleave = input_15_interleave_0, values = (cache_1_cast_fp16, key_1_cast_fp16))[name = tensor("input_15_cast_fp16")]; + tensor var_217_begin_0 = const()[name = tensor("op_217_begin_0"), val = tensor([0, 14, 0])]; + tensor var_217_end_0 = const()[name = tensor("op_217_end_0"), val = tensor([1, 70, 512])]; + tensor var_217_end_mask_0 = const()[name = tensor("op_217_end_mask_0"), val = tensor([true, true, true])]; + tensor var_217_cast_fp16 = slice_by_index(begin = var_217_begin_0, end = var_217_end_0, end_mask = var_217_end_mask_0, x = cache_1_cast_fp16)[name = tensor("op_217_cast_fp16")]; + tensor var_220_begin_0 = const()[name = tensor("op_220_begin_0"), val = tensor([0, 0, 0])]; + tensor var_220_end_0 = const()[name = tensor("op_220_end_0"), val = tensor([1, 14, 512])]; + tensor var_220_end_mask_0 = const()[name = tensor("op_220_end_mask_0"), val = tensor([true, false, true])]; + tensor var_220_cast_fp16 = slice_by_index(begin = var_220_begin_0, end = var_220_end_0, end_mask = var_220_end_mask_0, x = key_1_cast_fp16)[name = tensor("op_220_cast_fp16")]; + tensor var_223_interleave_0 = const()[name = tensor("op_223_interleave_0"), val = tensor(false)]; + tensor var_223_cast_fp16 = concat(axis = var_44, interleave = var_223_interleave_0, values = (var_217_cast_fp16, var_220_cast_fp16))[name = tensor("op_223_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4204352)))]; + tensor linear_2_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_q_weight_to_fp16, x = key_1_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_227 = const()[name = tensor("op_227"), val = tensor([1, -1, 8, 64])]; + tensor q_1_cast_fp16 = reshape(shape = var_227, x = linear_2_cast_fp16)[name = tensor("q_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4728704)))]; + tensor linear_3_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_k_weight_to_fp16, x = input_15_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor var_231 = const()[name = tensor("op_231"), val = tensor([1, -1, 8, 64])]; + tensor k_1_cast_fp16 = reshape(shape = var_231, x = linear_3_cast_fp16)[name = tensor("k_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5253056)))]; + tensor linear_4_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_v_weight_to_fp16, x = input_15_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor var_235 = const()[name = tensor("op_235"), val = tensor([1, -1, 8, 64])]; + tensor v_1_cast_fp16 = reshape(shape = var_235, x = linear_4_cast_fp16)[name = tensor("v_1_cast_fp16")]; + tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5777408)))]; + tensor var_247_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_247_cast_fp16")]; + tensor encoder_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5778496)))]; + tensor var_249_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_249_cast_fp16")]; + tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_3_transpose_x_0 = const()[name = tensor("x_3_transpose_x_0"), val = tensor(false)]; + tensor x_3_transpose_y_0 = const()[name = tensor("x_3_transpose_y_0"), val = tensor(false)]; + tensor var_251_to_fp16 = const()[name = tensor("op_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5779584)))]; + tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_249_cast_fp16)[name = tensor("transpose_237")]; + tensor x_3_cast_fp16 = matmul(transpose_x = x_3_transpose_x_0, transpose_y = x_3_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_251_to_fp16)[name = tensor("x_3_cast_fp16")]; + tensor x_5_pad_0 = const()[name = tensor("x_5_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_5_mode_0 = const()[name = tensor("x_5_mode_0"), val = tensor("constant")]; + tensor const_18_to_fp16 = const()[name = tensor("const_18_to_fp16"), val = tensor(0x0p+0)]; + tensor x_5_cast_fp16 = pad(constant_val = const_18_to_fp16, mode = x_5_mode_0, pad = x_5_pad_0, x = x_3_cast_fp16)[name = tensor("x_5_cast_fp16")]; + tensor var_259 = const()[name = tensor("op_259"), val = tensor([1, 8, -1, 16])]; + tensor x_7_cast_fp16 = reshape(shape = var_259, x = x_5_cast_fp16)[name = tensor("x_7_cast_fp16")]; + tensor var_263_begin_0 = const()[name = tensor("op_263_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_263_end_0 = const()[name = tensor("op_263_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_263_end_mask_0 = const()[name = tensor("op_263_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_263_cast_fp16 = slice_by_index(begin = var_263_begin_0, end = var_263_end_0, end_mask = var_263_end_mask_0, x = x_7_cast_fp16)[name = tensor("op_263_cast_fp16")]; + tensor var_264 = const()[name = tensor("op_264"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_1_cast_fp16 = reshape(shape = var_264, x = var_263_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; + tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; + tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_235")]; + tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_247_cast_fp16)[name = tensor("transpose_236")]; + tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; + tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; + tensor var_273_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_273_cast_fp16")]; + tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_1_cast_fp16 = mul(x = var_273_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; + tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; + tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; + tensor var_24_to_fp16 = const()[name = tensor("op_24_to_fp16"), val = tensor(-0x1.388p+13)]; + tensor scores_3_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_3)[name = tensor("scores_3_cast_fp16")]; + tensor var_279_cast_fp16 = softmax(axis = var_30, x = scores_3_cast_fp16)[name = tensor("op_279_cast_fp16")]; + tensor var_23_to_fp16 = const()[name = tensor("op_23_to_fp16"), val = tensor(0x0p+0)]; + tensor input_17_cast_fp16 = select(a = var_23_to_fp16, b = var_279_cast_fp16, cond = mask_3)[name = tensor("input_17_cast_fp16")]; + tensor x_9_transpose_x_0 = const()[name = tensor("x_9_transpose_x_0"), val = tensor(false)]; + tensor x_9_transpose_y_0 = const()[name = tensor("x_9_transpose_y_0"), val = tensor(false)]; + tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_238")]; + tensor x_9_cast_fp16 = matmul(transpose_x = x_9_transpose_x_0, transpose_y = x_9_transpose_y_0, x = input_17_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_9_cast_fp16")]; + tensor var_283_perm_0 = const()[name = tensor("op_283_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_284 = const()[name = tensor("op_284"), val = tensor([1, -1, 512])]; + tensor var_283_cast_fp16 = transpose(perm = var_283_perm_0, x = x_9_cast_fp16)[name = tensor("transpose_234")]; + tensor input_19_cast_fp16 = reshape(shape = var_284, x = var_283_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5954752)))]; + tensor linear_6_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_out_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("linear_6_cast_fp16")]; + tensor input_23_cast_fp16 = add(x = input_13_cast_fp16, y = linear_6_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor x_13_axes_0 = const()[name = tensor("x_13_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6479104)))]; + tensor encoder_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6480192)))]; + tensor x_13_cast_fp16 = layer_norm(axes = x_13_axes_0, beta = encoder_layers_0_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_0_norm_conv_weight_to_fp16, x = input_23_cast_fp16)[name = tensor("x_13_cast_fp16")]; + tensor input_25_perm_0 = const()[name = tensor("input_25_perm_0"), val = tensor([0, 2, 1])]; + tensor input_27_pad_type_0 = const()[name = tensor("input_27_pad_type_0"), val = tensor("valid")]; + tensor input_27_strides_0 = const()[name = tensor("input_27_strides_0"), val = tensor([1])]; + tensor input_27_pad_0 = const()[name = tensor("input_27_pad_0"), val = tensor([0, 0])]; + tensor input_27_dilations_0 = const()[name = tensor("input_27_dilations_0"), val = tensor([1])]; + tensor input_27_groups_0 = const()[name = tensor("input_27_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6481280)))]; + tensor input_25_cast_fp16 = transpose(perm = input_25_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_233")]; + tensor input_27_cast_fp16 = conv(dilations = input_27_dilations_0, groups = input_27_groups_0, pad = input_27_pad_0, pad_type = input_27_pad_type_0, strides = input_27_strides_0, weight = encoder_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor x_15_split_num_splits_0 = const()[name = tensor("x_15_split_num_splits_0"), val = tensor(2)]; + tensor x_15_split_axis_0 = const()[name = tensor("x_15_split_axis_0"), val = tensor(1)]; + tensor x_15_split_cast_fp16_0, tensor x_15_split_cast_fp16_1 = split(axis = x_15_split_axis_0, num_splits = x_15_split_num_splits_0, x = input_27_cast_fp16)[name = tensor("x_15_split_cast_fp16")]; + tensor x_15_split_1_sigmoid_cast_fp16 = sigmoid(x = x_15_split_cast_fp16_1)[name = tensor("x_15_split_1_sigmoid_cast_fp16")]; + tensor x_15_cast_fp16 = mul(x = x_15_split_cast_fp16_0, y = x_15_split_1_sigmoid_cast_fp16)[name = tensor("x_15_cast_fp16")]; + tensor var_309_axes_0 = const()[name = tensor("op_309_axes_0"), val = tensor([1])]; + tensor var_309 = expand_dims(axes = var_309_axes_0, x = pad_mask)[name = tensor("op_309")]; + tensor input_29_cast_fp16 = select(a = var_23_to_fp16, b = x_15_cast_fp16, cond = var_309)[name = tensor("input_29_cast_fp16")]; + tensor new_x_3_interleave_0 = const()[name = tensor("new_x_3_interleave_0"), val = tensor(false)]; + tensor new_x_3_cast_fp16 = concat(axis = var_30, interleave = new_x_3_interleave_0, values = (cache_3_cast_fp16, input_29_cast_fp16))[name = tensor("new_x_3_cast_fp16")]; + tensor next_cache_1_begin_0 = const()[name = tensor("next_cache_1_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_1_end_0 = const()[name = tensor("next_cache_1_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_1_end_mask_0 = const()[name = tensor("next_cache_1_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_1_cast_fp16 = slice_by_index(begin = next_cache_1_begin_0, end = next_cache_1_end_0, end_mask = next_cache_1_end_mask_0, x = new_x_3_cast_fp16)[name = tensor("next_cache_1_cast_fp16")]; + tensor var_325_begin_0 = const()[name = tensor("op_325_begin_0"), val = tensor([0, 0, 14])]; + tensor var_325_end_0 = const()[name = tensor("op_325_end_0"), val = tensor([1, 512, 22])]; + tensor var_325_end_mask_0 = const()[name = tensor("op_325_end_mask_0"), val = tensor([true, true, true])]; + tensor var_325_cast_fp16 = slice_by_index(begin = var_325_begin_0, end = var_325_end_0, end_mask = var_325_end_mask_0, x = next_cache_1_cast_fp16)[name = tensor("op_325_cast_fp16")]; + tensor x_17_pad_type_0 = const()[name = tensor("x_17_pad_type_0"), val = tensor("valid")]; + tensor x_17_groups_0 = const()[name = tensor("x_17_groups_0"), val = tensor(512)]; + tensor x_17_strides_0 = const()[name = tensor("x_17_strides_0"), val = tensor([1])]; + tensor x_17_pad_0 = const()[name = tensor("x_17_pad_0"), val = tensor([0, 0])]; + tensor x_17_dilations_0 = const()[name = tensor("x_17_dilations_0"), val = tensor([1])]; + tensor encoder_layers_0_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7529920)))]; + tensor x_17_cast_fp16 = conv(dilations = x_17_dilations_0, groups = x_17_groups_0, pad = x_17_pad_0, pad_type = x_17_pad_type_0, strides = x_17_strides_0, weight = encoder_layers_0_conv_depthwise_conv_weight_to_fp16, x = new_x_3_cast_fp16)[name = tensor("x_17_cast_fp16")]; + tensor input_31_perm_0 = const()[name = tensor("input_31_perm_0"), val = tensor([0, 2, 1])]; + tensor x_19_axes_0 = const()[name = tensor("x_19_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7539200)))]; + tensor encoder_layers_0_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7540288)))]; + tensor input_31_cast_fp16 = transpose(perm = input_31_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_232")]; + tensor x_19_cast_fp16 = layer_norm(axes = x_19_axes_0, beta = encoder_layers_0_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_0_conv_batch_norm_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("x_19_cast_fp16")]; + tensor input_33_perm_0 = const()[name = tensor("input_33_perm_0"), val = tensor([0, 2, 1])]; + tensor input_33_cast_fp16 = transpose(perm = input_33_perm_0, x = x_19_cast_fp16)[name = tensor("transpose_231")]; + tensor input_35_cast_fp16 = silu(x = input_33_cast_fp16)[name = tensor("input_35_cast_fp16")]; + tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; + tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; + tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; + tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; + tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7541376)))]; + tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = encoder_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("x_21_cast_fp16")]; + tensor input_37_perm_0 = const()[name = tensor("input_37_perm_0"), val = tensor([0, 2, 1])]; + tensor input_37_cast_fp16 = transpose(perm = input_37_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_230")]; + tensor input_39_cast_fp16 = add(x = input_23_cast_fp16, y = input_37_cast_fp16)[name = tensor("input_39_cast_fp16")]; + tensor input_41_axes_0 = const()[name = tensor("input_41_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8065728)))]; + tensor encoder_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8066816)))]; + tensor input_41_cast_fp16 = layer_norm(axes = input_41_axes_0, beta = encoder_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_0_norm_feed_forward2_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("input_41_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8067904)))]; + tensor linear_7_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor input_45_cast_fp16 = silu(x = linear_7_cast_fp16)[name = tensor("input_45_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10165120)))]; + tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_45_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor var_366_to_fp16 = const()[name = tensor("op_366_to_fp16"), val = tensor(0x1p-1)]; + tensor var_367_cast_fp16 = mul(x = linear_8_cast_fp16, y = var_366_to_fp16)[name = tensor("op_367_cast_fp16")]; + tensor input_51_cast_fp16 = add(x = input_39_cast_fp16, y = var_367_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor input_53_axes_0 = const()[name = tensor("input_53_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12262336)))]; + tensor encoder_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12263424)))]; + tensor input_53_cast_fp16 = layer_norm(axes = input_53_axes_0, beta = encoder_layers_0_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_0_norm_out_weight_to_fp16, x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; + tensor cache_5_begin_0 = const()[name = tensor("cache_5_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_5_end_0 = const()[name = tensor("cache_5_end_0"), val = tensor([2, 1, 70, 512])]; + tensor cache_5_end_mask_0 = const()[name = tensor("cache_5_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_5_squeeze_mask_0 = const()[name = tensor("cache_5_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_5_cast_fp16 = slice_by_index(begin = cache_5_begin_0, end = cache_5_end_0, end_mask = cache_5_end_mask_0, squeeze_mask = cache_5_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_5_cast_fp16")]; + tensor cache_7_begin_0 = const()[name = tensor("cache_7_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_7_end_0 = const()[name = tensor("cache_7_end_0"), val = tensor([2, 1, 512, 8])]; + tensor cache_7_end_mask_0 = const()[name = tensor("cache_7_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_7_squeeze_mask_0 = const()[name = tensor("cache_7_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_7_cast_fp16 = slice_by_index(begin = cache_7_begin_0, end = cache_7_end_0, end_mask = cache_7_end_mask_0, squeeze_mask = cache_7_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_7_cast_fp16")]; + tensor input_55_axes_0 = const()[name = tensor("input_55_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12264512)))]; + tensor encoder_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12265600)))]; + tensor input_55_cast_fp16 = layer_norm(axes = input_55_axes_0, beta = encoder_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_1_norm_feed_forward1_weight_to_fp16, x = input_53_cast_fp16)[name = tensor("input_55_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12266688)))]; + tensor linear_9_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_55_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor input_59_cast_fp16 = silu(x = linear_9_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14363904)))]; + tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor var_401_to_fp16 = const()[name = tensor("op_401_to_fp16"), val = tensor(0x1p-1)]; + tensor var_402_cast_fp16 = mul(x = linear_10_cast_fp16, y = var_401_to_fp16)[name = tensor("op_402_cast_fp16")]; + tensor input_65_cast_fp16 = add(x = input_53_cast_fp16, y = var_402_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor key_3_axes_0 = const()[name = tensor("key_3_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16461120)))]; + tensor encoder_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16462208)))]; + tensor key_3_cast_fp16 = layer_norm(axes = key_3_axes_0, beta = encoder_layers_1_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_1_norm_self_att_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("key_3_cast_fp16")]; + tensor input_67_interleave_0 = const()[name = tensor("input_67_interleave_0"), val = tensor(false)]; + tensor input_67_cast_fp16 = concat(axis = var_44, interleave = input_67_interleave_0, values = (cache_5_cast_fp16, key_3_cast_fp16))[name = tensor("input_67_cast_fp16")]; + tensor var_424_begin_0 = const()[name = tensor("op_424_begin_0"), val = tensor([0, 14, 0])]; + tensor var_424_end_0 = const()[name = tensor("op_424_end_0"), val = tensor([1, 70, 512])]; + tensor var_424_end_mask_0 = const()[name = tensor("op_424_end_mask_0"), val = tensor([true, true, true])]; + tensor var_424_cast_fp16 = slice_by_index(begin = var_424_begin_0, end = var_424_end_0, end_mask = var_424_end_mask_0, x = cache_5_cast_fp16)[name = tensor("op_424_cast_fp16")]; + tensor var_427_begin_0 = const()[name = tensor("op_427_begin_0"), val = tensor([0, 0, 0])]; + tensor var_427_end_0 = const()[name = tensor("op_427_end_0"), val = tensor([1, 14, 512])]; + tensor var_427_end_mask_0 = const()[name = tensor("op_427_end_mask_0"), val = tensor([true, false, true])]; + tensor var_427_cast_fp16 = slice_by_index(begin = var_427_begin_0, end = var_427_end_0, end_mask = var_427_end_mask_0, x = key_3_cast_fp16)[name = tensor("op_427_cast_fp16")]; + tensor var_430_interleave_0 = const()[name = tensor("op_430_interleave_0"), val = tensor(false)]; + tensor var_430_cast_fp16 = concat(axis = var_44, interleave = var_430_interleave_0, values = (var_424_cast_fp16, var_427_cast_fp16))[name = tensor("op_430_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16463296)))]; + tensor linear_11_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_q_weight_to_fp16, x = key_3_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor var_434 = const()[name = tensor("op_434"), val = tensor([1, -1, 8, 64])]; + tensor q_7_cast_fp16 = reshape(shape = var_434, x = linear_11_cast_fp16)[name = tensor("q_7_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16987648)))]; + tensor linear_12_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_k_weight_to_fp16, x = input_67_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_438 = const()[name = tensor("op_438"), val = tensor([1, -1, 8, 64])]; + tensor k_5_cast_fp16 = reshape(shape = var_438, x = linear_12_cast_fp16)[name = tensor("k_5_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17512000)))]; + tensor linear_13_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_v_weight_to_fp16, x = input_67_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_442 = const()[name = tensor("op_442"), val = tensor([1, -1, 8, 64])]; + tensor v_3_cast_fp16 = reshape(shape = var_442, x = linear_13_cast_fp16)[name = tensor("v_3_cast_fp16")]; + tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18036352)))]; + tensor var_454_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_454_cast_fp16")]; + tensor encoder_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18037440)))]; + tensor var_456_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_456_cast_fp16")]; + tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; + tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; + tensor var_458_to_fp16 = const()[name = tensor("op_458_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18038528)))]; + tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_456_cast_fp16)[name = tensor("transpose_228")]; + tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_458_to_fp16)[name = tensor("x_29_cast_fp16")]; + tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; + tensor const_31_to_fp16 = const()[name = tensor("const_31_to_fp16"), val = tensor(0x0p+0)]; + tensor x_31_cast_fp16 = pad(constant_val = const_31_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; + tensor var_466 = const()[name = tensor("op_466"), val = tensor([1, 8, -1, 16])]; + tensor x_33_cast_fp16 = reshape(shape = var_466, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; + tensor var_470_begin_0 = const()[name = tensor("op_470_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_470_end_0 = const()[name = tensor("op_470_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_470_end_mask_0 = const()[name = tensor("op_470_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_470_cast_fp16 = slice_by_index(begin = var_470_begin_0, end = var_470_end_0, end_mask = var_470_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_470_cast_fp16")]; + tensor var_471 = const()[name = tensor("op_471"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_5_cast_fp16 = reshape(shape = var_471, x = var_470_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; + tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; + tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_226")]; + tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_454_cast_fp16)[name = tensor("transpose_227")]; + tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; + tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; + tensor var_480_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_480_cast_fp16")]; + tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_5_cast_fp16 = mul(x = var_480_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; + tensor scores_7_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_3)[name = tensor("scores_7_cast_fp16")]; + tensor var_486_cast_fp16 = softmax(axis = var_30, x = scores_7_cast_fp16)[name = tensor("op_486_cast_fp16")]; + tensor input_69_cast_fp16 = select(a = var_23_to_fp16, b = var_486_cast_fp16, cond = mask_3)[name = tensor("input_69_cast_fp16")]; + tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; + tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; + tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_229")]; + tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_69_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; + tensor var_490_perm_0 = const()[name = tensor("op_490_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_491 = const()[name = tensor("op_491"), val = tensor([1, -1, 512])]; + tensor var_490_cast_fp16 = transpose(perm = var_490_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_225")]; + tensor input_71_cast_fp16 = reshape(shape = var_491, x = var_490_cast_fp16)[name = tensor("input_71_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18213696)))]; + tensor linear_15_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_out_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("linear_15_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_65_cast_fp16, y = linear_15_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18738048)))]; + tensor encoder_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18739136)))]; + tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = encoder_layers_1_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_1_norm_conv_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("x_39_cast_fp16")]; + tensor input_77_perm_0 = const()[name = tensor("input_77_perm_0"), val = tensor([0, 2, 1])]; + tensor input_79_pad_type_0 = const()[name = tensor("input_79_pad_type_0"), val = tensor("valid")]; + tensor input_79_strides_0 = const()[name = tensor("input_79_strides_0"), val = tensor([1])]; + tensor input_79_pad_0 = const()[name = tensor("input_79_pad_0"), val = tensor([0, 0])]; + tensor input_79_dilations_0 = const()[name = tensor("input_79_dilations_0"), val = tensor([1])]; + tensor input_79_groups_0 = const()[name = tensor("input_79_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18740224)))]; + tensor input_77_cast_fp16 = transpose(perm = input_77_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_224")]; + tensor input_79_cast_fp16 = conv(dilations = input_79_dilations_0, groups = input_79_groups_0, pad = input_79_pad_0, pad_type = input_79_pad_type_0, strides = input_79_strides_0, weight = encoder_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("input_79_cast_fp16")]; + tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; + tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; + tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_79_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; + tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; + tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; + tensor input_81_cast_fp16 = select(a = var_23_to_fp16, b = x_41_cast_fp16, cond = var_309)[name = tensor("input_81_cast_fp16")]; + tensor new_x_7_interleave_0 = const()[name = tensor("new_x_7_interleave_0"), val = tensor(false)]; + tensor new_x_7_cast_fp16 = concat(axis = var_30, interleave = new_x_7_interleave_0, values = (cache_7_cast_fp16, input_81_cast_fp16))[name = tensor("new_x_7_cast_fp16")]; + tensor next_cache_3_begin_0 = const()[name = tensor("next_cache_3_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_3_end_0 = const()[name = tensor("next_cache_3_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_3_end_mask_0 = const()[name = tensor("next_cache_3_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_3_cast_fp16 = slice_by_index(begin = next_cache_3_begin_0, end = next_cache_3_end_0, end_mask = next_cache_3_end_mask_0, x = new_x_7_cast_fp16)[name = tensor("next_cache_3_cast_fp16")]; + tensor var_532_begin_0 = const()[name = tensor("op_532_begin_0"), val = tensor([0, 0, 14])]; + tensor var_532_end_0 = const()[name = tensor("op_532_end_0"), val = tensor([1, 512, 22])]; + tensor var_532_end_mask_0 = const()[name = tensor("op_532_end_mask_0"), val = tensor([true, true, true])]; + tensor var_532_cast_fp16 = slice_by_index(begin = var_532_begin_0, end = var_532_end_0, end_mask = var_532_end_mask_0, x = next_cache_3_cast_fp16)[name = tensor("op_532_cast_fp16")]; + tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; + tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(512)]; + tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; + tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; + tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; + tensor encoder_layers_1_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19788864)))]; + tensor x_43_cast_fp16 = conv(dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = encoder_layers_1_conv_depthwise_conv_weight_to_fp16, x = new_x_7_cast_fp16)[name = tensor("x_43_cast_fp16")]; + tensor input_83_perm_0 = const()[name = tensor("input_83_perm_0"), val = tensor([0, 2, 1])]; + tensor x_45_axes_0 = const()[name = tensor("x_45_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19798144)))]; + tensor encoder_layers_1_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19799232)))]; + tensor input_83_cast_fp16 = transpose(perm = input_83_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_223")]; + tensor x_45_cast_fp16 = layer_norm(axes = x_45_axes_0, beta = encoder_layers_1_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_1_conv_batch_norm_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("x_45_cast_fp16")]; + tensor input_85_perm_0 = const()[name = tensor("input_85_perm_0"), val = tensor([0, 2, 1])]; + tensor input_85_cast_fp16 = transpose(perm = input_85_perm_0, x = x_45_cast_fp16)[name = tensor("transpose_222")]; + tensor input_87_cast_fp16 = silu(x = input_85_cast_fp16)[name = tensor("input_87_cast_fp16")]; + tensor x_47_pad_type_0 = const()[name = tensor("x_47_pad_type_0"), val = tensor("valid")]; + tensor x_47_strides_0 = const()[name = tensor("x_47_strides_0"), val = tensor([1])]; + tensor x_47_pad_0 = const()[name = tensor("x_47_pad_0"), val = tensor([0, 0])]; + tensor x_47_dilations_0 = const()[name = tensor("x_47_dilations_0"), val = tensor([1])]; + tensor x_47_groups_0 = const()[name = tensor("x_47_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19800320)))]; + tensor x_47_cast_fp16 = conv(dilations = x_47_dilations_0, groups = x_47_groups_0, pad = x_47_pad_0, pad_type = x_47_pad_type_0, strides = x_47_strides_0, weight = encoder_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_87_cast_fp16)[name = tensor("x_47_cast_fp16")]; + tensor input_89_perm_0 = const()[name = tensor("input_89_perm_0"), val = tensor([0, 2, 1])]; + tensor input_89_cast_fp16 = transpose(perm = input_89_perm_0, x = x_47_cast_fp16)[name = tensor("transpose_221")]; + tensor input_91_cast_fp16 = add(x = input_75_cast_fp16, y = input_89_cast_fp16)[name = tensor("input_91_cast_fp16")]; + tensor input_93_axes_0 = const()[name = tensor("input_93_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20324672)))]; + tensor encoder_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20325760)))]; + tensor input_93_cast_fp16 = layer_norm(axes = input_93_axes_0, beta = encoder_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_1_norm_feed_forward2_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("input_93_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20326848)))]; + tensor linear_16_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_97_cast_fp16 = silu(x = linear_16_cast_fp16)[name = tensor("input_97_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22424064)))]; + tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_97_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor var_573_to_fp16 = const()[name = tensor("op_573_to_fp16"), val = tensor(0x1p-1)]; + tensor var_574_cast_fp16 = mul(x = linear_17_cast_fp16, y = var_573_to_fp16)[name = tensor("op_574_cast_fp16")]; + tensor input_103_cast_fp16 = add(x = input_91_cast_fp16, y = var_574_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor input_105_axes_0 = const()[name = tensor("input_105_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24521280)))]; + tensor encoder_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24522368)))]; + tensor input_105_cast_fp16 = layer_norm(axes = input_105_axes_0, beta = encoder_layers_1_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_1_norm_out_weight_to_fp16, x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; + tensor cache_9_begin_0 = const()[name = tensor("cache_9_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_9_end_0 = const()[name = tensor("cache_9_end_0"), val = tensor([3, 1, 70, 512])]; + tensor cache_9_end_mask_0 = const()[name = tensor("cache_9_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_9_squeeze_mask_0 = const()[name = tensor("cache_9_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_9_cast_fp16 = slice_by_index(begin = cache_9_begin_0, end = cache_9_end_0, end_mask = cache_9_end_mask_0, squeeze_mask = cache_9_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_9_cast_fp16")]; + tensor cache_11_begin_0 = const()[name = tensor("cache_11_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_11_end_0 = const()[name = tensor("cache_11_end_0"), val = tensor([3, 1, 512, 8])]; + tensor cache_11_end_mask_0 = const()[name = tensor("cache_11_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_11_squeeze_mask_0 = const()[name = tensor("cache_11_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_11_cast_fp16 = slice_by_index(begin = cache_11_begin_0, end = cache_11_end_0, end_mask = cache_11_end_mask_0, squeeze_mask = cache_11_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_11_cast_fp16")]; + tensor input_107_axes_0 = const()[name = tensor("input_107_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24523456)))]; + tensor encoder_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24524544)))]; + tensor input_107_cast_fp16 = layer_norm(axes = input_107_axes_0, beta = encoder_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_2_norm_feed_forward1_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("input_107_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24525632)))]; + tensor linear_18_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor input_111_cast_fp16 = silu(x = linear_18_cast_fp16)[name = tensor("input_111_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26622848)))]; + tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor var_608_to_fp16 = const()[name = tensor("op_608_to_fp16"), val = tensor(0x1p-1)]; + tensor var_609_cast_fp16 = mul(x = linear_19_cast_fp16, y = var_608_to_fp16)[name = tensor("op_609_cast_fp16")]; + tensor input_117_cast_fp16 = add(x = input_105_cast_fp16, y = var_609_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor key_5_axes_0 = const()[name = tensor("key_5_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28720064)))]; + tensor encoder_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28721152)))]; + tensor key_5_cast_fp16 = layer_norm(axes = key_5_axes_0, beta = encoder_layers_2_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_2_norm_self_att_weight_to_fp16, x = input_117_cast_fp16)[name = tensor("key_5_cast_fp16")]; + tensor input_119_interleave_0 = const()[name = tensor("input_119_interleave_0"), val = tensor(false)]; + tensor input_119_cast_fp16 = concat(axis = var_44, interleave = input_119_interleave_0, values = (cache_9_cast_fp16, key_5_cast_fp16))[name = tensor("input_119_cast_fp16")]; + tensor var_631_begin_0 = const()[name = tensor("op_631_begin_0"), val = tensor([0, 14, 0])]; + tensor var_631_end_0 = const()[name = tensor("op_631_end_0"), val = tensor([1, 70, 512])]; + tensor var_631_end_mask_0 = const()[name = tensor("op_631_end_mask_0"), val = tensor([true, true, true])]; + tensor var_631_cast_fp16 = slice_by_index(begin = var_631_begin_0, end = var_631_end_0, end_mask = var_631_end_mask_0, x = cache_9_cast_fp16)[name = tensor("op_631_cast_fp16")]; + tensor var_634_begin_0 = const()[name = tensor("op_634_begin_0"), val = tensor([0, 0, 0])]; + tensor var_634_end_0 = const()[name = tensor("op_634_end_0"), val = tensor([1, 14, 512])]; + tensor var_634_end_mask_0 = const()[name = tensor("op_634_end_mask_0"), val = tensor([true, false, true])]; + tensor var_634_cast_fp16 = slice_by_index(begin = var_634_begin_0, end = var_634_end_0, end_mask = var_634_end_mask_0, x = key_5_cast_fp16)[name = tensor("op_634_cast_fp16")]; + tensor var_637_interleave_0 = const()[name = tensor("op_637_interleave_0"), val = tensor(false)]; + tensor var_637_cast_fp16 = concat(axis = var_44, interleave = var_637_interleave_0, values = (var_631_cast_fp16, var_634_cast_fp16))[name = tensor("op_637_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28722240)))]; + tensor linear_20_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_q_weight_to_fp16, x = key_5_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_641 = const()[name = tensor("op_641"), val = tensor([1, -1, 8, 64])]; + tensor q_13_cast_fp16 = reshape(shape = var_641, x = linear_20_cast_fp16)[name = tensor("q_13_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29246592)))]; + tensor linear_21_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_k_weight_to_fp16, x = input_119_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor var_645 = const()[name = tensor("op_645"), val = tensor([1, -1, 8, 64])]; + tensor k_9_cast_fp16 = reshape(shape = var_645, x = linear_21_cast_fp16)[name = tensor("k_9_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29770944)))]; + tensor linear_22_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_v_weight_to_fp16, x = input_119_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor var_649 = const()[name = tensor("op_649"), val = tensor([1, -1, 8, 64])]; + tensor v_5_cast_fp16 = reshape(shape = var_649, x = linear_22_cast_fp16)[name = tensor("v_5_cast_fp16")]; + tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30295296)))]; + tensor var_661_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_661_cast_fp16")]; + tensor encoder_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30296384)))]; + tensor var_663_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_663_cast_fp16")]; + tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_55_transpose_x_0 = const()[name = tensor("x_55_transpose_x_0"), val = tensor(false)]; + tensor x_55_transpose_y_0 = const()[name = tensor("x_55_transpose_y_0"), val = tensor(false)]; + tensor var_665_to_fp16 = const()[name = tensor("op_665_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30297472)))]; + tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_663_cast_fp16)[name = tensor("transpose_219")]; + tensor x_55_cast_fp16 = matmul(transpose_x = x_55_transpose_x_0, transpose_y = x_55_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_665_to_fp16)[name = tensor("x_55_cast_fp16")]; + tensor x_57_pad_0 = const()[name = tensor("x_57_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_57_mode_0 = const()[name = tensor("x_57_mode_0"), val = tensor("constant")]; + tensor const_44_to_fp16 = const()[name = tensor("const_44_to_fp16"), val = tensor(0x0p+0)]; + tensor x_57_cast_fp16 = pad(constant_val = const_44_to_fp16, mode = x_57_mode_0, pad = x_57_pad_0, x = x_55_cast_fp16)[name = tensor("x_57_cast_fp16")]; + tensor var_673 = const()[name = tensor("op_673"), val = tensor([1, 8, -1, 16])]; + tensor x_59_cast_fp16 = reshape(shape = var_673, x = x_57_cast_fp16)[name = tensor("x_59_cast_fp16")]; + tensor var_677_begin_0 = const()[name = tensor("op_677_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_677_end_0 = const()[name = tensor("op_677_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_677_end_mask_0 = const()[name = tensor("op_677_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_677_cast_fp16 = slice_by_index(begin = var_677_begin_0, end = var_677_end_0, end_mask = var_677_end_mask_0, x = x_59_cast_fp16)[name = tensor("op_677_cast_fp16")]; + tensor var_678 = const()[name = tensor("op_678"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_9_cast_fp16 = reshape(shape = var_678, x = var_677_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; + tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; + tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_217")]; + tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_661_cast_fp16)[name = tensor("transpose_218")]; + tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; + tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; + tensor var_687_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_687_cast_fp16")]; + tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_9_cast_fp16 = mul(x = var_687_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; + tensor scores_11_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_3)[name = tensor("scores_11_cast_fp16")]; + tensor var_693_cast_fp16 = softmax(axis = var_30, x = scores_11_cast_fp16)[name = tensor("op_693_cast_fp16")]; + tensor input_121_cast_fp16 = select(a = var_23_to_fp16, b = var_693_cast_fp16, cond = mask_3)[name = tensor("input_121_cast_fp16")]; + tensor x_61_transpose_x_0 = const()[name = tensor("x_61_transpose_x_0"), val = tensor(false)]; + tensor x_61_transpose_y_0 = const()[name = tensor("x_61_transpose_y_0"), val = tensor(false)]; + tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_220")]; + tensor x_61_cast_fp16 = matmul(transpose_x = x_61_transpose_x_0, transpose_y = x_61_transpose_y_0, x = input_121_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_61_cast_fp16")]; + tensor var_697_perm_0 = const()[name = tensor("op_697_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_698 = const()[name = tensor("op_698"), val = tensor([1, -1, 512])]; + tensor var_697_cast_fp16 = transpose(perm = var_697_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_216")]; + tensor input_123_cast_fp16 = reshape(shape = var_698, x = var_697_cast_fp16)[name = tensor("input_123_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30472640)))]; + tensor linear_24_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_out_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("linear_24_cast_fp16")]; + tensor input_127_cast_fp16 = add(x = input_117_cast_fp16, y = linear_24_cast_fp16)[name = tensor("input_127_cast_fp16")]; + tensor x_65_axes_0 = const()[name = tensor("x_65_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30996992)))]; + tensor encoder_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30998080)))]; + tensor x_65_cast_fp16 = layer_norm(axes = x_65_axes_0, beta = encoder_layers_2_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_2_norm_conv_weight_to_fp16, x = input_127_cast_fp16)[name = tensor("x_65_cast_fp16")]; + tensor input_129_perm_0 = const()[name = tensor("input_129_perm_0"), val = tensor([0, 2, 1])]; + tensor input_131_pad_type_0 = const()[name = tensor("input_131_pad_type_0"), val = tensor("valid")]; + tensor input_131_strides_0 = const()[name = tensor("input_131_strides_0"), val = tensor([1])]; + tensor input_131_pad_0 = const()[name = tensor("input_131_pad_0"), val = tensor([0, 0])]; + tensor input_131_dilations_0 = const()[name = tensor("input_131_dilations_0"), val = tensor([1])]; + tensor input_131_groups_0 = const()[name = tensor("input_131_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30999168)))]; + tensor input_129_cast_fp16 = transpose(perm = input_129_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_215")]; + tensor input_131_cast_fp16 = conv(dilations = input_131_dilations_0, groups = input_131_groups_0, pad = input_131_pad_0, pad_type = input_131_pad_type_0, strides = input_131_strides_0, weight = encoder_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor x_67_split_num_splits_0 = const()[name = tensor("x_67_split_num_splits_0"), val = tensor(2)]; + tensor x_67_split_axis_0 = const()[name = tensor("x_67_split_axis_0"), val = tensor(1)]; + tensor x_67_split_cast_fp16_0, tensor x_67_split_cast_fp16_1 = split(axis = x_67_split_axis_0, num_splits = x_67_split_num_splits_0, x = input_131_cast_fp16)[name = tensor("x_67_split_cast_fp16")]; + tensor x_67_split_1_sigmoid_cast_fp16 = sigmoid(x = x_67_split_cast_fp16_1)[name = tensor("x_67_split_1_sigmoid_cast_fp16")]; + tensor x_67_cast_fp16 = mul(x = x_67_split_cast_fp16_0, y = x_67_split_1_sigmoid_cast_fp16)[name = tensor("x_67_cast_fp16")]; + tensor input_133_cast_fp16 = select(a = var_23_to_fp16, b = x_67_cast_fp16, cond = var_309)[name = tensor("input_133_cast_fp16")]; + tensor new_x_11_interleave_0 = const()[name = tensor("new_x_11_interleave_0"), val = tensor(false)]; + tensor new_x_11_cast_fp16 = concat(axis = var_30, interleave = new_x_11_interleave_0, values = (cache_11_cast_fp16, input_133_cast_fp16))[name = tensor("new_x_11_cast_fp16")]; + tensor next_cache_5_begin_0 = const()[name = tensor("next_cache_5_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_5_end_0 = const()[name = tensor("next_cache_5_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_5_end_mask_0 = const()[name = tensor("next_cache_5_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_5_cast_fp16 = slice_by_index(begin = next_cache_5_begin_0, end = next_cache_5_end_0, end_mask = next_cache_5_end_mask_0, x = new_x_11_cast_fp16)[name = tensor("next_cache_5_cast_fp16")]; + tensor var_739_begin_0 = const()[name = tensor("op_739_begin_0"), val = tensor([0, 0, 14])]; + tensor var_739_end_0 = const()[name = tensor("op_739_end_0"), val = tensor([1, 512, 22])]; + tensor var_739_end_mask_0 = const()[name = tensor("op_739_end_mask_0"), val = tensor([true, true, true])]; + tensor var_739_cast_fp16 = slice_by_index(begin = var_739_begin_0, end = var_739_end_0, end_mask = var_739_end_mask_0, x = next_cache_5_cast_fp16)[name = tensor("op_739_cast_fp16")]; + tensor x_69_pad_type_0 = const()[name = tensor("x_69_pad_type_0"), val = tensor("valid")]; + tensor x_69_groups_0 = const()[name = tensor("x_69_groups_0"), val = tensor(512)]; + tensor x_69_strides_0 = const()[name = tensor("x_69_strides_0"), val = tensor([1])]; + tensor x_69_pad_0 = const()[name = tensor("x_69_pad_0"), val = tensor([0, 0])]; + tensor x_69_dilations_0 = const()[name = tensor("x_69_dilations_0"), val = tensor([1])]; + tensor encoder_layers_2_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32047808)))]; + tensor x_69_cast_fp16 = conv(dilations = x_69_dilations_0, groups = x_69_groups_0, pad = x_69_pad_0, pad_type = x_69_pad_type_0, strides = x_69_strides_0, weight = encoder_layers_2_conv_depthwise_conv_weight_to_fp16, x = new_x_11_cast_fp16)[name = tensor("x_69_cast_fp16")]; + tensor input_135_perm_0 = const()[name = tensor("input_135_perm_0"), val = tensor([0, 2, 1])]; + tensor x_71_axes_0 = const()[name = tensor("x_71_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32057088)))]; + tensor encoder_layers_2_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32058176)))]; + tensor input_135_cast_fp16 = transpose(perm = input_135_perm_0, x = x_69_cast_fp16)[name = tensor("transpose_214")]; + tensor x_71_cast_fp16 = layer_norm(axes = x_71_axes_0, beta = encoder_layers_2_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_2_conv_batch_norm_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("x_71_cast_fp16")]; + tensor input_137_perm_0 = const()[name = tensor("input_137_perm_0"), val = tensor([0, 2, 1])]; + tensor input_137_cast_fp16 = transpose(perm = input_137_perm_0, x = x_71_cast_fp16)[name = tensor("transpose_213")]; + tensor input_139_cast_fp16 = silu(x = input_137_cast_fp16)[name = tensor("input_139_cast_fp16")]; + tensor x_73_pad_type_0 = const()[name = tensor("x_73_pad_type_0"), val = tensor("valid")]; + tensor x_73_strides_0 = const()[name = tensor("x_73_strides_0"), val = tensor([1])]; + tensor x_73_pad_0 = const()[name = tensor("x_73_pad_0"), val = tensor([0, 0])]; + tensor x_73_dilations_0 = const()[name = tensor("x_73_dilations_0"), val = tensor([1])]; + tensor x_73_groups_0 = const()[name = tensor("x_73_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32059264)))]; + tensor x_73_cast_fp16 = conv(dilations = x_73_dilations_0, groups = x_73_groups_0, pad = x_73_pad_0, pad_type = x_73_pad_type_0, strides = x_73_strides_0, weight = encoder_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("x_73_cast_fp16")]; + tensor input_141_perm_0 = const()[name = tensor("input_141_perm_0"), val = tensor([0, 2, 1])]; + tensor input_141_cast_fp16 = transpose(perm = input_141_perm_0, x = x_73_cast_fp16)[name = tensor("transpose_212")]; + tensor input_143_cast_fp16 = add(x = input_127_cast_fp16, y = input_141_cast_fp16)[name = tensor("input_143_cast_fp16")]; + tensor input_145_axes_0 = const()[name = tensor("input_145_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32583616)))]; + tensor encoder_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32584704)))]; + tensor input_145_cast_fp16 = layer_norm(axes = input_145_axes_0, beta = encoder_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_2_norm_feed_forward2_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("input_145_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32585792)))]; + tensor linear_25_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor input_149_cast_fp16 = silu(x = linear_25_cast_fp16)[name = tensor("input_149_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34683008)))]; + tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_149_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor var_780_to_fp16 = const()[name = tensor("op_780_to_fp16"), val = tensor(0x1p-1)]; + tensor var_781_cast_fp16 = mul(x = linear_26_cast_fp16, y = var_780_to_fp16)[name = tensor("op_781_cast_fp16")]; + tensor input_155_cast_fp16 = add(x = input_143_cast_fp16, y = var_781_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor input_157_axes_0 = const()[name = tensor("input_157_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36780224)))]; + tensor encoder_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36781312)))]; + tensor input_157_cast_fp16 = layer_norm(axes = input_157_axes_0, beta = encoder_layers_2_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_2_norm_out_weight_to_fp16, x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; + tensor cache_13_begin_0 = const()[name = tensor("cache_13_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_13_end_0 = const()[name = tensor("cache_13_end_0"), val = tensor([4, 1, 70, 512])]; + tensor cache_13_end_mask_0 = const()[name = tensor("cache_13_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_13_squeeze_mask_0 = const()[name = tensor("cache_13_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_13_cast_fp16 = slice_by_index(begin = cache_13_begin_0, end = cache_13_end_0, end_mask = cache_13_end_mask_0, squeeze_mask = cache_13_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_13_cast_fp16")]; + tensor cache_15_begin_0 = const()[name = tensor("cache_15_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_15_end_0 = const()[name = tensor("cache_15_end_0"), val = tensor([4, 1, 512, 8])]; + tensor cache_15_end_mask_0 = const()[name = tensor("cache_15_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_15_squeeze_mask_0 = const()[name = tensor("cache_15_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_15_cast_fp16 = slice_by_index(begin = cache_15_begin_0, end = cache_15_end_0, end_mask = cache_15_end_mask_0, squeeze_mask = cache_15_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_15_cast_fp16")]; + tensor input_159_axes_0 = const()[name = tensor("input_159_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36782400)))]; + tensor encoder_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36783488)))]; + tensor input_159_cast_fp16 = layer_norm(axes = input_159_axes_0, beta = encoder_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_3_norm_feed_forward1_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("input_159_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36784576)))]; + tensor linear_27_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_159_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor input_163_cast_fp16 = silu(x = linear_27_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38881792)))]; + tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor var_815_to_fp16 = const()[name = tensor("op_815_to_fp16"), val = tensor(0x1p-1)]; + tensor var_816_cast_fp16 = mul(x = linear_28_cast_fp16, y = var_815_to_fp16)[name = tensor("op_816_cast_fp16")]; + tensor input_169_cast_fp16 = add(x = input_157_cast_fp16, y = var_816_cast_fp16)[name = tensor("input_169_cast_fp16")]; + tensor key_7_axes_0 = const()[name = tensor("key_7_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40979008)))]; + tensor encoder_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40980096)))]; + tensor key_7_cast_fp16 = layer_norm(axes = key_7_axes_0, beta = encoder_layers_3_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_3_norm_self_att_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("key_7_cast_fp16")]; + tensor input_171_interleave_0 = const()[name = tensor("input_171_interleave_0"), val = tensor(false)]; + tensor input_171_cast_fp16 = concat(axis = var_44, interleave = input_171_interleave_0, values = (cache_13_cast_fp16, key_7_cast_fp16))[name = tensor("input_171_cast_fp16")]; + tensor var_838_begin_0 = const()[name = tensor("op_838_begin_0"), val = tensor([0, 14, 0])]; + tensor var_838_end_0 = const()[name = tensor("op_838_end_0"), val = tensor([1, 70, 512])]; + tensor var_838_end_mask_0 = const()[name = tensor("op_838_end_mask_0"), val = tensor([true, true, true])]; + tensor var_838_cast_fp16 = slice_by_index(begin = var_838_begin_0, end = var_838_end_0, end_mask = var_838_end_mask_0, x = cache_13_cast_fp16)[name = tensor("op_838_cast_fp16")]; + tensor var_841_begin_0 = const()[name = tensor("op_841_begin_0"), val = tensor([0, 0, 0])]; + tensor var_841_end_0 = const()[name = tensor("op_841_end_0"), val = tensor([1, 14, 512])]; + tensor var_841_end_mask_0 = const()[name = tensor("op_841_end_mask_0"), val = tensor([true, false, true])]; + tensor var_841_cast_fp16 = slice_by_index(begin = var_841_begin_0, end = var_841_end_0, end_mask = var_841_end_mask_0, x = key_7_cast_fp16)[name = tensor("op_841_cast_fp16")]; + tensor var_844_interleave_0 = const()[name = tensor("op_844_interleave_0"), val = tensor(false)]; + tensor var_844_cast_fp16 = concat(axis = var_44, interleave = var_844_interleave_0, values = (var_838_cast_fp16, var_841_cast_fp16))[name = tensor("op_844_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40981184)))]; + tensor linear_29_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_q_weight_to_fp16, x = key_7_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor var_848 = const()[name = tensor("op_848"), val = tensor([1, -1, 8, 64])]; + tensor q_19_cast_fp16 = reshape(shape = var_848, x = linear_29_cast_fp16)[name = tensor("q_19_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41505536)))]; + tensor linear_30_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_k_weight_to_fp16, x = input_171_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_852 = const()[name = tensor("op_852"), val = tensor([1, -1, 8, 64])]; + tensor k_13_cast_fp16 = reshape(shape = var_852, x = linear_30_cast_fp16)[name = tensor("k_13_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42029888)))]; + tensor linear_31_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_v_weight_to_fp16, x = input_171_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_856 = const()[name = tensor("op_856"), val = tensor([1, -1, 8, 64])]; + tensor v_7_cast_fp16 = reshape(shape = var_856, x = linear_31_cast_fp16)[name = tensor("v_7_cast_fp16")]; + tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42554240)))]; + tensor var_868_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_868_cast_fp16")]; + tensor encoder_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42555328)))]; + tensor var_870_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_870_cast_fp16")]; + tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_81_transpose_x_0 = const()[name = tensor("x_81_transpose_x_0"), val = tensor(false)]; + tensor x_81_transpose_y_0 = const()[name = tensor("x_81_transpose_y_0"), val = tensor(false)]; + tensor var_872_to_fp16 = const()[name = tensor("op_872_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42556416)))]; + tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_870_cast_fp16)[name = tensor("transpose_210")]; + tensor x_81_cast_fp16 = matmul(transpose_x = x_81_transpose_x_0, transpose_y = x_81_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_872_to_fp16)[name = tensor("x_81_cast_fp16")]; + tensor x_83_pad_0 = const()[name = tensor("x_83_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_83_mode_0 = const()[name = tensor("x_83_mode_0"), val = tensor("constant")]; + tensor const_57_to_fp16 = const()[name = tensor("const_57_to_fp16"), val = tensor(0x0p+0)]; + tensor x_83_cast_fp16 = pad(constant_val = const_57_to_fp16, mode = x_83_mode_0, pad = x_83_pad_0, x = x_81_cast_fp16)[name = tensor("x_83_cast_fp16")]; + tensor var_880 = const()[name = tensor("op_880"), val = tensor([1, 8, -1, 16])]; + tensor x_85_cast_fp16 = reshape(shape = var_880, x = x_83_cast_fp16)[name = tensor("x_85_cast_fp16")]; + tensor var_884_begin_0 = const()[name = tensor("op_884_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_884_end_0 = const()[name = tensor("op_884_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_884_end_mask_0 = const()[name = tensor("op_884_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_884_cast_fp16 = slice_by_index(begin = var_884_begin_0, end = var_884_end_0, end_mask = var_884_end_mask_0, x = x_85_cast_fp16)[name = tensor("op_884_cast_fp16")]; + tensor var_885 = const()[name = tensor("op_885"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_13_cast_fp16 = reshape(shape = var_885, x = var_884_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; + tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; + tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_208")]; + tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_868_cast_fp16)[name = tensor("transpose_209")]; + tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; + tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; + tensor var_894_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_894_cast_fp16")]; + tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_13_cast_fp16 = mul(x = var_894_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; + tensor scores_15_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_3)[name = tensor("scores_15_cast_fp16")]; + tensor var_900_cast_fp16 = softmax(axis = var_30, x = scores_15_cast_fp16)[name = tensor("op_900_cast_fp16")]; + tensor input_173_cast_fp16 = select(a = var_23_to_fp16, b = var_900_cast_fp16, cond = mask_3)[name = tensor("input_173_cast_fp16")]; + tensor x_87_transpose_x_0 = const()[name = tensor("x_87_transpose_x_0"), val = tensor(false)]; + tensor x_87_transpose_y_0 = const()[name = tensor("x_87_transpose_y_0"), val = tensor(false)]; + tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_211")]; + tensor x_87_cast_fp16 = matmul(transpose_x = x_87_transpose_x_0, transpose_y = x_87_transpose_y_0, x = input_173_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_87_cast_fp16")]; + tensor var_904_perm_0 = const()[name = tensor("op_904_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_905 = const()[name = tensor("op_905"), val = tensor([1, -1, 512])]; + tensor var_904_cast_fp16 = transpose(perm = var_904_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_207")]; + tensor input_175_cast_fp16 = reshape(shape = var_905, x = var_904_cast_fp16)[name = tensor("input_175_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42731584)))]; + tensor linear_33_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_out_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("linear_33_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_169_cast_fp16, y = linear_33_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor x_91_axes_0 = const()[name = tensor("x_91_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43255936)))]; + tensor encoder_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43257024)))]; + tensor x_91_cast_fp16 = layer_norm(axes = x_91_axes_0, beta = encoder_layers_3_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_3_norm_conv_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("x_91_cast_fp16")]; + tensor input_181_perm_0 = const()[name = tensor("input_181_perm_0"), val = tensor([0, 2, 1])]; + tensor input_183_pad_type_0 = const()[name = tensor("input_183_pad_type_0"), val = tensor("valid")]; + tensor input_183_strides_0 = const()[name = tensor("input_183_strides_0"), val = tensor([1])]; + tensor input_183_pad_0 = const()[name = tensor("input_183_pad_0"), val = tensor([0, 0])]; + tensor input_183_dilations_0 = const()[name = tensor("input_183_dilations_0"), val = tensor([1])]; + tensor input_183_groups_0 = const()[name = tensor("input_183_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43258112)))]; + tensor input_181_cast_fp16 = transpose(perm = input_181_perm_0, x = x_91_cast_fp16)[name = tensor("transpose_206")]; + tensor input_183_cast_fp16 = conv(dilations = input_183_dilations_0, groups = input_183_groups_0, pad = input_183_pad_0, pad_type = input_183_pad_type_0, strides = input_183_strides_0, weight = encoder_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor x_93_split_num_splits_0 = const()[name = tensor("x_93_split_num_splits_0"), val = tensor(2)]; + tensor x_93_split_axis_0 = const()[name = tensor("x_93_split_axis_0"), val = tensor(1)]; + tensor x_93_split_cast_fp16_0, tensor x_93_split_cast_fp16_1 = split(axis = x_93_split_axis_0, num_splits = x_93_split_num_splits_0, x = input_183_cast_fp16)[name = tensor("x_93_split_cast_fp16")]; + tensor x_93_split_1_sigmoid_cast_fp16 = sigmoid(x = x_93_split_cast_fp16_1)[name = tensor("x_93_split_1_sigmoid_cast_fp16")]; + tensor x_93_cast_fp16 = mul(x = x_93_split_cast_fp16_0, y = x_93_split_1_sigmoid_cast_fp16)[name = tensor("x_93_cast_fp16")]; + tensor input_185_cast_fp16 = select(a = var_23_to_fp16, b = x_93_cast_fp16, cond = var_309)[name = tensor("input_185_cast_fp16")]; + tensor new_x_15_interleave_0 = const()[name = tensor("new_x_15_interleave_0"), val = tensor(false)]; + tensor new_x_15_cast_fp16 = concat(axis = var_30, interleave = new_x_15_interleave_0, values = (cache_15_cast_fp16, input_185_cast_fp16))[name = tensor("new_x_15_cast_fp16")]; + tensor next_cache_7_begin_0 = const()[name = tensor("next_cache_7_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_7_end_0 = const()[name = tensor("next_cache_7_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_7_end_mask_0 = const()[name = tensor("next_cache_7_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_7_cast_fp16 = slice_by_index(begin = next_cache_7_begin_0, end = next_cache_7_end_0, end_mask = next_cache_7_end_mask_0, x = new_x_15_cast_fp16)[name = tensor("next_cache_7_cast_fp16")]; + tensor var_946_begin_0 = const()[name = tensor("op_946_begin_0"), val = tensor([0, 0, 14])]; + tensor var_946_end_0 = const()[name = tensor("op_946_end_0"), val = tensor([1, 512, 22])]; + tensor var_946_end_mask_0 = const()[name = tensor("op_946_end_mask_0"), val = tensor([true, true, true])]; + tensor var_946_cast_fp16 = slice_by_index(begin = var_946_begin_0, end = var_946_end_0, end_mask = var_946_end_mask_0, x = next_cache_7_cast_fp16)[name = tensor("op_946_cast_fp16")]; + tensor x_95_pad_type_0 = const()[name = tensor("x_95_pad_type_0"), val = tensor("valid")]; + tensor x_95_groups_0 = const()[name = tensor("x_95_groups_0"), val = tensor(512)]; + tensor x_95_strides_0 = const()[name = tensor("x_95_strides_0"), val = tensor([1])]; + tensor x_95_pad_0 = const()[name = tensor("x_95_pad_0"), val = tensor([0, 0])]; + tensor x_95_dilations_0 = const()[name = tensor("x_95_dilations_0"), val = tensor([1])]; + tensor encoder_layers_3_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44306752)))]; + tensor x_95_cast_fp16 = conv(dilations = x_95_dilations_0, groups = x_95_groups_0, pad = x_95_pad_0, pad_type = x_95_pad_type_0, strides = x_95_strides_0, weight = encoder_layers_3_conv_depthwise_conv_weight_to_fp16, x = new_x_15_cast_fp16)[name = tensor("x_95_cast_fp16")]; + tensor input_187_perm_0 = const()[name = tensor("input_187_perm_0"), val = tensor([0, 2, 1])]; + tensor x_97_axes_0 = const()[name = tensor("x_97_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44316032)))]; + tensor encoder_layers_3_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44317120)))]; + tensor input_187_cast_fp16 = transpose(perm = input_187_perm_0, x = x_95_cast_fp16)[name = tensor("transpose_205")]; + tensor x_97_cast_fp16 = layer_norm(axes = x_97_axes_0, beta = encoder_layers_3_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_3_conv_batch_norm_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("x_97_cast_fp16")]; + tensor input_189_perm_0 = const()[name = tensor("input_189_perm_0"), val = tensor([0, 2, 1])]; + tensor input_189_cast_fp16 = transpose(perm = input_189_perm_0, x = x_97_cast_fp16)[name = tensor("transpose_204")]; + tensor input_191_cast_fp16 = silu(x = input_189_cast_fp16)[name = tensor("input_191_cast_fp16")]; + tensor x_99_pad_type_0 = const()[name = tensor("x_99_pad_type_0"), val = tensor("valid")]; + tensor x_99_strides_0 = const()[name = tensor("x_99_strides_0"), val = tensor([1])]; + tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0])]; + tensor x_99_dilations_0 = const()[name = tensor("x_99_dilations_0"), val = tensor([1])]; + tensor x_99_groups_0 = const()[name = tensor("x_99_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44318208)))]; + tensor x_99_cast_fp16 = conv(dilations = x_99_dilations_0, groups = x_99_groups_0, pad = x_99_pad_0, pad_type = x_99_pad_type_0, strides = x_99_strides_0, weight = encoder_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_191_cast_fp16)[name = tensor("x_99_cast_fp16")]; + tensor input_193_perm_0 = const()[name = tensor("input_193_perm_0"), val = tensor([0, 2, 1])]; + tensor input_193_cast_fp16 = transpose(perm = input_193_perm_0, x = x_99_cast_fp16)[name = tensor("transpose_203")]; + tensor input_195_cast_fp16 = add(x = input_179_cast_fp16, y = input_193_cast_fp16)[name = tensor("input_195_cast_fp16")]; + tensor input_197_axes_0 = const()[name = tensor("input_197_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44842560)))]; + tensor encoder_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44843648)))]; + tensor input_197_cast_fp16 = layer_norm(axes = input_197_axes_0, beta = encoder_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_3_norm_feed_forward2_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("input_197_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44844736)))]; + tensor linear_34_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_197_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_201_cast_fp16 = silu(x = linear_34_cast_fp16)[name = tensor("input_201_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46941952)))]; + tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_201_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor var_987_to_fp16 = const()[name = tensor("op_987_to_fp16"), val = tensor(0x1p-1)]; + tensor var_988_cast_fp16 = mul(x = linear_35_cast_fp16, y = var_987_to_fp16)[name = tensor("op_988_cast_fp16")]; + tensor input_207_cast_fp16 = add(x = input_195_cast_fp16, y = var_988_cast_fp16)[name = tensor("input_207_cast_fp16")]; + tensor input_209_axes_0 = const()[name = tensor("input_209_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49039168)))]; + tensor encoder_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49040256)))]; + tensor input_209_cast_fp16 = layer_norm(axes = input_209_axes_0, beta = encoder_layers_3_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_3_norm_out_weight_to_fp16, x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; + tensor cache_17_begin_0 = const()[name = tensor("cache_17_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_17_end_0 = const()[name = tensor("cache_17_end_0"), val = tensor([5, 1, 70, 512])]; + tensor cache_17_end_mask_0 = const()[name = tensor("cache_17_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_17_squeeze_mask_0 = const()[name = tensor("cache_17_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_17_cast_fp16 = slice_by_index(begin = cache_17_begin_0, end = cache_17_end_0, end_mask = cache_17_end_mask_0, squeeze_mask = cache_17_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_17_cast_fp16")]; + tensor cache_19_begin_0 = const()[name = tensor("cache_19_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_19_end_0 = const()[name = tensor("cache_19_end_0"), val = tensor([5, 1, 512, 8])]; + tensor cache_19_end_mask_0 = const()[name = tensor("cache_19_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_19_squeeze_mask_0 = const()[name = tensor("cache_19_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_19_cast_fp16 = slice_by_index(begin = cache_19_begin_0, end = cache_19_end_0, end_mask = cache_19_end_mask_0, squeeze_mask = cache_19_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_19_cast_fp16")]; + tensor input_211_axes_0 = const()[name = tensor("input_211_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49041344)))]; + tensor encoder_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49042432)))]; + tensor input_211_cast_fp16 = layer_norm(axes = input_211_axes_0, beta = encoder_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_4_norm_feed_forward1_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("input_211_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49043520)))]; + tensor linear_36_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor input_215_cast_fp16 = silu(x = linear_36_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51140736)))]; + tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor var_1022_to_fp16 = const()[name = tensor("op_1022_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1023_cast_fp16 = mul(x = linear_37_cast_fp16, y = var_1022_to_fp16)[name = tensor("op_1023_cast_fp16")]; + tensor input_221_cast_fp16 = add(x = input_209_cast_fp16, y = var_1023_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor key_9_axes_0 = const()[name = tensor("key_9_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53237952)))]; + tensor encoder_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53239040)))]; + tensor key_9_cast_fp16 = layer_norm(axes = key_9_axes_0, beta = encoder_layers_4_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_4_norm_self_att_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("key_9_cast_fp16")]; + tensor input_223_interleave_0 = const()[name = tensor("input_223_interleave_0"), val = tensor(false)]; + tensor input_223_cast_fp16 = concat(axis = var_44, interleave = input_223_interleave_0, values = (cache_17_cast_fp16, key_9_cast_fp16))[name = tensor("input_223_cast_fp16")]; + tensor var_1045_begin_0 = const()[name = tensor("op_1045_begin_0"), val = tensor([0, 14, 0])]; + tensor var_1045_end_0 = const()[name = tensor("op_1045_end_0"), val = tensor([1, 70, 512])]; + tensor var_1045_end_mask_0 = const()[name = tensor("op_1045_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1045_cast_fp16 = slice_by_index(begin = var_1045_begin_0, end = var_1045_end_0, end_mask = var_1045_end_mask_0, x = cache_17_cast_fp16)[name = tensor("op_1045_cast_fp16")]; + tensor var_1048_begin_0 = const()[name = tensor("op_1048_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1048_end_0 = const()[name = tensor("op_1048_end_0"), val = tensor([1, 14, 512])]; + tensor var_1048_end_mask_0 = const()[name = tensor("op_1048_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1048_cast_fp16 = slice_by_index(begin = var_1048_begin_0, end = var_1048_end_0, end_mask = var_1048_end_mask_0, x = key_9_cast_fp16)[name = tensor("op_1048_cast_fp16")]; + tensor var_1051_interleave_0 = const()[name = tensor("op_1051_interleave_0"), val = tensor(false)]; + tensor var_1051_cast_fp16 = concat(axis = var_44, interleave = var_1051_interleave_0, values = (var_1045_cast_fp16, var_1048_cast_fp16))[name = tensor("op_1051_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53240128)))]; + tensor linear_38_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_q_weight_to_fp16, x = key_9_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_1055 = const()[name = tensor("op_1055"), val = tensor([1, -1, 8, 64])]; + tensor q_25_cast_fp16 = reshape(shape = var_1055, x = linear_38_cast_fp16)[name = tensor("q_25_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53764480)))]; + tensor linear_39_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_k_weight_to_fp16, x = input_223_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor var_1059 = const()[name = tensor("op_1059"), val = tensor([1, -1, 8, 64])]; + tensor k_17_cast_fp16 = reshape(shape = var_1059, x = linear_39_cast_fp16)[name = tensor("k_17_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54288832)))]; + tensor linear_40_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_v_weight_to_fp16, x = input_223_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor var_1063 = const()[name = tensor("op_1063"), val = tensor([1, -1, 8, 64])]; + tensor v_9_cast_fp16 = reshape(shape = var_1063, x = linear_40_cast_fp16)[name = tensor("v_9_cast_fp16")]; + tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54813184)))]; + tensor var_1075_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1075_cast_fp16")]; + tensor encoder_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54814272)))]; + tensor var_1077_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1077_cast_fp16")]; + tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_107_transpose_x_0 = const()[name = tensor("x_107_transpose_x_0"), val = tensor(false)]; + tensor x_107_transpose_y_0 = const()[name = tensor("x_107_transpose_y_0"), val = tensor(false)]; + tensor var_1079_to_fp16 = const()[name = tensor("op_1079_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54815360)))]; + tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1077_cast_fp16)[name = tensor("transpose_201")]; + tensor x_107_cast_fp16 = matmul(transpose_x = x_107_transpose_x_0, transpose_y = x_107_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1079_to_fp16)[name = tensor("x_107_cast_fp16")]; + tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_109_mode_0 = const()[name = tensor("x_109_mode_0"), val = tensor("constant")]; + tensor const_70_to_fp16 = const()[name = tensor("const_70_to_fp16"), val = tensor(0x0p+0)]; + tensor x_109_cast_fp16 = pad(constant_val = const_70_to_fp16, mode = x_109_mode_0, pad = x_109_pad_0, x = x_107_cast_fp16)[name = tensor("x_109_cast_fp16")]; + tensor var_1087 = const()[name = tensor("op_1087"), val = tensor([1, 8, -1, 16])]; + tensor x_111_cast_fp16 = reshape(shape = var_1087, x = x_109_cast_fp16)[name = tensor("x_111_cast_fp16")]; + tensor var_1091_begin_0 = const()[name = tensor("op_1091_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1091_end_0 = const()[name = tensor("op_1091_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_1091_end_mask_0 = const()[name = tensor("op_1091_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1091_cast_fp16 = slice_by_index(begin = var_1091_begin_0, end = var_1091_end_0, end_mask = var_1091_end_mask_0, x = x_111_cast_fp16)[name = tensor("op_1091_cast_fp16")]; + tensor var_1092 = const()[name = tensor("op_1092"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1092, x = var_1091_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; + tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; + tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_199")]; + tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1075_cast_fp16)[name = tensor("transpose_200")]; + tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; + tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; + tensor var_1101_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1101_cast_fp16")]; + tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_17_cast_fp16 = mul(x = var_1101_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; + tensor scores_19_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_3)[name = tensor("scores_19_cast_fp16")]; + tensor var_1107_cast_fp16 = softmax(axis = var_30, x = scores_19_cast_fp16)[name = tensor("op_1107_cast_fp16")]; + tensor input_225_cast_fp16 = select(a = var_23_to_fp16, b = var_1107_cast_fp16, cond = mask_3)[name = tensor("input_225_cast_fp16")]; + tensor x_113_transpose_x_0 = const()[name = tensor("x_113_transpose_x_0"), val = tensor(false)]; + tensor x_113_transpose_y_0 = const()[name = tensor("x_113_transpose_y_0"), val = tensor(false)]; + tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_202")]; + tensor x_113_cast_fp16 = matmul(transpose_x = x_113_transpose_x_0, transpose_y = x_113_transpose_y_0, x = input_225_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_113_cast_fp16")]; + tensor var_1111_perm_0 = const()[name = tensor("op_1111_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1112 = const()[name = tensor("op_1112"), val = tensor([1, -1, 512])]; + tensor var_1111_cast_fp16 = transpose(perm = var_1111_perm_0, x = x_113_cast_fp16)[name = tensor("transpose_198")]; + tensor input_227_cast_fp16 = reshape(shape = var_1112, x = var_1111_cast_fp16)[name = tensor("input_227_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54990528)))]; + tensor linear_42_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_out_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("linear_42_cast_fp16")]; + tensor input_231_cast_fp16 = add(x = input_221_cast_fp16, y = linear_42_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor x_117_axes_0 = const()[name = tensor("x_117_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55514880)))]; + tensor encoder_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55515968)))]; + tensor x_117_cast_fp16 = layer_norm(axes = x_117_axes_0, beta = encoder_layers_4_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_4_norm_conv_weight_to_fp16, x = input_231_cast_fp16)[name = tensor("x_117_cast_fp16")]; + tensor input_233_perm_0 = const()[name = tensor("input_233_perm_0"), val = tensor([0, 2, 1])]; + tensor input_235_pad_type_0 = const()[name = tensor("input_235_pad_type_0"), val = tensor("valid")]; + tensor input_235_strides_0 = const()[name = tensor("input_235_strides_0"), val = tensor([1])]; + tensor input_235_pad_0 = const()[name = tensor("input_235_pad_0"), val = tensor([0, 0])]; + tensor input_235_dilations_0 = const()[name = tensor("input_235_dilations_0"), val = tensor([1])]; + tensor input_235_groups_0 = const()[name = tensor("input_235_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55517056)))]; + tensor input_233_cast_fp16 = transpose(perm = input_233_perm_0, x = x_117_cast_fp16)[name = tensor("transpose_197")]; + tensor input_235_cast_fp16 = conv(dilations = input_235_dilations_0, groups = input_235_groups_0, pad = input_235_pad_0, pad_type = input_235_pad_type_0, strides = input_235_strides_0, weight = encoder_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor x_119_split_num_splits_0 = const()[name = tensor("x_119_split_num_splits_0"), val = tensor(2)]; + tensor x_119_split_axis_0 = const()[name = tensor("x_119_split_axis_0"), val = tensor(1)]; + tensor x_119_split_cast_fp16_0, tensor x_119_split_cast_fp16_1 = split(axis = x_119_split_axis_0, num_splits = x_119_split_num_splits_0, x = input_235_cast_fp16)[name = tensor("x_119_split_cast_fp16")]; + tensor x_119_split_1_sigmoid_cast_fp16 = sigmoid(x = x_119_split_cast_fp16_1)[name = tensor("x_119_split_1_sigmoid_cast_fp16")]; + tensor x_119_cast_fp16 = mul(x = x_119_split_cast_fp16_0, y = x_119_split_1_sigmoid_cast_fp16)[name = tensor("x_119_cast_fp16")]; + tensor input_237_cast_fp16 = select(a = var_23_to_fp16, b = x_119_cast_fp16, cond = var_309)[name = tensor("input_237_cast_fp16")]; + tensor new_x_19_interleave_0 = const()[name = tensor("new_x_19_interleave_0"), val = tensor(false)]; + tensor new_x_19_cast_fp16 = concat(axis = var_30, interleave = new_x_19_interleave_0, values = (cache_19_cast_fp16, input_237_cast_fp16))[name = tensor("new_x_19_cast_fp16")]; + tensor next_cache_9_begin_0 = const()[name = tensor("next_cache_9_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_9_end_0 = const()[name = tensor("next_cache_9_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_9_end_mask_0 = const()[name = tensor("next_cache_9_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_9_cast_fp16 = slice_by_index(begin = next_cache_9_begin_0, end = next_cache_9_end_0, end_mask = next_cache_9_end_mask_0, x = new_x_19_cast_fp16)[name = tensor("next_cache_9_cast_fp16")]; + tensor var_1153_begin_0 = const()[name = tensor("op_1153_begin_0"), val = tensor([0, 0, 14])]; + tensor var_1153_end_0 = const()[name = tensor("op_1153_end_0"), val = tensor([1, 512, 22])]; + tensor var_1153_end_mask_0 = const()[name = tensor("op_1153_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1153_cast_fp16 = slice_by_index(begin = var_1153_begin_0, end = var_1153_end_0, end_mask = var_1153_end_mask_0, x = next_cache_9_cast_fp16)[name = tensor("op_1153_cast_fp16")]; + tensor x_121_pad_type_0 = const()[name = tensor("x_121_pad_type_0"), val = tensor("valid")]; + tensor x_121_groups_0 = const()[name = tensor("x_121_groups_0"), val = tensor(512)]; + tensor x_121_strides_0 = const()[name = tensor("x_121_strides_0"), val = tensor([1])]; + tensor x_121_pad_0 = const()[name = tensor("x_121_pad_0"), val = tensor([0, 0])]; + tensor x_121_dilations_0 = const()[name = tensor("x_121_dilations_0"), val = tensor([1])]; + tensor encoder_layers_4_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56565696)))]; + tensor x_121_cast_fp16 = conv(dilations = x_121_dilations_0, groups = x_121_groups_0, pad = x_121_pad_0, pad_type = x_121_pad_type_0, strides = x_121_strides_0, weight = encoder_layers_4_conv_depthwise_conv_weight_to_fp16, x = new_x_19_cast_fp16)[name = tensor("x_121_cast_fp16")]; + tensor input_239_perm_0 = const()[name = tensor("input_239_perm_0"), val = tensor([0, 2, 1])]; + tensor x_123_axes_0 = const()[name = tensor("x_123_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56574976)))]; + tensor encoder_layers_4_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56576064)))]; + tensor input_239_cast_fp16 = transpose(perm = input_239_perm_0, x = x_121_cast_fp16)[name = tensor("transpose_196")]; + tensor x_123_cast_fp16 = layer_norm(axes = x_123_axes_0, beta = encoder_layers_4_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_4_conv_batch_norm_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("x_123_cast_fp16")]; + tensor input_241_perm_0 = const()[name = tensor("input_241_perm_0"), val = tensor([0, 2, 1])]; + tensor input_241_cast_fp16 = transpose(perm = input_241_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_195")]; + tensor input_243_cast_fp16 = silu(x = input_241_cast_fp16)[name = tensor("input_243_cast_fp16")]; + tensor x_125_pad_type_0 = const()[name = tensor("x_125_pad_type_0"), val = tensor("valid")]; + tensor x_125_strides_0 = const()[name = tensor("x_125_strides_0"), val = tensor([1])]; + tensor x_125_pad_0 = const()[name = tensor("x_125_pad_0"), val = tensor([0, 0])]; + tensor x_125_dilations_0 = const()[name = tensor("x_125_dilations_0"), val = tensor([1])]; + tensor x_125_groups_0 = const()[name = tensor("x_125_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56577152)))]; + tensor x_125_cast_fp16 = conv(dilations = x_125_dilations_0, groups = x_125_groups_0, pad = x_125_pad_0, pad_type = x_125_pad_type_0, strides = x_125_strides_0, weight = encoder_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("x_125_cast_fp16")]; + tensor input_245_perm_0 = const()[name = tensor("input_245_perm_0"), val = tensor([0, 2, 1])]; + tensor input_245_cast_fp16 = transpose(perm = input_245_perm_0, x = x_125_cast_fp16)[name = tensor("transpose_194")]; + tensor input_247_cast_fp16 = add(x = input_231_cast_fp16, y = input_245_cast_fp16)[name = tensor("input_247_cast_fp16")]; + tensor input_249_axes_0 = const()[name = tensor("input_249_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57101504)))]; + tensor encoder_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57102592)))]; + tensor input_249_cast_fp16 = layer_norm(axes = input_249_axes_0, beta = encoder_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_4_norm_feed_forward2_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("input_249_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57103680)))]; + tensor linear_43_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor input_253_cast_fp16 = silu(x = linear_43_cast_fp16)[name = tensor("input_253_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59200896)))]; + tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_253_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor var_1194_to_fp16 = const()[name = tensor("op_1194_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1195_cast_fp16 = mul(x = linear_44_cast_fp16, y = var_1194_to_fp16)[name = tensor("op_1195_cast_fp16")]; + tensor input_259_cast_fp16 = add(x = input_247_cast_fp16, y = var_1195_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor input_261_axes_0 = const()[name = tensor("input_261_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61298112)))]; + tensor encoder_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61299200)))]; + tensor input_261_cast_fp16 = layer_norm(axes = input_261_axes_0, beta = encoder_layers_4_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_4_norm_out_weight_to_fp16, x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; + tensor cache_21_begin_0 = const()[name = tensor("cache_21_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_21_end_0 = const()[name = tensor("cache_21_end_0"), val = tensor([6, 1, 70, 512])]; + tensor cache_21_end_mask_0 = const()[name = tensor("cache_21_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_21_squeeze_mask_0 = const()[name = tensor("cache_21_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_21_cast_fp16 = slice_by_index(begin = cache_21_begin_0, end = cache_21_end_0, end_mask = cache_21_end_mask_0, squeeze_mask = cache_21_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_21_cast_fp16")]; + tensor cache_23_begin_0 = const()[name = tensor("cache_23_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_23_end_0 = const()[name = tensor("cache_23_end_0"), val = tensor([6, 1, 512, 8])]; + tensor cache_23_end_mask_0 = const()[name = tensor("cache_23_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_23_squeeze_mask_0 = const()[name = tensor("cache_23_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_23_cast_fp16 = slice_by_index(begin = cache_23_begin_0, end = cache_23_end_0, end_mask = cache_23_end_mask_0, squeeze_mask = cache_23_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_23_cast_fp16")]; + tensor input_263_axes_0 = const()[name = tensor("input_263_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61300288)))]; + tensor encoder_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61301376)))]; + tensor input_263_cast_fp16 = layer_norm(axes = input_263_axes_0, beta = encoder_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_5_norm_feed_forward1_weight_to_fp16, x = input_261_cast_fp16)[name = tensor("input_263_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61302464)))]; + tensor linear_45_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_263_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor input_267_cast_fp16 = silu(x = linear_45_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63399680)))]; + tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor var_1229_to_fp16 = const()[name = tensor("op_1229_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1230_cast_fp16 = mul(x = linear_46_cast_fp16, y = var_1229_to_fp16)[name = tensor("op_1230_cast_fp16")]; + tensor input_273_cast_fp16 = add(x = input_261_cast_fp16, y = var_1230_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor key_11_axes_0 = const()[name = tensor("key_11_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65496896)))]; + tensor encoder_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65497984)))]; + tensor key_11_cast_fp16 = layer_norm(axes = key_11_axes_0, beta = encoder_layers_5_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_5_norm_self_att_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("key_11_cast_fp16")]; + tensor input_275_interleave_0 = const()[name = tensor("input_275_interleave_0"), val = tensor(false)]; + tensor input_275_cast_fp16 = concat(axis = var_44, interleave = input_275_interleave_0, values = (cache_21_cast_fp16, key_11_cast_fp16))[name = tensor("input_275_cast_fp16")]; + tensor var_1252_begin_0 = const()[name = tensor("op_1252_begin_0"), val = tensor([0, 14, 0])]; + tensor var_1252_end_0 = const()[name = tensor("op_1252_end_0"), val = tensor([1, 70, 512])]; + tensor var_1252_end_mask_0 = const()[name = tensor("op_1252_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1252_cast_fp16 = slice_by_index(begin = var_1252_begin_0, end = var_1252_end_0, end_mask = var_1252_end_mask_0, x = cache_21_cast_fp16)[name = tensor("op_1252_cast_fp16")]; + tensor var_1255_begin_0 = const()[name = tensor("op_1255_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1255_end_0 = const()[name = tensor("op_1255_end_0"), val = tensor([1, 14, 512])]; + tensor var_1255_end_mask_0 = const()[name = tensor("op_1255_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1255_cast_fp16 = slice_by_index(begin = var_1255_begin_0, end = var_1255_end_0, end_mask = var_1255_end_mask_0, x = key_11_cast_fp16)[name = tensor("op_1255_cast_fp16")]; + tensor var_1258_interleave_0 = const()[name = tensor("op_1258_interleave_0"), val = tensor(false)]; + tensor var_1258_cast_fp16 = concat(axis = var_44, interleave = var_1258_interleave_0, values = (var_1252_cast_fp16, var_1255_cast_fp16))[name = tensor("op_1258_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65499072)))]; + tensor linear_47_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_q_weight_to_fp16, x = key_11_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor var_1262 = const()[name = tensor("op_1262"), val = tensor([1, -1, 8, 64])]; + tensor q_31_cast_fp16 = reshape(shape = var_1262, x = linear_47_cast_fp16)[name = tensor("q_31_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66023424)))]; + tensor linear_48_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_k_weight_to_fp16, x = input_275_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_1266 = const()[name = tensor("op_1266"), val = tensor([1, -1, 8, 64])]; + tensor k_21_cast_fp16 = reshape(shape = var_1266, x = linear_48_cast_fp16)[name = tensor("k_21_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66547776)))]; + tensor linear_49_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_v_weight_to_fp16, x = input_275_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_1270 = const()[name = tensor("op_1270"), val = tensor([1, -1, 8, 64])]; + tensor v_11_cast_fp16 = reshape(shape = var_1270, x = linear_49_cast_fp16)[name = tensor("v_11_cast_fp16")]; + tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67072128)))]; + tensor var_1282_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1282_cast_fp16")]; + tensor encoder_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67073216)))]; + tensor var_1284_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1284_cast_fp16")]; + tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_133_transpose_x_0 = const()[name = tensor("x_133_transpose_x_0"), val = tensor(false)]; + tensor x_133_transpose_y_0 = const()[name = tensor("x_133_transpose_y_0"), val = tensor(false)]; + tensor var_1286_to_fp16 = const()[name = tensor("op_1286_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67074304)))]; + tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1284_cast_fp16)[name = tensor("transpose_192")]; + tensor x_133_cast_fp16 = matmul(transpose_x = x_133_transpose_x_0, transpose_y = x_133_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1286_to_fp16)[name = tensor("x_133_cast_fp16")]; + tensor x_135_pad_0 = const()[name = tensor("x_135_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_135_mode_0 = const()[name = tensor("x_135_mode_0"), val = tensor("constant")]; + tensor const_83_to_fp16 = const()[name = tensor("const_83_to_fp16"), val = tensor(0x0p+0)]; + tensor x_135_cast_fp16 = pad(constant_val = const_83_to_fp16, mode = x_135_mode_0, pad = x_135_pad_0, x = x_133_cast_fp16)[name = tensor("x_135_cast_fp16")]; + tensor var_1294 = const()[name = tensor("op_1294"), val = tensor([1, 8, -1, 16])]; + tensor x_137_cast_fp16 = reshape(shape = var_1294, x = x_135_cast_fp16)[name = tensor("x_137_cast_fp16")]; + tensor var_1298_begin_0 = const()[name = tensor("op_1298_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1298_end_0 = const()[name = tensor("op_1298_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_1298_end_mask_0 = const()[name = tensor("op_1298_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1298_cast_fp16 = slice_by_index(begin = var_1298_begin_0, end = var_1298_end_0, end_mask = var_1298_end_mask_0, x = x_137_cast_fp16)[name = tensor("op_1298_cast_fp16")]; + tensor var_1299 = const()[name = tensor("op_1299"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1299, x = var_1298_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; + tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; + tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_190")]; + tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1282_cast_fp16)[name = tensor("transpose_191")]; + tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; + tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; + tensor var_1308_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1308_cast_fp16")]; + tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_21_cast_fp16 = mul(x = var_1308_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; + tensor scores_23_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_3)[name = tensor("scores_23_cast_fp16")]; + tensor var_1314_cast_fp16 = softmax(axis = var_30, x = scores_23_cast_fp16)[name = tensor("op_1314_cast_fp16")]; + tensor input_277_cast_fp16 = select(a = var_23_to_fp16, b = var_1314_cast_fp16, cond = mask_3)[name = tensor("input_277_cast_fp16")]; + tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; + tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; + tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_193")]; + tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = input_277_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_139_cast_fp16")]; + tensor var_1318_perm_0 = const()[name = tensor("op_1318_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1319 = const()[name = tensor("op_1319"), val = tensor([1, -1, 512])]; + tensor var_1318_cast_fp16 = transpose(perm = var_1318_perm_0, x = x_139_cast_fp16)[name = tensor("transpose_189")]; + tensor input_279_cast_fp16 = reshape(shape = var_1319, x = var_1318_cast_fp16)[name = tensor("input_279_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67249472)))]; + tensor linear_51_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_out_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("linear_51_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_273_cast_fp16, y = linear_51_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor x_143_axes_0 = const()[name = tensor("x_143_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67773824)))]; + tensor encoder_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67774912)))]; + tensor x_143_cast_fp16 = layer_norm(axes = x_143_axes_0, beta = encoder_layers_5_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_5_norm_conv_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("x_143_cast_fp16")]; + tensor input_285_perm_0 = const()[name = tensor("input_285_perm_0"), val = tensor([0, 2, 1])]; + tensor input_287_pad_type_0 = const()[name = tensor("input_287_pad_type_0"), val = tensor("valid")]; + tensor input_287_strides_0 = const()[name = tensor("input_287_strides_0"), val = tensor([1])]; + tensor input_287_pad_0 = const()[name = tensor("input_287_pad_0"), val = tensor([0, 0])]; + tensor input_287_dilations_0 = const()[name = tensor("input_287_dilations_0"), val = tensor([1])]; + tensor input_287_groups_0 = const()[name = tensor("input_287_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67776000)))]; + tensor input_285_cast_fp16 = transpose(perm = input_285_perm_0, x = x_143_cast_fp16)[name = tensor("transpose_188")]; + tensor input_287_cast_fp16 = conv(dilations = input_287_dilations_0, groups = input_287_groups_0, pad = input_287_pad_0, pad_type = input_287_pad_type_0, strides = input_287_strides_0, weight = encoder_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("input_287_cast_fp16")]; + tensor x_145_split_num_splits_0 = const()[name = tensor("x_145_split_num_splits_0"), val = tensor(2)]; + tensor x_145_split_axis_0 = const()[name = tensor("x_145_split_axis_0"), val = tensor(1)]; + tensor x_145_split_cast_fp16_0, tensor x_145_split_cast_fp16_1 = split(axis = x_145_split_axis_0, num_splits = x_145_split_num_splits_0, x = input_287_cast_fp16)[name = tensor("x_145_split_cast_fp16")]; + tensor x_145_split_1_sigmoid_cast_fp16 = sigmoid(x = x_145_split_cast_fp16_1)[name = tensor("x_145_split_1_sigmoid_cast_fp16")]; + tensor x_145_cast_fp16 = mul(x = x_145_split_cast_fp16_0, y = x_145_split_1_sigmoid_cast_fp16)[name = tensor("x_145_cast_fp16")]; + tensor input_289_cast_fp16 = select(a = var_23_to_fp16, b = x_145_cast_fp16, cond = var_309)[name = tensor("input_289_cast_fp16")]; + tensor new_x_23_interleave_0 = const()[name = tensor("new_x_23_interleave_0"), val = tensor(false)]; + tensor new_x_23_cast_fp16 = concat(axis = var_30, interleave = new_x_23_interleave_0, values = (cache_23_cast_fp16, input_289_cast_fp16))[name = tensor("new_x_23_cast_fp16")]; + tensor next_cache_11_begin_0 = const()[name = tensor("next_cache_11_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_11_end_0 = const()[name = tensor("next_cache_11_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_11_end_mask_0 = const()[name = tensor("next_cache_11_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_11_cast_fp16 = slice_by_index(begin = next_cache_11_begin_0, end = next_cache_11_end_0, end_mask = next_cache_11_end_mask_0, x = new_x_23_cast_fp16)[name = tensor("next_cache_11_cast_fp16")]; + tensor var_1360_begin_0 = const()[name = tensor("op_1360_begin_0"), val = tensor([0, 0, 14])]; + tensor var_1360_end_0 = const()[name = tensor("op_1360_end_0"), val = tensor([1, 512, 22])]; + tensor var_1360_end_mask_0 = const()[name = tensor("op_1360_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1360_cast_fp16 = slice_by_index(begin = var_1360_begin_0, end = var_1360_end_0, end_mask = var_1360_end_mask_0, x = next_cache_11_cast_fp16)[name = tensor("op_1360_cast_fp16")]; + tensor x_147_pad_type_0 = const()[name = tensor("x_147_pad_type_0"), val = tensor("valid")]; + tensor x_147_groups_0 = const()[name = tensor("x_147_groups_0"), val = tensor(512)]; + tensor x_147_strides_0 = const()[name = tensor("x_147_strides_0"), val = tensor([1])]; + tensor x_147_pad_0 = const()[name = tensor("x_147_pad_0"), val = tensor([0, 0])]; + tensor x_147_dilations_0 = const()[name = tensor("x_147_dilations_0"), val = tensor([1])]; + tensor encoder_layers_5_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68824640)))]; + tensor x_147_cast_fp16 = conv(dilations = x_147_dilations_0, groups = x_147_groups_0, pad = x_147_pad_0, pad_type = x_147_pad_type_0, strides = x_147_strides_0, weight = encoder_layers_5_conv_depthwise_conv_weight_to_fp16, x = new_x_23_cast_fp16)[name = tensor("x_147_cast_fp16")]; + tensor input_291_perm_0 = const()[name = tensor("input_291_perm_0"), val = tensor([0, 2, 1])]; + tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68833920)))]; + tensor encoder_layers_5_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68835008)))]; + tensor input_291_cast_fp16 = transpose(perm = input_291_perm_0, x = x_147_cast_fp16)[name = tensor("transpose_187")]; + tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = encoder_layers_5_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_5_conv_batch_norm_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("x_149_cast_fp16")]; + tensor input_293_perm_0 = const()[name = tensor("input_293_perm_0"), val = tensor([0, 2, 1])]; + tensor input_293_cast_fp16 = transpose(perm = input_293_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_186")]; + tensor input_295_cast_fp16 = silu(x = input_293_cast_fp16)[name = tensor("input_295_cast_fp16")]; + tensor x_151_pad_type_0 = const()[name = tensor("x_151_pad_type_0"), val = tensor("valid")]; + tensor x_151_strides_0 = const()[name = tensor("x_151_strides_0"), val = tensor([1])]; + tensor x_151_pad_0 = const()[name = tensor("x_151_pad_0"), val = tensor([0, 0])]; + tensor x_151_dilations_0 = const()[name = tensor("x_151_dilations_0"), val = tensor([1])]; + tensor x_151_groups_0 = const()[name = tensor("x_151_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68836096)))]; + tensor x_151_cast_fp16 = conv(dilations = x_151_dilations_0, groups = x_151_groups_0, pad = x_151_pad_0, pad_type = x_151_pad_type_0, strides = x_151_strides_0, weight = encoder_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_295_cast_fp16)[name = tensor("x_151_cast_fp16")]; + tensor input_297_perm_0 = const()[name = tensor("input_297_perm_0"), val = tensor([0, 2, 1])]; + tensor input_297_cast_fp16 = transpose(perm = input_297_perm_0, x = x_151_cast_fp16)[name = tensor("transpose_185")]; + tensor input_299_cast_fp16 = add(x = input_283_cast_fp16, y = input_297_cast_fp16)[name = tensor("input_299_cast_fp16")]; + tensor input_301_axes_0 = const()[name = tensor("input_301_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69360448)))]; + tensor encoder_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69361536)))]; + tensor input_301_cast_fp16 = layer_norm(axes = input_301_axes_0, beta = encoder_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_5_norm_feed_forward2_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("input_301_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69362624)))]; + tensor linear_52_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_305_cast_fp16 = silu(x = linear_52_cast_fp16)[name = tensor("input_305_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71459840)))]; + tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_305_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor var_1401_to_fp16 = const()[name = tensor("op_1401_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1402_cast_fp16 = mul(x = linear_53_cast_fp16, y = var_1401_to_fp16)[name = tensor("op_1402_cast_fp16")]; + tensor input_311_cast_fp16 = add(x = input_299_cast_fp16, y = var_1402_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor input_313_axes_0 = const()[name = tensor("input_313_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73557056)))]; + tensor encoder_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73558144)))]; + tensor input_313_cast_fp16 = layer_norm(axes = input_313_axes_0, beta = encoder_layers_5_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_5_norm_out_weight_to_fp16, x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; + tensor cache_25_begin_0 = const()[name = tensor("cache_25_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_25_end_0 = const()[name = tensor("cache_25_end_0"), val = tensor([7, 1, 70, 512])]; + tensor cache_25_end_mask_0 = const()[name = tensor("cache_25_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_25_squeeze_mask_0 = const()[name = tensor("cache_25_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_25_cast_fp16 = slice_by_index(begin = cache_25_begin_0, end = cache_25_end_0, end_mask = cache_25_end_mask_0, squeeze_mask = cache_25_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_25_cast_fp16")]; + tensor cache_27_begin_0 = const()[name = tensor("cache_27_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_27_end_0 = const()[name = tensor("cache_27_end_0"), val = tensor([7, 1, 512, 8])]; + tensor cache_27_end_mask_0 = const()[name = tensor("cache_27_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_27_squeeze_mask_0 = const()[name = tensor("cache_27_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_27_cast_fp16 = slice_by_index(begin = cache_27_begin_0, end = cache_27_end_0, end_mask = cache_27_end_mask_0, squeeze_mask = cache_27_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_27_cast_fp16")]; + tensor input_315_axes_0 = const()[name = tensor("input_315_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73559232)))]; + tensor encoder_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73560320)))]; + tensor input_315_cast_fp16 = layer_norm(axes = input_315_axes_0, beta = encoder_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_6_norm_feed_forward1_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("input_315_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73561408)))]; + tensor linear_54_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor input_319_cast_fp16 = silu(x = linear_54_cast_fp16)[name = tensor("input_319_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75658624)))]; + tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor var_1436_to_fp16 = const()[name = tensor("op_1436_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1437_cast_fp16 = mul(x = linear_55_cast_fp16, y = var_1436_to_fp16)[name = tensor("op_1437_cast_fp16")]; + tensor input_325_cast_fp16 = add(x = input_313_cast_fp16, y = var_1437_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor key_13_axes_0 = const()[name = tensor("key_13_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77755840)))]; + tensor encoder_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77756928)))]; + tensor key_13_cast_fp16 = layer_norm(axes = key_13_axes_0, beta = encoder_layers_6_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_6_norm_self_att_weight_to_fp16, x = input_325_cast_fp16)[name = tensor("key_13_cast_fp16")]; + tensor input_327_interleave_0 = const()[name = tensor("input_327_interleave_0"), val = tensor(false)]; + tensor input_327_cast_fp16 = concat(axis = var_44, interleave = input_327_interleave_0, values = (cache_25_cast_fp16, key_13_cast_fp16))[name = tensor("input_327_cast_fp16")]; + tensor var_1459_begin_0 = const()[name = tensor("op_1459_begin_0"), val = tensor([0, 14, 0])]; + tensor var_1459_end_0 = const()[name = tensor("op_1459_end_0"), val = tensor([1, 70, 512])]; + tensor var_1459_end_mask_0 = const()[name = tensor("op_1459_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1459_cast_fp16 = slice_by_index(begin = var_1459_begin_0, end = var_1459_end_0, end_mask = var_1459_end_mask_0, x = cache_25_cast_fp16)[name = tensor("op_1459_cast_fp16")]; + tensor var_1462_begin_0 = const()[name = tensor("op_1462_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1462_end_0 = const()[name = tensor("op_1462_end_0"), val = tensor([1, 14, 512])]; + tensor var_1462_end_mask_0 = const()[name = tensor("op_1462_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1462_cast_fp16 = slice_by_index(begin = var_1462_begin_0, end = var_1462_end_0, end_mask = var_1462_end_mask_0, x = key_13_cast_fp16)[name = tensor("op_1462_cast_fp16")]; + tensor var_1465_interleave_0 = const()[name = tensor("op_1465_interleave_0"), val = tensor(false)]; + tensor var_1465_cast_fp16 = concat(axis = var_44, interleave = var_1465_interleave_0, values = (var_1459_cast_fp16, var_1462_cast_fp16))[name = tensor("op_1465_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77758016)))]; + tensor linear_56_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_q_weight_to_fp16, x = key_13_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_1469 = const()[name = tensor("op_1469"), val = tensor([1, -1, 8, 64])]; + tensor q_37_cast_fp16 = reshape(shape = var_1469, x = linear_56_cast_fp16)[name = tensor("q_37_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78282368)))]; + tensor linear_57_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_k_weight_to_fp16, x = input_327_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor var_1473 = const()[name = tensor("op_1473"), val = tensor([1, -1, 8, 64])]; + tensor k_25_cast_fp16 = reshape(shape = var_1473, x = linear_57_cast_fp16)[name = tensor("k_25_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78806720)))]; + tensor linear_58_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_v_weight_to_fp16, x = input_327_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor var_1477 = const()[name = tensor("op_1477"), val = tensor([1, -1, 8, 64])]; + tensor v_13_cast_fp16 = reshape(shape = var_1477, x = linear_58_cast_fp16)[name = tensor("v_13_cast_fp16")]; + tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79331072)))]; + tensor var_1489_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1489_cast_fp16")]; + tensor encoder_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79332160)))]; + tensor var_1491_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1491_cast_fp16")]; + tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_159_transpose_x_0 = const()[name = tensor("x_159_transpose_x_0"), val = tensor(false)]; + tensor x_159_transpose_y_0 = const()[name = tensor("x_159_transpose_y_0"), val = tensor(false)]; + tensor var_1493_to_fp16 = const()[name = tensor("op_1493_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79333248)))]; + tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1491_cast_fp16)[name = tensor("transpose_183")]; + tensor x_159_cast_fp16 = matmul(transpose_x = x_159_transpose_x_0, transpose_y = x_159_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1493_to_fp16)[name = tensor("x_159_cast_fp16")]; + tensor x_161_pad_0 = const()[name = tensor("x_161_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_161_mode_0 = const()[name = tensor("x_161_mode_0"), val = tensor("constant")]; + tensor const_96_to_fp16 = const()[name = tensor("const_96_to_fp16"), val = tensor(0x0p+0)]; + tensor x_161_cast_fp16 = pad(constant_val = const_96_to_fp16, mode = x_161_mode_0, pad = x_161_pad_0, x = x_159_cast_fp16)[name = tensor("x_161_cast_fp16")]; + tensor var_1501 = const()[name = tensor("op_1501"), val = tensor([1, 8, -1, 16])]; + tensor x_163_cast_fp16 = reshape(shape = var_1501, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; + tensor var_1505_begin_0 = const()[name = tensor("op_1505_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1505_end_0 = const()[name = tensor("op_1505_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_1505_end_mask_0 = const()[name = tensor("op_1505_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1505_cast_fp16 = slice_by_index(begin = var_1505_begin_0, end = var_1505_end_0, end_mask = var_1505_end_mask_0, x = x_163_cast_fp16)[name = tensor("op_1505_cast_fp16")]; + tensor var_1506 = const()[name = tensor("op_1506"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1506, x = var_1505_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; + tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; + tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_181")]; + tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1489_cast_fp16)[name = tensor("transpose_182")]; + tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; + tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; + tensor var_1515_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1515_cast_fp16")]; + tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_25_cast_fp16 = mul(x = var_1515_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; + tensor scores_27_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_3)[name = tensor("scores_27_cast_fp16")]; + tensor var_1521_cast_fp16 = softmax(axis = var_30, x = scores_27_cast_fp16)[name = tensor("op_1521_cast_fp16")]; + tensor input_329_cast_fp16 = select(a = var_23_to_fp16, b = var_1521_cast_fp16, cond = mask_3)[name = tensor("input_329_cast_fp16")]; + tensor x_165_transpose_x_0 = const()[name = tensor("x_165_transpose_x_0"), val = tensor(false)]; + tensor x_165_transpose_y_0 = const()[name = tensor("x_165_transpose_y_0"), val = tensor(false)]; + tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_184")]; + tensor x_165_cast_fp16 = matmul(transpose_x = x_165_transpose_x_0, transpose_y = x_165_transpose_y_0, x = input_329_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_165_cast_fp16")]; + tensor var_1525_perm_0 = const()[name = tensor("op_1525_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1526 = const()[name = tensor("op_1526"), val = tensor([1, -1, 512])]; + tensor var_1525_cast_fp16 = transpose(perm = var_1525_perm_0, x = x_165_cast_fp16)[name = tensor("transpose_180")]; + tensor input_331_cast_fp16 = reshape(shape = var_1526, x = var_1525_cast_fp16)[name = tensor("input_331_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79508416)))]; + tensor linear_60_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_out_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("linear_60_cast_fp16")]; + tensor input_335_cast_fp16 = add(x = input_325_cast_fp16, y = linear_60_cast_fp16)[name = tensor("input_335_cast_fp16")]; + tensor x_169_axes_0 = const()[name = tensor("x_169_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80032768)))]; + tensor encoder_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80033856)))]; + tensor x_169_cast_fp16 = layer_norm(axes = x_169_axes_0, beta = encoder_layers_6_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_6_norm_conv_weight_to_fp16, x = input_335_cast_fp16)[name = tensor("x_169_cast_fp16")]; + tensor input_337_perm_0 = const()[name = tensor("input_337_perm_0"), val = tensor([0, 2, 1])]; + tensor input_339_pad_type_0 = const()[name = tensor("input_339_pad_type_0"), val = tensor("valid")]; + tensor input_339_strides_0 = const()[name = tensor("input_339_strides_0"), val = tensor([1])]; + tensor input_339_pad_0 = const()[name = tensor("input_339_pad_0"), val = tensor([0, 0])]; + tensor input_339_dilations_0 = const()[name = tensor("input_339_dilations_0"), val = tensor([1])]; + tensor input_339_groups_0 = const()[name = tensor("input_339_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80034944)))]; + tensor input_337_cast_fp16 = transpose(perm = input_337_perm_0, x = x_169_cast_fp16)[name = tensor("transpose_179")]; + tensor input_339_cast_fp16 = conv(dilations = input_339_dilations_0, groups = input_339_groups_0, pad = input_339_pad_0, pad_type = input_339_pad_type_0, strides = input_339_strides_0, weight = encoder_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor x_171_split_num_splits_0 = const()[name = tensor("x_171_split_num_splits_0"), val = tensor(2)]; + tensor x_171_split_axis_0 = const()[name = tensor("x_171_split_axis_0"), val = tensor(1)]; + tensor x_171_split_cast_fp16_0, tensor x_171_split_cast_fp16_1 = split(axis = x_171_split_axis_0, num_splits = x_171_split_num_splits_0, x = input_339_cast_fp16)[name = tensor("x_171_split_cast_fp16")]; + tensor x_171_split_1_sigmoid_cast_fp16 = sigmoid(x = x_171_split_cast_fp16_1)[name = tensor("x_171_split_1_sigmoid_cast_fp16")]; + tensor x_171_cast_fp16 = mul(x = x_171_split_cast_fp16_0, y = x_171_split_1_sigmoid_cast_fp16)[name = tensor("x_171_cast_fp16")]; + tensor input_341_cast_fp16 = select(a = var_23_to_fp16, b = x_171_cast_fp16, cond = var_309)[name = tensor("input_341_cast_fp16")]; + tensor new_x_27_interleave_0 = const()[name = tensor("new_x_27_interleave_0"), val = tensor(false)]; + tensor new_x_27_cast_fp16 = concat(axis = var_30, interleave = new_x_27_interleave_0, values = (cache_27_cast_fp16, input_341_cast_fp16))[name = tensor("new_x_27_cast_fp16")]; + tensor next_cache_13_begin_0 = const()[name = tensor("next_cache_13_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_13_end_0 = const()[name = tensor("next_cache_13_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_13_end_mask_0 = const()[name = tensor("next_cache_13_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_13_cast_fp16 = slice_by_index(begin = next_cache_13_begin_0, end = next_cache_13_end_0, end_mask = next_cache_13_end_mask_0, x = new_x_27_cast_fp16)[name = tensor("next_cache_13_cast_fp16")]; + tensor var_1567_begin_0 = const()[name = tensor("op_1567_begin_0"), val = tensor([0, 0, 14])]; + tensor var_1567_end_0 = const()[name = tensor("op_1567_end_0"), val = tensor([1, 512, 22])]; + tensor var_1567_end_mask_0 = const()[name = tensor("op_1567_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1567_cast_fp16 = slice_by_index(begin = var_1567_begin_0, end = var_1567_end_0, end_mask = var_1567_end_mask_0, x = next_cache_13_cast_fp16)[name = tensor("op_1567_cast_fp16")]; + tensor x_173_pad_type_0 = const()[name = tensor("x_173_pad_type_0"), val = tensor("valid")]; + tensor x_173_groups_0 = const()[name = tensor("x_173_groups_0"), val = tensor(512)]; + tensor x_173_strides_0 = const()[name = tensor("x_173_strides_0"), val = tensor([1])]; + tensor x_173_pad_0 = const()[name = tensor("x_173_pad_0"), val = tensor([0, 0])]; + tensor x_173_dilations_0 = const()[name = tensor("x_173_dilations_0"), val = tensor([1])]; + tensor encoder_layers_6_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81083584)))]; + tensor x_173_cast_fp16 = conv(dilations = x_173_dilations_0, groups = x_173_groups_0, pad = x_173_pad_0, pad_type = x_173_pad_type_0, strides = x_173_strides_0, weight = encoder_layers_6_conv_depthwise_conv_weight_to_fp16, x = new_x_27_cast_fp16)[name = tensor("x_173_cast_fp16")]; + tensor input_343_perm_0 = const()[name = tensor("input_343_perm_0"), val = tensor([0, 2, 1])]; + tensor x_175_axes_0 = const()[name = tensor("x_175_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81092864)))]; + tensor encoder_layers_6_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81093952)))]; + tensor input_343_cast_fp16 = transpose(perm = input_343_perm_0, x = x_173_cast_fp16)[name = tensor("transpose_178")]; + tensor x_175_cast_fp16 = layer_norm(axes = x_175_axes_0, beta = encoder_layers_6_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_6_conv_batch_norm_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("x_175_cast_fp16")]; + tensor input_345_perm_0 = const()[name = tensor("input_345_perm_0"), val = tensor([0, 2, 1])]; + tensor input_345_cast_fp16 = transpose(perm = input_345_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_177")]; + tensor input_347_cast_fp16 = silu(x = input_345_cast_fp16)[name = tensor("input_347_cast_fp16")]; + tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; + tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; + tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; + tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; + tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81095040)))]; + tensor x_177_cast_fp16 = conv(dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = encoder_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("x_177_cast_fp16")]; + tensor input_349_perm_0 = const()[name = tensor("input_349_perm_0"), val = tensor([0, 2, 1])]; + tensor input_349_cast_fp16 = transpose(perm = input_349_perm_0, x = x_177_cast_fp16)[name = tensor("transpose_176")]; + tensor input_351_cast_fp16 = add(x = input_335_cast_fp16, y = input_349_cast_fp16)[name = tensor("input_351_cast_fp16")]; + tensor input_353_axes_0 = const()[name = tensor("input_353_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81619392)))]; + tensor encoder_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81620480)))]; + tensor input_353_cast_fp16 = layer_norm(axes = input_353_axes_0, beta = encoder_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_6_norm_feed_forward2_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("input_353_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81621568)))]; + tensor linear_61_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor input_357_cast_fp16 = silu(x = linear_61_cast_fp16)[name = tensor("input_357_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83718784)))]; + tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_357_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor var_1608_to_fp16 = const()[name = tensor("op_1608_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1609_cast_fp16 = mul(x = linear_62_cast_fp16, y = var_1608_to_fp16)[name = tensor("op_1609_cast_fp16")]; + tensor input_363_cast_fp16 = add(x = input_351_cast_fp16, y = var_1609_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor input_365_axes_0 = const()[name = tensor("input_365_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85816000)))]; + tensor encoder_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85817088)))]; + tensor input_365_cast_fp16 = layer_norm(axes = input_365_axes_0, beta = encoder_layers_6_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_6_norm_out_weight_to_fp16, x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; + tensor cache_29_begin_0 = const()[name = tensor("cache_29_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_29_end_0 = const()[name = tensor("cache_29_end_0"), val = tensor([8, 1, 70, 512])]; + tensor cache_29_end_mask_0 = const()[name = tensor("cache_29_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_29_squeeze_mask_0 = const()[name = tensor("cache_29_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_29_cast_fp16 = slice_by_index(begin = cache_29_begin_0, end = cache_29_end_0, end_mask = cache_29_end_mask_0, squeeze_mask = cache_29_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_29_cast_fp16")]; + tensor cache_31_begin_0 = const()[name = tensor("cache_31_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_31_end_0 = const()[name = tensor("cache_31_end_0"), val = tensor([8, 1, 512, 8])]; + tensor cache_31_end_mask_0 = const()[name = tensor("cache_31_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_31_squeeze_mask_0 = const()[name = tensor("cache_31_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_31_cast_fp16 = slice_by_index(begin = cache_31_begin_0, end = cache_31_end_0, end_mask = cache_31_end_mask_0, squeeze_mask = cache_31_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_31_cast_fp16")]; + tensor input_367_axes_0 = const()[name = tensor("input_367_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85818176)))]; + tensor encoder_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85819264)))]; + tensor input_367_cast_fp16 = layer_norm(axes = input_367_axes_0, beta = encoder_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_7_norm_feed_forward1_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("input_367_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85820352)))]; + tensor linear_63_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_367_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor input_371_cast_fp16 = silu(x = linear_63_cast_fp16)[name = tensor("input_371_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87917568)))]; + tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor var_1643_to_fp16 = const()[name = tensor("op_1643_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1644_cast_fp16 = mul(x = linear_64_cast_fp16, y = var_1643_to_fp16)[name = tensor("op_1644_cast_fp16")]; + tensor input_377_cast_fp16 = add(x = input_365_cast_fp16, y = var_1644_cast_fp16)[name = tensor("input_377_cast_fp16")]; + tensor key_15_axes_0 = const()[name = tensor("key_15_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90014784)))]; + tensor encoder_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90015872)))]; + tensor key_15_cast_fp16 = layer_norm(axes = key_15_axes_0, beta = encoder_layers_7_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_7_norm_self_att_weight_to_fp16, x = input_377_cast_fp16)[name = tensor("key_15_cast_fp16")]; + tensor input_379_interleave_0 = const()[name = tensor("input_379_interleave_0"), val = tensor(false)]; + tensor input_379_cast_fp16 = concat(axis = var_44, interleave = input_379_interleave_0, values = (cache_29_cast_fp16, key_15_cast_fp16))[name = tensor("input_379_cast_fp16")]; + tensor var_1666_begin_0 = const()[name = tensor("op_1666_begin_0"), val = tensor([0, 14, 0])]; + tensor var_1666_end_0 = const()[name = tensor("op_1666_end_0"), val = tensor([1, 70, 512])]; + tensor var_1666_end_mask_0 = const()[name = tensor("op_1666_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1666_cast_fp16 = slice_by_index(begin = var_1666_begin_0, end = var_1666_end_0, end_mask = var_1666_end_mask_0, x = cache_29_cast_fp16)[name = tensor("op_1666_cast_fp16")]; + tensor var_1669_begin_0 = const()[name = tensor("op_1669_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1669_end_0 = const()[name = tensor("op_1669_end_0"), val = tensor([1, 14, 512])]; + tensor var_1669_end_mask_0 = const()[name = tensor("op_1669_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1669_cast_fp16 = slice_by_index(begin = var_1669_begin_0, end = var_1669_end_0, end_mask = var_1669_end_mask_0, x = key_15_cast_fp16)[name = tensor("op_1669_cast_fp16")]; + tensor var_1672_interleave_0 = const()[name = tensor("op_1672_interleave_0"), val = tensor(false)]; + tensor var_1672_cast_fp16 = concat(axis = var_44, interleave = var_1672_interleave_0, values = (var_1666_cast_fp16, var_1669_cast_fp16))[name = tensor("op_1672_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90016960)))]; + tensor linear_65_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_q_weight_to_fp16, x = key_15_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor var_1676 = const()[name = tensor("op_1676"), val = tensor([1, -1, 8, 64])]; + tensor q_43_cast_fp16 = reshape(shape = var_1676, x = linear_65_cast_fp16)[name = tensor("q_43_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90541312)))]; + tensor linear_66_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_k_weight_to_fp16, x = input_379_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1680 = const()[name = tensor("op_1680"), val = tensor([1, -1, 8, 64])]; + tensor k_29_cast_fp16 = reshape(shape = var_1680, x = linear_66_cast_fp16)[name = tensor("k_29_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91065664)))]; + tensor linear_67_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_v_weight_to_fp16, x = input_379_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1684 = const()[name = tensor("op_1684"), val = tensor([1, -1, 8, 64])]; + tensor v_15_cast_fp16 = reshape(shape = var_1684, x = linear_67_cast_fp16)[name = tensor("v_15_cast_fp16")]; + tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91590016)))]; + tensor var_1696_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1696_cast_fp16")]; + tensor encoder_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91591104)))]; + tensor var_1698_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1698_cast_fp16")]; + tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_185_transpose_x_0 = const()[name = tensor("x_185_transpose_x_0"), val = tensor(false)]; + tensor x_185_transpose_y_0 = const()[name = tensor("x_185_transpose_y_0"), val = tensor(false)]; + tensor var_1700_to_fp16 = const()[name = tensor("op_1700_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91592192)))]; + tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1698_cast_fp16)[name = tensor("transpose_174")]; + tensor x_185_cast_fp16 = matmul(transpose_x = x_185_transpose_x_0, transpose_y = x_185_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1700_to_fp16)[name = tensor("x_185_cast_fp16")]; + tensor x_187_pad_0 = const()[name = tensor("x_187_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_187_mode_0 = const()[name = tensor("x_187_mode_0"), val = tensor("constant")]; + tensor const_109_to_fp16 = const()[name = tensor("const_109_to_fp16"), val = tensor(0x0p+0)]; + tensor x_187_cast_fp16 = pad(constant_val = const_109_to_fp16, mode = x_187_mode_0, pad = x_187_pad_0, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; + tensor var_1708 = const()[name = tensor("op_1708"), val = tensor([1, 8, -1, 16])]; + tensor x_189_cast_fp16 = reshape(shape = var_1708, x = x_187_cast_fp16)[name = tensor("x_189_cast_fp16")]; + tensor var_1712_begin_0 = const()[name = tensor("op_1712_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1712_end_0 = const()[name = tensor("op_1712_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_1712_end_mask_0 = const()[name = tensor("op_1712_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1712_cast_fp16 = slice_by_index(begin = var_1712_begin_0, end = var_1712_end_0, end_mask = var_1712_end_mask_0, x = x_189_cast_fp16)[name = tensor("op_1712_cast_fp16")]; + tensor var_1713 = const()[name = tensor("op_1713"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1713, x = var_1712_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; + tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; + tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_172")]; + tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1696_cast_fp16)[name = tensor("transpose_173")]; + tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; + tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; + tensor var_1722_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1722_cast_fp16")]; + tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_29_cast_fp16 = mul(x = var_1722_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; + tensor scores_31_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_3)[name = tensor("scores_31_cast_fp16")]; + tensor var_1728_cast_fp16 = softmax(axis = var_30, x = scores_31_cast_fp16)[name = tensor("op_1728_cast_fp16")]; + tensor input_381_cast_fp16 = select(a = var_23_to_fp16, b = var_1728_cast_fp16, cond = mask_3)[name = tensor("input_381_cast_fp16")]; + tensor x_191_transpose_x_0 = const()[name = tensor("x_191_transpose_x_0"), val = tensor(false)]; + tensor x_191_transpose_y_0 = const()[name = tensor("x_191_transpose_y_0"), val = tensor(false)]; + tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_175")]; + tensor x_191_cast_fp16 = matmul(transpose_x = x_191_transpose_x_0, transpose_y = x_191_transpose_y_0, x = input_381_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_191_cast_fp16")]; + tensor var_1732_perm_0 = const()[name = tensor("op_1732_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1733 = const()[name = tensor("op_1733"), val = tensor([1, -1, 512])]; + tensor var_1732_cast_fp16 = transpose(perm = var_1732_perm_0, x = x_191_cast_fp16)[name = tensor("transpose_171")]; + tensor input_383_cast_fp16 = reshape(shape = var_1733, x = var_1732_cast_fp16)[name = tensor("input_383_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91767360)))]; + tensor linear_69_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_out_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("linear_69_cast_fp16")]; + tensor input_387_cast_fp16 = add(x = input_377_cast_fp16, y = linear_69_cast_fp16)[name = tensor("input_387_cast_fp16")]; + tensor x_195_axes_0 = const()[name = tensor("x_195_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92291712)))]; + tensor encoder_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92292800)))]; + tensor x_195_cast_fp16 = layer_norm(axes = x_195_axes_0, beta = encoder_layers_7_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_7_norm_conv_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("x_195_cast_fp16")]; + tensor input_389_perm_0 = const()[name = tensor("input_389_perm_0"), val = tensor([0, 2, 1])]; + tensor input_391_pad_type_0 = const()[name = tensor("input_391_pad_type_0"), val = tensor("valid")]; + tensor input_391_strides_0 = const()[name = tensor("input_391_strides_0"), val = tensor([1])]; + tensor input_391_pad_0 = const()[name = tensor("input_391_pad_0"), val = tensor([0, 0])]; + tensor input_391_dilations_0 = const()[name = tensor("input_391_dilations_0"), val = tensor([1])]; + tensor input_391_groups_0 = const()[name = tensor("input_391_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92293888)))]; + tensor input_389_cast_fp16 = transpose(perm = input_389_perm_0, x = x_195_cast_fp16)[name = tensor("transpose_170")]; + tensor input_391_cast_fp16 = conv(dilations = input_391_dilations_0, groups = input_391_groups_0, pad = input_391_pad_0, pad_type = input_391_pad_type_0, strides = input_391_strides_0, weight = encoder_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; + tensor x_197_split_num_splits_0 = const()[name = tensor("x_197_split_num_splits_0"), val = tensor(2)]; + tensor x_197_split_axis_0 = const()[name = tensor("x_197_split_axis_0"), val = tensor(1)]; + tensor x_197_split_cast_fp16_0, tensor x_197_split_cast_fp16_1 = split(axis = x_197_split_axis_0, num_splits = x_197_split_num_splits_0, x = input_391_cast_fp16)[name = tensor("x_197_split_cast_fp16")]; + tensor x_197_split_1_sigmoid_cast_fp16 = sigmoid(x = x_197_split_cast_fp16_1)[name = tensor("x_197_split_1_sigmoid_cast_fp16")]; + tensor x_197_cast_fp16 = mul(x = x_197_split_cast_fp16_0, y = x_197_split_1_sigmoid_cast_fp16)[name = tensor("x_197_cast_fp16")]; + tensor input_393_cast_fp16 = select(a = var_23_to_fp16, b = x_197_cast_fp16, cond = var_309)[name = tensor("input_393_cast_fp16")]; + tensor new_x_31_interleave_0 = const()[name = tensor("new_x_31_interleave_0"), val = tensor(false)]; + tensor new_x_31_cast_fp16 = concat(axis = var_30, interleave = new_x_31_interleave_0, values = (cache_31_cast_fp16, input_393_cast_fp16))[name = tensor("new_x_31_cast_fp16")]; + tensor next_cache_15_begin_0 = const()[name = tensor("next_cache_15_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_15_end_0 = const()[name = tensor("next_cache_15_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_15_end_mask_0 = const()[name = tensor("next_cache_15_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_15_cast_fp16 = slice_by_index(begin = next_cache_15_begin_0, end = next_cache_15_end_0, end_mask = next_cache_15_end_mask_0, x = new_x_31_cast_fp16)[name = tensor("next_cache_15_cast_fp16")]; + tensor var_1774_begin_0 = const()[name = tensor("op_1774_begin_0"), val = tensor([0, 0, 14])]; + tensor var_1774_end_0 = const()[name = tensor("op_1774_end_0"), val = tensor([1, 512, 22])]; + tensor var_1774_end_mask_0 = const()[name = tensor("op_1774_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1774_cast_fp16 = slice_by_index(begin = var_1774_begin_0, end = var_1774_end_0, end_mask = var_1774_end_mask_0, x = next_cache_15_cast_fp16)[name = tensor("op_1774_cast_fp16")]; + tensor x_199_pad_type_0 = const()[name = tensor("x_199_pad_type_0"), val = tensor("valid")]; + tensor x_199_groups_0 = const()[name = tensor("x_199_groups_0"), val = tensor(512)]; + tensor x_199_strides_0 = const()[name = tensor("x_199_strides_0"), val = tensor([1])]; + tensor x_199_pad_0 = const()[name = tensor("x_199_pad_0"), val = tensor([0, 0])]; + tensor x_199_dilations_0 = const()[name = tensor("x_199_dilations_0"), val = tensor([1])]; + tensor encoder_layers_7_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93342528)))]; + tensor x_199_cast_fp16 = conv(dilations = x_199_dilations_0, groups = x_199_groups_0, pad = x_199_pad_0, pad_type = x_199_pad_type_0, strides = x_199_strides_0, weight = encoder_layers_7_conv_depthwise_conv_weight_to_fp16, x = new_x_31_cast_fp16)[name = tensor("x_199_cast_fp16")]; + tensor input_395_perm_0 = const()[name = tensor("input_395_perm_0"), val = tensor([0, 2, 1])]; + tensor x_201_axes_0 = const()[name = tensor("x_201_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93351808)))]; + tensor encoder_layers_7_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93352896)))]; + tensor input_395_cast_fp16 = transpose(perm = input_395_perm_0, x = x_199_cast_fp16)[name = tensor("transpose_169")]; + tensor x_201_cast_fp16 = layer_norm(axes = x_201_axes_0, beta = encoder_layers_7_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_7_conv_batch_norm_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("x_201_cast_fp16")]; + tensor input_397_perm_0 = const()[name = tensor("input_397_perm_0"), val = tensor([0, 2, 1])]; + tensor input_397_cast_fp16 = transpose(perm = input_397_perm_0, x = x_201_cast_fp16)[name = tensor("transpose_168")]; + tensor input_399_cast_fp16 = silu(x = input_397_cast_fp16)[name = tensor("input_399_cast_fp16")]; + tensor x_203_pad_type_0 = const()[name = tensor("x_203_pad_type_0"), val = tensor("valid")]; + tensor x_203_strides_0 = const()[name = tensor("x_203_strides_0"), val = tensor([1])]; + tensor x_203_pad_0 = const()[name = tensor("x_203_pad_0"), val = tensor([0, 0])]; + tensor x_203_dilations_0 = const()[name = tensor("x_203_dilations_0"), val = tensor([1])]; + tensor x_203_groups_0 = const()[name = tensor("x_203_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93353984)))]; + tensor x_203_cast_fp16 = conv(dilations = x_203_dilations_0, groups = x_203_groups_0, pad = x_203_pad_0, pad_type = x_203_pad_type_0, strides = x_203_strides_0, weight = encoder_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_399_cast_fp16)[name = tensor("x_203_cast_fp16")]; + tensor input_401_perm_0 = const()[name = tensor("input_401_perm_0"), val = tensor([0, 2, 1])]; + tensor input_401_cast_fp16 = transpose(perm = input_401_perm_0, x = x_203_cast_fp16)[name = tensor("transpose_167")]; + tensor input_403_cast_fp16 = add(x = input_387_cast_fp16, y = input_401_cast_fp16)[name = tensor("input_403_cast_fp16")]; + tensor input_405_axes_0 = const()[name = tensor("input_405_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93878336)))]; + tensor encoder_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93879424)))]; + tensor input_405_cast_fp16 = layer_norm(axes = input_405_axes_0, beta = encoder_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_7_norm_feed_forward2_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("input_405_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93880512)))]; + tensor linear_70_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_405_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_409_cast_fp16 = silu(x = linear_70_cast_fp16)[name = tensor("input_409_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95977728)))]; + tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_409_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor var_1815_to_fp16 = const()[name = tensor("op_1815_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1816_cast_fp16 = mul(x = linear_71_cast_fp16, y = var_1815_to_fp16)[name = tensor("op_1816_cast_fp16")]; + tensor input_415_cast_fp16 = add(x = input_403_cast_fp16, y = var_1816_cast_fp16)[name = tensor("input_415_cast_fp16")]; + tensor input_417_axes_0 = const()[name = tensor("input_417_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98074944)))]; + tensor encoder_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98076032)))]; + tensor input_417_cast_fp16 = layer_norm(axes = input_417_axes_0, beta = encoder_layers_7_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_7_norm_out_weight_to_fp16, x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; + tensor cache_33_begin_0 = const()[name = tensor("cache_33_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_33_end_0 = const()[name = tensor("cache_33_end_0"), val = tensor([9, 1, 70, 512])]; + tensor cache_33_end_mask_0 = const()[name = tensor("cache_33_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_33_squeeze_mask_0 = const()[name = tensor("cache_33_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_33_cast_fp16 = slice_by_index(begin = cache_33_begin_0, end = cache_33_end_0, end_mask = cache_33_end_mask_0, squeeze_mask = cache_33_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_33_cast_fp16")]; + tensor cache_35_begin_0 = const()[name = tensor("cache_35_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_35_end_0 = const()[name = tensor("cache_35_end_0"), val = tensor([9, 1, 512, 8])]; + tensor cache_35_end_mask_0 = const()[name = tensor("cache_35_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_35_squeeze_mask_0 = const()[name = tensor("cache_35_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_35_cast_fp16 = slice_by_index(begin = cache_35_begin_0, end = cache_35_end_0, end_mask = cache_35_end_mask_0, squeeze_mask = cache_35_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_35_cast_fp16")]; + tensor input_419_axes_0 = const()[name = tensor("input_419_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98077120)))]; + tensor encoder_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98078208)))]; + tensor input_419_cast_fp16 = layer_norm(axes = input_419_axes_0, beta = encoder_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_8_norm_feed_forward1_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("input_419_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98079296)))]; + tensor linear_72_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor input_423_cast_fp16 = silu(x = linear_72_cast_fp16)[name = tensor("input_423_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100176512)))]; + tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor var_1850_to_fp16 = const()[name = tensor("op_1850_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1851_cast_fp16 = mul(x = linear_73_cast_fp16, y = var_1850_to_fp16)[name = tensor("op_1851_cast_fp16")]; + tensor input_429_cast_fp16 = add(x = input_417_cast_fp16, y = var_1851_cast_fp16)[name = tensor("input_429_cast_fp16")]; + tensor key_17_axes_0 = const()[name = tensor("key_17_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102273728)))]; + tensor encoder_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102274816)))]; + tensor key_17_cast_fp16 = layer_norm(axes = key_17_axes_0, beta = encoder_layers_8_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_8_norm_self_att_weight_to_fp16, x = input_429_cast_fp16)[name = tensor("key_17_cast_fp16")]; + tensor input_431_interleave_0 = const()[name = tensor("input_431_interleave_0"), val = tensor(false)]; + tensor input_431_cast_fp16 = concat(axis = var_44, interleave = input_431_interleave_0, values = (cache_33_cast_fp16, key_17_cast_fp16))[name = tensor("input_431_cast_fp16")]; + tensor var_1873_begin_0 = const()[name = tensor("op_1873_begin_0"), val = tensor([0, 14, 0])]; + tensor var_1873_end_0 = const()[name = tensor("op_1873_end_0"), val = tensor([1, 70, 512])]; + tensor var_1873_end_mask_0 = const()[name = tensor("op_1873_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1873_cast_fp16 = slice_by_index(begin = var_1873_begin_0, end = var_1873_end_0, end_mask = var_1873_end_mask_0, x = cache_33_cast_fp16)[name = tensor("op_1873_cast_fp16")]; + tensor var_1876_begin_0 = const()[name = tensor("op_1876_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1876_end_0 = const()[name = tensor("op_1876_end_0"), val = tensor([1, 14, 512])]; + tensor var_1876_end_mask_0 = const()[name = tensor("op_1876_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1876_cast_fp16 = slice_by_index(begin = var_1876_begin_0, end = var_1876_end_0, end_mask = var_1876_end_mask_0, x = key_17_cast_fp16)[name = tensor("op_1876_cast_fp16")]; + tensor var_1879_interleave_0 = const()[name = tensor("op_1879_interleave_0"), val = tensor(false)]; + tensor var_1879_cast_fp16 = concat(axis = var_44, interleave = var_1879_interleave_0, values = (var_1873_cast_fp16, var_1876_cast_fp16))[name = tensor("op_1879_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102275904)))]; + tensor linear_74_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_q_weight_to_fp16, x = key_17_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_1883 = const()[name = tensor("op_1883"), val = tensor([1, -1, 8, 64])]; + tensor q_49_cast_fp16 = reshape(shape = var_1883, x = linear_74_cast_fp16)[name = tensor("q_49_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102800256)))]; + tensor linear_75_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_k_weight_to_fp16, x = input_431_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor var_1887 = const()[name = tensor("op_1887"), val = tensor([1, -1, 8, 64])]; + tensor k_33_cast_fp16 = reshape(shape = var_1887, x = linear_75_cast_fp16)[name = tensor("k_33_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103324608)))]; + tensor linear_76_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_v_weight_to_fp16, x = input_431_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor var_1891 = const()[name = tensor("op_1891"), val = tensor([1, -1, 8, 64])]; + tensor v_17_cast_fp16 = reshape(shape = var_1891, x = linear_76_cast_fp16)[name = tensor("v_17_cast_fp16")]; + tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103848960)))]; + tensor var_1903_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1903_cast_fp16")]; + tensor encoder_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103850048)))]; + tensor var_1905_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1905_cast_fp16")]; + tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; + tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; + tensor var_1907_to_fp16 = const()[name = tensor("op_1907_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103851136)))]; + tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1905_cast_fp16)[name = tensor("transpose_165")]; + tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1907_to_fp16)[name = tensor("x_211_cast_fp16")]; + tensor x_213_pad_0 = const()[name = tensor("x_213_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_213_mode_0 = const()[name = tensor("x_213_mode_0"), val = tensor("constant")]; + tensor const_122_to_fp16 = const()[name = tensor("const_122_to_fp16"), val = tensor(0x0p+0)]; + tensor x_213_cast_fp16 = pad(constant_val = const_122_to_fp16, mode = x_213_mode_0, pad = x_213_pad_0, x = x_211_cast_fp16)[name = tensor("x_213_cast_fp16")]; + tensor var_1915 = const()[name = tensor("op_1915"), val = tensor([1, 8, -1, 16])]; + tensor x_215_cast_fp16 = reshape(shape = var_1915, x = x_213_cast_fp16)[name = tensor("x_215_cast_fp16")]; + tensor var_1919_begin_0 = const()[name = tensor("op_1919_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1919_end_0 = const()[name = tensor("op_1919_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_1919_end_mask_0 = const()[name = tensor("op_1919_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1919_cast_fp16 = slice_by_index(begin = var_1919_begin_0, end = var_1919_end_0, end_mask = var_1919_end_mask_0, x = x_215_cast_fp16)[name = tensor("op_1919_cast_fp16")]; + tensor var_1920 = const()[name = tensor("op_1920"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1920, x = var_1919_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; + tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; + tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_163")]; + tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1903_cast_fp16)[name = tensor("transpose_164")]; + tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; + tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; + tensor var_1929_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1929_cast_fp16")]; + tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_33_cast_fp16 = mul(x = var_1929_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; + tensor scores_35_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_3)[name = tensor("scores_35_cast_fp16")]; + tensor var_1935_cast_fp16 = softmax(axis = var_30, x = scores_35_cast_fp16)[name = tensor("op_1935_cast_fp16")]; + tensor input_433_cast_fp16 = select(a = var_23_to_fp16, b = var_1935_cast_fp16, cond = mask_3)[name = tensor("input_433_cast_fp16")]; + tensor x_217_transpose_x_0 = const()[name = tensor("x_217_transpose_x_0"), val = tensor(false)]; + tensor x_217_transpose_y_0 = const()[name = tensor("x_217_transpose_y_0"), val = tensor(false)]; + tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_166")]; + tensor x_217_cast_fp16 = matmul(transpose_x = x_217_transpose_x_0, transpose_y = x_217_transpose_y_0, x = input_433_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_217_cast_fp16")]; + tensor var_1939_perm_0 = const()[name = tensor("op_1939_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1940 = const()[name = tensor("op_1940"), val = tensor([1, -1, 512])]; + tensor var_1939_cast_fp16 = transpose(perm = var_1939_perm_0, x = x_217_cast_fp16)[name = tensor("transpose_162")]; + tensor input_435_cast_fp16 = reshape(shape = var_1940, x = var_1939_cast_fp16)[name = tensor("input_435_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104026304)))]; + tensor linear_78_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_out_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("linear_78_cast_fp16")]; + tensor input_439_cast_fp16 = add(x = input_429_cast_fp16, y = linear_78_cast_fp16)[name = tensor("input_439_cast_fp16")]; + tensor x_221_axes_0 = const()[name = tensor("x_221_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104550656)))]; + tensor encoder_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104551744)))]; + tensor x_221_cast_fp16 = layer_norm(axes = x_221_axes_0, beta = encoder_layers_8_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_8_norm_conv_weight_to_fp16, x = input_439_cast_fp16)[name = tensor("x_221_cast_fp16")]; + tensor input_441_perm_0 = const()[name = tensor("input_441_perm_0"), val = tensor([0, 2, 1])]; + tensor input_443_pad_type_0 = const()[name = tensor("input_443_pad_type_0"), val = tensor("valid")]; + tensor input_443_strides_0 = const()[name = tensor("input_443_strides_0"), val = tensor([1])]; + tensor input_443_pad_0 = const()[name = tensor("input_443_pad_0"), val = tensor([0, 0])]; + tensor input_443_dilations_0 = const()[name = tensor("input_443_dilations_0"), val = tensor([1])]; + tensor input_443_groups_0 = const()[name = tensor("input_443_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104552832)))]; + tensor input_441_cast_fp16 = transpose(perm = input_441_perm_0, x = x_221_cast_fp16)[name = tensor("transpose_161")]; + tensor input_443_cast_fp16 = conv(dilations = input_443_dilations_0, groups = input_443_groups_0, pad = input_443_pad_0, pad_type = input_443_pad_type_0, strides = input_443_strides_0, weight = encoder_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("input_443_cast_fp16")]; + tensor x_223_split_num_splits_0 = const()[name = tensor("x_223_split_num_splits_0"), val = tensor(2)]; + tensor x_223_split_axis_0 = const()[name = tensor("x_223_split_axis_0"), val = tensor(1)]; + tensor x_223_split_cast_fp16_0, tensor x_223_split_cast_fp16_1 = split(axis = x_223_split_axis_0, num_splits = x_223_split_num_splits_0, x = input_443_cast_fp16)[name = tensor("x_223_split_cast_fp16")]; + tensor x_223_split_1_sigmoid_cast_fp16 = sigmoid(x = x_223_split_cast_fp16_1)[name = tensor("x_223_split_1_sigmoid_cast_fp16")]; + tensor x_223_cast_fp16 = mul(x = x_223_split_cast_fp16_0, y = x_223_split_1_sigmoid_cast_fp16)[name = tensor("x_223_cast_fp16")]; + tensor input_445_cast_fp16 = select(a = var_23_to_fp16, b = x_223_cast_fp16, cond = var_309)[name = tensor("input_445_cast_fp16")]; + tensor new_x_35_interleave_0 = const()[name = tensor("new_x_35_interleave_0"), val = tensor(false)]; + tensor new_x_35_cast_fp16 = concat(axis = var_30, interleave = new_x_35_interleave_0, values = (cache_35_cast_fp16, input_445_cast_fp16))[name = tensor("new_x_35_cast_fp16")]; + tensor next_cache_17_begin_0 = const()[name = tensor("next_cache_17_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_17_end_0 = const()[name = tensor("next_cache_17_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_17_end_mask_0 = const()[name = tensor("next_cache_17_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_17_cast_fp16 = slice_by_index(begin = next_cache_17_begin_0, end = next_cache_17_end_0, end_mask = next_cache_17_end_mask_0, x = new_x_35_cast_fp16)[name = tensor("next_cache_17_cast_fp16")]; + tensor var_1981_begin_0 = const()[name = tensor("op_1981_begin_0"), val = tensor([0, 0, 14])]; + tensor var_1981_end_0 = const()[name = tensor("op_1981_end_0"), val = tensor([1, 512, 22])]; + tensor var_1981_end_mask_0 = const()[name = tensor("op_1981_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1981_cast_fp16 = slice_by_index(begin = var_1981_begin_0, end = var_1981_end_0, end_mask = var_1981_end_mask_0, x = next_cache_17_cast_fp16)[name = tensor("op_1981_cast_fp16")]; + tensor x_225_pad_type_0 = const()[name = tensor("x_225_pad_type_0"), val = tensor("valid")]; + tensor x_225_groups_0 = const()[name = tensor("x_225_groups_0"), val = tensor(512)]; + tensor x_225_strides_0 = const()[name = tensor("x_225_strides_0"), val = tensor([1])]; + tensor x_225_pad_0 = const()[name = tensor("x_225_pad_0"), val = tensor([0, 0])]; + tensor x_225_dilations_0 = const()[name = tensor("x_225_dilations_0"), val = tensor([1])]; + tensor encoder_layers_8_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105601472)))]; + tensor x_225_cast_fp16 = conv(dilations = x_225_dilations_0, groups = x_225_groups_0, pad = x_225_pad_0, pad_type = x_225_pad_type_0, strides = x_225_strides_0, weight = encoder_layers_8_conv_depthwise_conv_weight_to_fp16, x = new_x_35_cast_fp16)[name = tensor("x_225_cast_fp16")]; + tensor input_447_perm_0 = const()[name = tensor("input_447_perm_0"), val = tensor([0, 2, 1])]; + tensor x_227_axes_0 = const()[name = tensor("x_227_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105610752)))]; + tensor encoder_layers_8_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105611840)))]; + tensor input_447_cast_fp16 = transpose(perm = input_447_perm_0, x = x_225_cast_fp16)[name = tensor("transpose_160")]; + tensor x_227_cast_fp16 = layer_norm(axes = x_227_axes_0, beta = encoder_layers_8_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_8_conv_batch_norm_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("x_227_cast_fp16")]; + tensor input_449_perm_0 = const()[name = tensor("input_449_perm_0"), val = tensor([0, 2, 1])]; + tensor input_449_cast_fp16 = transpose(perm = input_449_perm_0, x = x_227_cast_fp16)[name = tensor("transpose_159")]; + tensor input_451_cast_fp16 = silu(x = input_449_cast_fp16)[name = tensor("input_451_cast_fp16")]; + tensor x_229_pad_type_0 = const()[name = tensor("x_229_pad_type_0"), val = tensor("valid")]; + tensor x_229_strides_0 = const()[name = tensor("x_229_strides_0"), val = tensor([1])]; + tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0])]; + tensor x_229_dilations_0 = const()[name = tensor("x_229_dilations_0"), val = tensor([1])]; + tensor x_229_groups_0 = const()[name = tensor("x_229_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105612928)))]; + tensor x_229_cast_fp16 = conv(dilations = x_229_dilations_0, groups = x_229_groups_0, pad = x_229_pad_0, pad_type = x_229_pad_type_0, strides = x_229_strides_0, weight = encoder_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("x_229_cast_fp16")]; + tensor input_453_perm_0 = const()[name = tensor("input_453_perm_0"), val = tensor([0, 2, 1])]; + tensor input_453_cast_fp16 = transpose(perm = input_453_perm_0, x = x_229_cast_fp16)[name = tensor("transpose_158")]; + tensor input_455_cast_fp16 = add(x = input_439_cast_fp16, y = input_453_cast_fp16)[name = tensor("input_455_cast_fp16")]; + tensor input_457_axes_0 = const()[name = tensor("input_457_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106137280)))]; + tensor encoder_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106138368)))]; + tensor input_457_cast_fp16 = layer_norm(axes = input_457_axes_0, beta = encoder_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_8_norm_feed_forward2_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("input_457_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106139456)))]; + tensor linear_79_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor input_461_cast_fp16 = silu(x = linear_79_cast_fp16)[name = tensor("input_461_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108236672)))]; + tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_461_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor var_2022_to_fp16 = const()[name = tensor("op_2022_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2023_cast_fp16 = mul(x = linear_80_cast_fp16, y = var_2022_to_fp16)[name = tensor("op_2023_cast_fp16")]; + tensor input_467_cast_fp16 = add(x = input_455_cast_fp16, y = var_2023_cast_fp16)[name = tensor("input_467_cast_fp16")]; + tensor input_469_axes_0 = const()[name = tensor("input_469_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110333888)))]; + tensor encoder_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110334976)))]; + tensor input_469_cast_fp16 = layer_norm(axes = input_469_axes_0, beta = encoder_layers_8_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_8_norm_out_weight_to_fp16, x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; + tensor cache_37_begin_0 = const()[name = tensor("cache_37_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_37_end_0 = const()[name = tensor("cache_37_end_0"), val = tensor([10, 1, 70, 512])]; + tensor cache_37_end_mask_0 = const()[name = tensor("cache_37_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_37_squeeze_mask_0 = const()[name = tensor("cache_37_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_37_cast_fp16 = slice_by_index(begin = cache_37_begin_0, end = cache_37_end_0, end_mask = cache_37_end_mask_0, squeeze_mask = cache_37_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_37_cast_fp16")]; + tensor cache_39_begin_0 = const()[name = tensor("cache_39_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_39_end_0 = const()[name = tensor("cache_39_end_0"), val = tensor([10, 1, 512, 8])]; + tensor cache_39_end_mask_0 = const()[name = tensor("cache_39_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_39_squeeze_mask_0 = const()[name = tensor("cache_39_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_39_cast_fp16 = slice_by_index(begin = cache_39_begin_0, end = cache_39_end_0, end_mask = cache_39_end_mask_0, squeeze_mask = cache_39_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_39_cast_fp16")]; + tensor input_471_axes_0 = const()[name = tensor("input_471_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110336064)))]; + tensor encoder_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110337152)))]; + tensor input_471_cast_fp16 = layer_norm(axes = input_471_axes_0, beta = encoder_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_9_norm_feed_forward1_weight_to_fp16, x = input_469_cast_fp16)[name = tensor("input_471_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110338240)))]; + tensor linear_81_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_471_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor input_475_cast_fp16 = silu(x = linear_81_cast_fp16)[name = tensor("input_475_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112435456)))]; + tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor var_2057_to_fp16 = const()[name = tensor("op_2057_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2058_cast_fp16 = mul(x = linear_82_cast_fp16, y = var_2057_to_fp16)[name = tensor("op_2058_cast_fp16")]; + tensor input_481_cast_fp16 = add(x = input_469_cast_fp16, y = var_2058_cast_fp16)[name = tensor("input_481_cast_fp16")]; + tensor key_19_axes_0 = const()[name = tensor("key_19_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114532672)))]; + tensor encoder_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114533760)))]; + tensor key_19_cast_fp16 = layer_norm(axes = key_19_axes_0, beta = encoder_layers_9_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_9_norm_self_att_weight_to_fp16, x = input_481_cast_fp16)[name = tensor("key_19_cast_fp16")]; + tensor input_483_interleave_0 = const()[name = tensor("input_483_interleave_0"), val = tensor(false)]; + tensor input_483_cast_fp16 = concat(axis = var_44, interleave = input_483_interleave_0, values = (cache_37_cast_fp16, key_19_cast_fp16))[name = tensor("input_483_cast_fp16")]; + tensor var_2080_begin_0 = const()[name = tensor("op_2080_begin_0"), val = tensor([0, 14, 0])]; + tensor var_2080_end_0 = const()[name = tensor("op_2080_end_0"), val = tensor([1, 70, 512])]; + tensor var_2080_end_mask_0 = const()[name = tensor("op_2080_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2080_cast_fp16 = slice_by_index(begin = var_2080_begin_0, end = var_2080_end_0, end_mask = var_2080_end_mask_0, x = cache_37_cast_fp16)[name = tensor("op_2080_cast_fp16")]; + tensor var_2083_begin_0 = const()[name = tensor("op_2083_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2083_end_0 = const()[name = tensor("op_2083_end_0"), val = tensor([1, 14, 512])]; + tensor var_2083_end_mask_0 = const()[name = tensor("op_2083_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2083_cast_fp16 = slice_by_index(begin = var_2083_begin_0, end = var_2083_end_0, end_mask = var_2083_end_mask_0, x = key_19_cast_fp16)[name = tensor("op_2083_cast_fp16")]; + tensor var_2086_interleave_0 = const()[name = tensor("op_2086_interleave_0"), val = tensor(false)]; + tensor var_2086_cast_fp16 = concat(axis = var_44, interleave = var_2086_interleave_0, values = (var_2080_cast_fp16, var_2083_cast_fp16))[name = tensor("op_2086_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114534848)))]; + tensor linear_83_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_q_weight_to_fp16, x = key_19_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor var_2090 = const()[name = tensor("op_2090"), val = tensor([1, -1, 8, 64])]; + tensor q_55_cast_fp16 = reshape(shape = var_2090, x = linear_83_cast_fp16)[name = tensor("q_55_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115059200)))]; + tensor linear_84_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_k_weight_to_fp16, x = input_483_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_2094 = const()[name = tensor("op_2094"), val = tensor([1, -1, 8, 64])]; + tensor k_37_cast_fp16 = reshape(shape = var_2094, x = linear_84_cast_fp16)[name = tensor("k_37_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115583552)))]; + tensor linear_85_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_v_weight_to_fp16, x = input_483_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_2098 = const()[name = tensor("op_2098"), val = tensor([1, -1, 8, 64])]; + tensor v_19_cast_fp16 = reshape(shape = var_2098, x = linear_85_cast_fp16)[name = tensor("v_19_cast_fp16")]; + tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116107904)))]; + tensor var_2110_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2110_cast_fp16")]; + tensor encoder_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116108992)))]; + tensor var_2112_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2112_cast_fp16")]; + tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_237_transpose_x_0 = const()[name = tensor("x_237_transpose_x_0"), val = tensor(false)]; + tensor x_237_transpose_y_0 = const()[name = tensor("x_237_transpose_y_0"), val = tensor(false)]; + tensor var_2114_to_fp16 = const()[name = tensor("op_2114_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116110080)))]; + tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_2112_cast_fp16)[name = tensor("transpose_156")]; + tensor x_237_cast_fp16 = matmul(transpose_x = x_237_transpose_x_0, transpose_y = x_237_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_2114_to_fp16)[name = tensor("x_237_cast_fp16")]; + tensor x_239_pad_0 = const()[name = tensor("x_239_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_239_mode_0 = const()[name = tensor("x_239_mode_0"), val = tensor("constant")]; + tensor const_135_to_fp16 = const()[name = tensor("const_135_to_fp16"), val = tensor(0x0p+0)]; + tensor x_239_cast_fp16 = pad(constant_val = const_135_to_fp16, mode = x_239_mode_0, pad = x_239_pad_0, x = x_237_cast_fp16)[name = tensor("x_239_cast_fp16")]; + tensor var_2122 = const()[name = tensor("op_2122"), val = tensor([1, 8, -1, 16])]; + tensor x_241_cast_fp16 = reshape(shape = var_2122, x = x_239_cast_fp16)[name = tensor("x_241_cast_fp16")]; + tensor var_2126_begin_0 = const()[name = tensor("op_2126_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2126_end_0 = const()[name = tensor("op_2126_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_2126_end_mask_0 = const()[name = tensor("op_2126_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2126_cast_fp16 = slice_by_index(begin = var_2126_begin_0, end = var_2126_end_0, end_mask = var_2126_end_mask_0, x = x_241_cast_fp16)[name = tensor("op_2126_cast_fp16")]; + tensor var_2127 = const()[name = tensor("op_2127"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_37_cast_fp16 = reshape(shape = var_2127, x = var_2126_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; + tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; + tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_154")]; + tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_2110_cast_fp16)[name = tensor("transpose_155")]; + tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; + tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; + tensor var_2136_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_2136_cast_fp16")]; + tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_37_cast_fp16 = mul(x = var_2136_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; + tensor scores_39_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_3)[name = tensor("scores_39_cast_fp16")]; + tensor var_2142_cast_fp16 = softmax(axis = var_30, x = scores_39_cast_fp16)[name = tensor("op_2142_cast_fp16")]; + tensor input_485_cast_fp16 = select(a = var_23_to_fp16, b = var_2142_cast_fp16, cond = mask_3)[name = tensor("input_485_cast_fp16")]; + tensor x_243_transpose_x_0 = const()[name = tensor("x_243_transpose_x_0"), val = tensor(false)]; + tensor x_243_transpose_y_0 = const()[name = tensor("x_243_transpose_y_0"), val = tensor(false)]; + tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_157")]; + tensor x_243_cast_fp16 = matmul(transpose_x = x_243_transpose_x_0, transpose_y = x_243_transpose_y_0, x = input_485_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_243_cast_fp16")]; + tensor var_2146_perm_0 = const()[name = tensor("op_2146_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2147 = const()[name = tensor("op_2147"), val = tensor([1, -1, 512])]; + tensor var_2146_cast_fp16 = transpose(perm = var_2146_perm_0, x = x_243_cast_fp16)[name = tensor("transpose_153")]; + tensor input_487_cast_fp16 = reshape(shape = var_2147, x = var_2146_cast_fp16)[name = tensor("input_487_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116285248)))]; + tensor linear_87_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_out_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("linear_87_cast_fp16")]; + tensor input_491_cast_fp16 = add(x = input_481_cast_fp16, y = linear_87_cast_fp16)[name = tensor("input_491_cast_fp16")]; + tensor x_247_axes_0 = const()[name = tensor("x_247_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116809600)))]; + tensor encoder_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116810688)))]; + tensor x_247_cast_fp16 = layer_norm(axes = x_247_axes_0, beta = encoder_layers_9_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_9_norm_conv_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("x_247_cast_fp16")]; + tensor input_493_perm_0 = const()[name = tensor("input_493_perm_0"), val = tensor([0, 2, 1])]; + tensor input_495_pad_type_0 = const()[name = tensor("input_495_pad_type_0"), val = tensor("valid")]; + tensor input_495_strides_0 = const()[name = tensor("input_495_strides_0"), val = tensor([1])]; + tensor input_495_pad_0 = const()[name = tensor("input_495_pad_0"), val = tensor([0, 0])]; + tensor input_495_dilations_0 = const()[name = tensor("input_495_dilations_0"), val = tensor([1])]; + tensor input_495_groups_0 = const()[name = tensor("input_495_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116811776)))]; + tensor input_493_cast_fp16 = transpose(perm = input_493_perm_0, x = x_247_cast_fp16)[name = tensor("transpose_152")]; + tensor input_495_cast_fp16 = conv(dilations = input_495_dilations_0, groups = input_495_groups_0, pad = input_495_pad_0, pad_type = input_495_pad_type_0, strides = input_495_strides_0, weight = encoder_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("input_495_cast_fp16")]; + tensor x_249_split_num_splits_0 = const()[name = tensor("x_249_split_num_splits_0"), val = tensor(2)]; + tensor x_249_split_axis_0 = const()[name = tensor("x_249_split_axis_0"), val = tensor(1)]; + tensor x_249_split_cast_fp16_0, tensor x_249_split_cast_fp16_1 = split(axis = x_249_split_axis_0, num_splits = x_249_split_num_splits_0, x = input_495_cast_fp16)[name = tensor("x_249_split_cast_fp16")]; + tensor x_249_split_1_sigmoid_cast_fp16 = sigmoid(x = x_249_split_cast_fp16_1)[name = tensor("x_249_split_1_sigmoid_cast_fp16")]; + tensor x_249_cast_fp16 = mul(x = x_249_split_cast_fp16_0, y = x_249_split_1_sigmoid_cast_fp16)[name = tensor("x_249_cast_fp16")]; + tensor input_497_cast_fp16 = select(a = var_23_to_fp16, b = x_249_cast_fp16, cond = var_309)[name = tensor("input_497_cast_fp16")]; + tensor new_x_39_interleave_0 = const()[name = tensor("new_x_39_interleave_0"), val = tensor(false)]; + tensor new_x_39_cast_fp16 = concat(axis = var_30, interleave = new_x_39_interleave_0, values = (cache_39_cast_fp16, input_497_cast_fp16))[name = tensor("new_x_39_cast_fp16")]; + tensor next_cache_19_begin_0 = const()[name = tensor("next_cache_19_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_19_end_0 = const()[name = tensor("next_cache_19_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_19_end_mask_0 = const()[name = tensor("next_cache_19_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_19_cast_fp16 = slice_by_index(begin = next_cache_19_begin_0, end = next_cache_19_end_0, end_mask = next_cache_19_end_mask_0, x = new_x_39_cast_fp16)[name = tensor("next_cache_19_cast_fp16")]; + tensor var_2188_begin_0 = const()[name = tensor("op_2188_begin_0"), val = tensor([0, 0, 14])]; + tensor var_2188_end_0 = const()[name = tensor("op_2188_end_0"), val = tensor([1, 512, 22])]; + tensor var_2188_end_mask_0 = const()[name = tensor("op_2188_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2188_cast_fp16 = slice_by_index(begin = var_2188_begin_0, end = var_2188_end_0, end_mask = var_2188_end_mask_0, x = next_cache_19_cast_fp16)[name = tensor("op_2188_cast_fp16")]; + tensor x_251_pad_type_0 = const()[name = tensor("x_251_pad_type_0"), val = tensor("valid")]; + tensor x_251_groups_0 = const()[name = tensor("x_251_groups_0"), val = tensor(512)]; + tensor x_251_strides_0 = const()[name = tensor("x_251_strides_0"), val = tensor([1])]; + tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0])]; + tensor x_251_dilations_0 = const()[name = tensor("x_251_dilations_0"), val = tensor([1])]; + tensor encoder_layers_9_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117860416)))]; + tensor x_251_cast_fp16 = conv(dilations = x_251_dilations_0, groups = x_251_groups_0, pad = x_251_pad_0, pad_type = x_251_pad_type_0, strides = x_251_strides_0, weight = encoder_layers_9_conv_depthwise_conv_weight_to_fp16, x = new_x_39_cast_fp16)[name = tensor("x_251_cast_fp16")]; + tensor input_499_perm_0 = const()[name = tensor("input_499_perm_0"), val = tensor([0, 2, 1])]; + tensor x_253_axes_0 = const()[name = tensor("x_253_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117869696)))]; + tensor encoder_layers_9_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117870784)))]; + tensor input_499_cast_fp16 = transpose(perm = input_499_perm_0, x = x_251_cast_fp16)[name = tensor("transpose_151")]; + tensor x_253_cast_fp16 = layer_norm(axes = x_253_axes_0, beta = encoder_layers_9_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_9_conv_batch_norm_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("x_253_cast_fp16")]; + tensor input_501_perm_0 = const()[name = tensor("input_501_perm_0"), val = tensor([0, 2, 1])]; + tensor input_501_cast_fp16 = transpose(perm = input_501_perm_0, x = x_253_cast_fp16)[name = tensor("transpose_150")]; + tensor input_503_cast_fp16 = silu(x = input_501_cast_fp16)[name = tensor("input_503_cast_fp16")]; + tensor x_255_pad_type_0 = const()[name = tensor("x_255_pad_type_0"), val = tensor("valid")]; + tensor x_255_strides_0 = const()[name = tensor("x_255_strides_0"), val = tensor([1])]; + tensor x_255_pad_0 = const()[name = tensor("x_255_pad_0"), val = tensor([0, 0])]; + tensor x_255_dilations_0 = const()[name = tensor("x_255_dilations_0"), val = tensor([1])]; + tensor x_255_groups_0 = const()[name = tensor("x_255_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117871872)))]; + tensor x_255_cast_fp16 = conv(dilations = x_255_dilations_0, groups = x_255_groups_0, pad = x_255_pad_0, pad_type = x_255_pad_type_0, strides = x_255_strides_0, weight = encoder_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_503_cast_fp16)[name = tensor("x_255_cast_fp16")]; + tensor input_505_perm_0 = const()[name = tensor("input_505_perm_0"), val = tensor([0, 2, 1])]; + tensor input_505_cast_fp16 = transpose(perm = input_505_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_149")]; + tensor input_507_cast_fp16 = add(x = input_491_cast_fp16, y = input_505_cast_fp16)[name = tensor("input_507_cast_fp16")]; + tensor input_509_axes_0 = const()[name = tensor("input_509_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(118396224)))]; + tensor encoder_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(118397312)))]; + tensor input_509_cast_fp16 = layer_norm(axes = input_509_axes_0, beta = encoder_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_9_norm_feed_forward2_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("input_509_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(118398400)))]; + tensor linear_88_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_513_cast_fp16 = silu(x = linear_88_cast_fp16)[name = tensor("input_513_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120495616)))]; + tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_513_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor var_2229_to_fp16 = const()[name = tensor("op_2229_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2230_cast_fp16 = mul(x = linear_89_cast_fp16, y = var_2229_to_fp16)[name = tensor("op_2230_cast_fp16")]; + tensor input_519_cast_fp16 = add(x = input_507_cast_fp16, y = var_2230_cast_fp16)[name = tensor("input_519_cast_fp16")]; + tensor input_521_axes_0 = const()[name = tensor("input_521_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122592832)))]; + tensor encoder_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122593920)))]; + tensor input_521_cast_fp16 = layer_norm(axes = input_521_axes_0, beta = encoder_layers_9_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_9_norm_out_weight_to_fp16, x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; + tensor cache_41_begin_0 = const()[name = tensor("cache_41_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_41_end_0 = const()[name = tensor("cache_41_end_0"), val = tensor([11, 1, 70, 512])]; + tensor cache_41_end_mask_0 = const()[name = tensor("cache_41_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_41_squeeze_mask_0 = const()[name = tensor("cache_41_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_41_cast_fp16 = slice_by_index(begin = cache_41_begin_0, end = cache_41_end_0, end_mask = cache_41_end_mask_0, squeeze_mask = cache_41_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_41_cast_fp16")]; + tensor cache_43_begin_0 = const()[name = tensor("cache_43_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_43_end_0 = const()[name = tensor("cache_43_end_0"), val = tensor([11, 1, 512, 8])]; + tensor cache_43_end_mask_0 = const()[name = tensor("cache_43_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_43_squeeze_mask_0 = const()[name = tensor("cache_43_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_43_cast_fp16 = slice_by_index(begin = cache_43_begin_0, end = cache_43_end_0, end_mask = cache_43_end_mask_0, squeeze_mask = cache_43_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_43_cast_fp16")]; + tensor input_523_axes_0 = const()[name = tensor("input_523_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122595008)))]; + tensor encoder_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122596096)))]; + tensor input_523_cast_fp16 = layer_norm(axes = input_523_axes_0, beta = encoder_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_10_norm_feed_forward1_weight_to_fp16, x = input_521_cast_fp16)[name = tensor("input_523_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122597184)))]; + tensor linear_90_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_523_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor input_527_cast_fp16 = silu(x = linear_90_cast_fp16)[name = tensor("input_527_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(124694400)))]; + tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor var_2264_to_fp16 = const()[name = tensor("op_2264_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2265_cast_fp16 = mul(x = linear_91_cast_fp16, y = var_2264_to_fp16)[name = tensor("op_2265_cast_fp16")]; + tensor input_533_cast_fp16 = add(x = input_521_cast_fp16, y = var_2265_cast_fp16)[name = tensor("input_533_cast_fp16")]; + tensor key_21_axes_0 = const()[name = tensor("key_21_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126791616)))]; + tensor encoder_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126792704)))]; + tensor key_21_cast_fp16 = layer_norm(axes = key_21_axes_0, beta = encoder_layers_10_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_10_norm_self_att_weight_to_fp16, x = input_533_cast_fp16)[name = tensor("key_21_cast_fp16")]; + tensor input_535_interleave_0 = const()[name = tensor("input_535_interleave_0"), val = tensor(false)]; + tensor input_535_cast_fp16 = concat(axis = var_44, interleave = input_535_interleave_0, values = (cache_41_cast_fp16, key_21_cast_fp16))[name = tensor("input_535_cast_fp16")]; + tensor var_2287_begin_0 = const()[name = tensor("op_2287_begin_0"), val = tensor([0, 14, 0])]; + tensor var_2287_end_0 = const()[name = tensor("op_2287_end_0"), val = tensor([1, 70, 512])]; + tensor var_2287_end_mask_0 = const()[name = tensor("op_2287_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2287_cast_fp16 = slice_by_index(begin = var_2287_begin_0, end = var_2287_end_0, end_mask = var_2287_end_mask_0, x = cache_41_cast_fp16)[name = tensor("op_2287_cast_fp16")]; + tensor var_2290_begin_0 = const()[name = tensor("op_2290_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2290_end_0 = const()[name = tensor("op_2290_end_0"), val = tensor([1, 14, 512])]; + tensor var_2290_end_mask_0 = const()[name = tensor("op_2290_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2290_cast_fp16 = slice_by_index(begin = var_2290_begin_0, end = var_2290_end_0, end_mask = var_2290_end_mask_0, x = key_21_cast_fp16)[name = tensor("op_2290_cast_fp16")]; + tensor var_2293_interleave_0 = const()[name = tensor("op_2293_interleave_0"), val = tensor(false)]; + tensor var_2293_cast_fp16 = concat(axis = var_44, interleave = var_2293_interleave_0, values = (var_2287_cast_fp16, var_2290_cast_fp16))[name = tensor("op_2293_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126793792)))]; + tensor linear_92_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_q_weight_to_fp16, x = key_21_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_2297 = const()[name = tensor("op_2297"), val = tensor([1, -1, 8, 64])]; + tensor q_61_cast_fp16 = reshape(shape = var_2297, x = linear_92_cast_fp16)[name = tensor("q_61_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127318144)))]; + tensor linear_93_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_k_weight_to_fp16, x = input_535_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor var_2301 = const()[name = tensor("op_2301"), val = tensor([1, -1, 8, 64])]; + tensor k_41_cast_fp16 = reshape(shape = var_2301, x = linear_93_cast_fp16)[name = tensor("k_41_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127842496)))]; + tensor linear_94_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_v_weight_to_fp16, x = input_535_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor var_2305 = const()[name = tensor("op_2305"), val = tensor([1, -1, 8, 64])]; + tensor v_21_cast_fp16 = reshape(shape = var_2305, x = linear_94_cast_fp16)[name = tensor("v_21_cast_fp16")]; + tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128366848)))]; + tensor var_2317_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2317_cast_fp16")]; + tensor encoder_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128367936)))]; + tensor var_2319_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2319_cast_fp16")]; + tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_263_transpose_x_0 = const()[name = tensor("x_263_transpose_x_0"), val = tensor(false)]; + tensor x_263_transpose_y_0 = const()[name = tensor("x_263_transpose_y_0"), val = tensor(false)]; + tensor var_2321_to_fp16 = const()[name = tensor("op_2321_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128369024)))]; + tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2319_cast_fp16)[name = tensor("transpose_147")]; + tensor x_263_cast_fp16 = matmul(transpose_x = x_263_transpose_x_0, transpose_y = x_263_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2321_to_fp16)[name = tensor("x_263_cast_fp16")]; + tensor x_265_pad_0 = const()[name = tensor("x_265_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_265_mode_0 = const()[name = tensor("x_265_mode_0"), val = tensor("constant")]; + tensor const_148_to_fp16 = const()[name = tensor("const_148_to_fp16"), val = tensor(0x0p+0)]; + tensor x_265_cast_fp16 = pad(constant_val = const_148_to_fp16, mode = x_265_mode_0, pad = x_265_pad_0, x = x_263_cast_fp16)[name = tensor("x_265_cast_fp16")]; + tensor var_2329 = const()[name = tensor("op_2329"), val = tensor([1, 8, -1, 16])]; + tensor x_267_cast_fp16 = reshape(shape = var_2329, x = x_265_cast_fp16)[name = tensor("x_267_cast_fp16")]; + tensor var_2333_begin_0 = const()[name = tensor("op_2333_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2333_end_0 = const()[name = tensor("op_2333_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_2333_end_mask_0 = const()[name = tensor("op_2333_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2333_cast_fp16 = slice_by_index(begin = var_2333_begin_0, end = var_2333_end_0, end_mask = var_2333_end_mask_0, x = x_267_cast_fp16)[name = tensor("op_2333_cast_fp16")]; + tensor var_2334 = const()[name = tensor("op_2334"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2334, x = var_2333_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; + tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; + tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_145")]; + tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2317_cast_fp16)[name = tensor("transpose_146")]; + tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; + tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; + tensor var_2343_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2343_cast_fp16")]; + tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_41_cast_fp16 = mul(x = var_2343_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; + tensor scores_43_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_3)[name = tensor("scores_43_cast_fp16")]; + tensor var_2349_cast_fp16 = softmax(axis = var_30, x = scores_43_cast_fp16)[name = tensor("op_2349_cast_fp16")]; + tensor input_537_cast_fp16 = select(a = var_23_to_fp16, b = var_2349_cast_fp16, cond = mask_3)[name = tensor("input_537_cast_fp16")]; + tensor x_269_transpose_x_0 = const()[name = tensor("x_269_transpose_x_0"), val = tensor(false)]; + tensor x_269_transpose_y_0 = const()[name = tensor("x_269_transpose_y_0"), val = tensor(false)]; + tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_148")]; + tensor x_269_cast_fp16 = matmul(transpose_x = x_269_transpose_x_0, transpose_y = x_269_transpose_y_0, x = input_537_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_269_cast_fp16")]; + tensor var_2353_perm_0 = const()[name = tensor("op_2353_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2354 = const()[name = tensor("op_2354"), val = tensor([1, -1, 512])]; + tensor var_2353_cast_fp16 = transpose(perm = var_2353_perm_0, x = x_269_cast_fp16)[name = tensor("transpose_144")]; + tensor input_539_cast_fp16 = reshape(shape = var_2354, x = var_2353_cast_fp16)[name = tensor("input_539_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128544192)))]; + tensor linear_96_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_out_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("linear_96_cast_fp16")]; + tensor input_543_cast_fp16 = add(x = input_533_cast_fp16, y = linear_96_cast_fp16)[name = tensor("input_543_cast_fp16")]; + tensor x_273_axes_0 = const()[name = tensor("x_273_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129068544)))]; + tensor encoder_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129069632)))]; + tensor x_273_cast_fp16 = layer_norm(axes = x_273_axes_0, beta = encoder_layers_10_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_10_norm_conv_weight_to_fp16, x = input_543_cast_fp16)[name = tensor("x_273_cast_fp16")]; + tensor input_545_perm_0 = const()[name = tensor("input_545_perm_0"), val = tensor([0, 2, 1])]; + tensor input_547_pad_type_0 = const()[name = tensor("input_547_pad_type_0"), val = tensor("valid")]; + tensor input_547_strides_0 = const()[name = tensor("input_547_strides_0"), val = tensor([1])]; + tensor input_547_pad_0 = const()[name = tensor("input_547_pad_0"), val = tensor([0, 0])]; + tensor input_547_dilations_0 = const()[name = tensor("input_547_dilations_0"), val = tensor([1])]; + tensor input_547_groups_0 = const()[name = tensor("input_547_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129070720)))]; + tensor input_545_cast_fp16 = transpose(perm = input_545_perm_0, x = x_273_cast_fp16)[name = tensor("transpose_143")]; + tensor input_547_cast_fp16 = conv(dilations = input_547_dilations_0, groups = input_547_groups_0, pad = input_547_pad_0, pad_type = input_547_pad_type_0, strides = input_547_strides_0, weight = encoder_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("input_547_cast_fp16")]; + tensor x_275_split_num_splits_0 = const()[name = tensor("x_275_split_num_splits_0"), val = tensor(2)]; + tensor x_275_split_axis_0 = const()[name = tensor("x_275_split_axis_0"), val = tensor(1)]; + tensor x_275_split_cast_fp16_0, tensor x_275_split_cast_fp16_1 = split(axis = x_275_split_axis_0, num_splits = x_275_split_num_splits_0, x = input_547_cast_fp16)[name = tensor("x_275_split_cast_fp16")]; + tensor x_275_split_1_sigmoid_cast_fp16 = sigmoid(x = x_275_split_cast_fp16_1)[name = tensor("x_275_split_1_sigmoid_cast_fp16")]; + tensor x_275_cast_fp16 = mul(x = x_275_split_cast_fp16_0, y = x_275_split_1_sigmoid_cast_fp16)[name = tensor("x_275_cast_fp16")]; + tensor input_549_cast_fp16 = select(a = var_23_to_fp16, b = x_275_cast_fp16, cond = var_309)[name = tensor("input_549_cast_fp16")]; + tensor new_x_43_interleave_0 = const()[name = tensor("new_x_43_interleave_0"), val = tensor(false)]; + tensor new_x_43_cast_fp16 = concat(axis = var_30, interleave = new_x_43_interleave_0, values = (cache_43_cast_fp16, input_549_cast_fp16))[name = tensor("new_x_43_cast_fp16")]; + tensor next_cache_21_begin_0 = const()[name = tensor("next_cache_21_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_21_end_0 = const()[name = tensor("next_cache_21_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_21_end_mask_0 = const()[name = tensor("next_cache_21_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_21_cast_fp16 = slice_by_index(begin = next_cache_21_begin_0, end = next_cache_21_end_0, end_mask = next_cache_21_end_mask_0, x = new_x_43_cast_fp16)[name = tensor("next_cache_21_cast_fp16")]; + tensor var_2395_begin_0 = const()[name = tensor("op_2395_begin_0"), val = tensor([0, 0, 14])]; + tensor var_2395_end_0 = const()[name = tensor("op_2395_end_0"), val = tensor([1, 512, 22])]; + tensor var_2395_end_mask_0 = const()[name = tensor("op_2395_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2395_cast_fp16 = slice_by_index(begin = var_2395_begin_0, end = var_2395_end_0, end_mask = var_2395_end_mask_0, x = next_cache_21_cast_fp16)[name = tensor("op_2395_cast_fp16")]; + tensor x_277_pad_type_0 = const()[name = tensor("x_277_pad_type_0"), val = tensor("valid")]; + tensor x_277_groups_0 = const()[name = tensor("x_277_groups_0"), val = tensor(512)]; + tensor x_277_strides_0 = const()[name = tensor("x_277_strides_0"), val = tensor([1])]; + tensor x_277_pad_0 = const()[name = tensor("x_277_pad_0"), val = tensor([0, 0])]; + tensor x_277_dilations_0 = const()[name = tensor("x_277_dilations_0"), val = tensor([1])]; + tensor encoder_layers_10_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130119360)))]; + tensor x_277_cast_fp16 = conv(dilations = x_277_dilations_0, groups = x_277_groups_0, pad = x_277_pad_0, pad_type = x_277_pad_type_0, strides = x_277_strides_0, weight = encoder_layers_10_conv_depthwise_conv_weight_to_fp16, x = new_x_43_cast_fp16)[name = tensor("x_277_cast_fp16")]; + tensor input_551_perm_0 = const()[name = tensor("input_551_perm_0"), val = tensor([0, 2, 1])]; + tensor x_279_axes_0 = const()[name = tensor("x_279_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130128640)))]; + tensor encoder_layers_10_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130129728)))]; + tensor input_551_cast_fp16 = transpose(perm = input_551_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_142")]; + tensor x_279_cast_fp16 = layer_norm(axes = x_279_axes_0, beta = encoder_layers_10_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_10_conv_batch_norm_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("x_279_cast_fp16")]; + tensor input_553_perm_0 = const()[name = tensor("input_553_perm_0"), val = tensor([0, 2, 1])]; + tensor input_553_cast_fp16 = transpose(perm = input_553_perm_0, x = x_279_cast_fp16)[name = tensor("transpose_141")]; + tensor input_555_cast_fp16 = silu(x = input_553_cast_fp16)[name = tensor("input_555_cast_fp16")]; + tensor x_281_pad_type_0 = const()[name = tensor("x_281_pad_type_0"), val = tensor("valid")]; + tensor x_281_strides_0 = const()[name = tensor("x_281_strides_0"), val = tensor([1])]; + tensor x_281_pad_0 = const()[name = tensor("x_281_pad_0"), val = tensor([0, 0])]; + tensor x_281_dilations_0 = const()[name = tensor("x_281_dilations_0"), val = tensor([1])]; + tensor x_281_groups_0 = const()[name = tensor("x_281_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130130816)))]; + tensor x_281_cast_fp16 = conv(dilations = x_281_dilations_0, groups = x_281_groups_0, pad = x_281_pad_0, pad_type = x_281_pad_type_0, strides = x_281_strides_0, weight = encoder_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_555_cast_fp16)[name = tensor("x_281_cast_fp16")]; + tensor input_557_perm_0 = const()[name = tensor("input_557_perm_0"), val = tensor([0, 2, 1])]; + tensor input_557_cast_fp16 = transpose(perm = input_557_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_140")]; + tensor input_559_cast_fp16 = add(x = input_543_cast_fp16, y = input_557_cast_fp16)[name = tensor("input_559_cast_fp16")]; + tensor input_561_axes_0 = const()[name = tensor("input_561_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130655168)))]; + tensor encoder_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130656256)))]; + tensor input_561_cast_fp16 = layer_norm(axes = input_561_axes_0, beta = encoder_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_10_norm_feed_forward2_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("input_561_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130657344)))]; + tensor linear_97_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_561_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor input_565_cast_fp16 = silu(x = linear_97_cast_fp16)[name = tensor("input_565_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132754560)))]; + tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_565_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor var_2436_to_fp16 = const()[name = tensor("op_2436_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2437_cast_fp16 = mul(x = linear_98_cast_fp16, y = var_2436_to_fp16)[name = tensor("op_2437_cast_fp16")]; + tensor input_571_cast_fp16 = add(x = input_559_cast_fp16, y = var_2437_cast_fp16)[name = tensor("input_571_cast_fp16")]; + tensor input_573_axes_0 = const()[name = tensor("input_573_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134851776)))]; + tensor encoder_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134852864)))]; + tensor input_573_cast_fp16 = layer_norm(axes = input_573_axes_0, beta = encoder_layers_10_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_10_norm_out_weight_to_fp16, x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; + tensor cache_45_begin_0 = const()[name = tensor("cache_45_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_45_end_0 = const()[name = tensor("cache_45_end_0"), val = tensor([12, 1, 70, 512])]; + tensor cache_45_end_mask_0 = const()[name = tensor("cache_45_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_45_squeeze_mask_0 = const()[name = tensor("cache_45_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_45_cast_fp16 = slice_by_index(begin = cache_45_begin_0, end = cache_45_end_0, end_mask = cache_45_end_mask_0, squeeze_mask = cache_45_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_45_cast_fp16")]; + tensor cache_47_begin_0 = const()[name = tensor("cache_47_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_47_end_0 = const()[name = tensor("cache_47_end_0"), val = tensor([12, 1, 512, 8])]; + tensor cache_47_end_mask_0 = const()[name = tensor("cache_47_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_47_squeeze_mask_0 = const()[name = tensor("cache_47_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_47_cast_fp16 = slice_by_index(begin = cache_47_begin_0, end = cache_47_end_0, end_mask = cache_47_end_mask_0, squeeze_mask = cache_47_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_47_cast_fp16")]; + tensor input_575_axes_0 = const()[name = tensor("input_575_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134853952)))]; + tensor encoder_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134855040)))]; + tensor input_575_cast_fp16 = layer_norm(axes = input_575_axes_0, beta = encoder_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_11_norm_feed_forward1_weight_to_fp16, x = input_573_cast_fp16)[name = tensor("input_575_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134856128)))]; + tensor linear_99_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_575_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor input_579_cast_fp16 = silu(x = linear_99_cast_fp16)[name = tensor("input_579_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136953344)))]; + tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor var_2471_to_fp16 = const()[name = tensor("op_2471_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2472_cast_fp16 = mul(x = linear_100_cast_fp16, y = var_2471_to_fp16)[name = tensor("op_2472_cast_fp16")]; + tensor input_585_cast_fp16 = add(x = input_573_cast_fp16, y = var_2472_cast_fp16)[name = tensor("input_585_cast_fp16")]; + tensor key_23_axes_0 = const()[name = tensor("key_23_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139050560)))]; + tensor encoder_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139051648)))]; + tensor key_23_cast_fp16 = layer_norm(axes = key_23_axes_0, beta = encoder_layers_11_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_11_norm_self_att_weight_to_fp16, x = input_585_cast_fp16)[name = tensor("key_23_cast_fp16")]; + tensor input_587_interleave_0 = const()[name = tensor("input_587_interleave_0"), val = tensor(false)]; + tensor input_587_cast_fp16 = concat(axis = var_44, interleave = input_587_interleave_0, values = (cache_45_cast_fp16, key_23_cast_fp16))[name = tensor("input_587_cast_fp16")]; + tensor var_2494_begin_0 = const()[name = tensor("op_2494_begin_0"), val = tensor([0, 14, 0])]; + tensor var_2494_end_0 = const()[name = tensor("op_2494_end_0"), val = tensor([1, 70, 512])]; + tensor var_2494_end_mask_0 = const()[name = tensor("op_2494_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2494_cast_fp16 = slice_by_index(begin = var_2494_begin_0, end = var_2494_end_0, end_mask = var_2494_end_mask_0, x = cache_45_cast_fp16)[name = tensor("op_2494_cast_fp16")]; + tensor var_2497_begin_0 = const()[name = tensor("op_2497_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2497_end_0 = const()[name = tensor("op_2497_end_0"), val = tensor([1, 14, 512])]; + tensor var_2497_end_mask_0 = const()[name = tensor("op_2497_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2497_cast_fp16 = slice_by_index(begin = var_2497_begin_0, end = var_2497_end_0, end_mask = var_2497_end_mask_0, x = key_23_cast_fp16)[name = tensor("op_2497_cast_fp16")]; + tensor var_2500_interleave_0 = const()[name = tensor("op_2500_interleave_0"), val = tensor(false)]; + tensor var_2500_cast_fp16 = concat(axis = var_44, interleave = var_2500_interleave_0, values = (var_2494_cast_fp16, var_2497_cast_fp16))[name = tensor("op_2500_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139052736)))]; + tensor linear_101_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_q_weight_to_fp16, x = key_23_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor var_2504 = const()[name = tensor("op_2504"), val = tensor([1, -1, 8, 64])]; + tensor q_67_cast_fp16 = reshape(shape = var_2504, x = linear_101_cast_fp16)[name = tensor("q_67_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139577088)))]; + tensor linear_102_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_k_weight_to_fp16, x = input_587_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_2508 = const()[name = tensor("op_2508"), val = tensor([1, -1, 8, 64])]; + tensor k_45_cast_fp16 = reshape(shape = var_2508, x = linear_102_cast_fp16)[name = tensor("k_45_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140101440)))]; + tensor linear_103_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_v_weight_to_fp16, x = input_587_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_2512 = const()[name = tensor("op_2512"), val = tensor([1, -1, 8, 64])]; + tensor v_23_cast_fp16 = reshape(shape = var_2512, x = linear_103_cast_fp16)[name = tensor("v_23_cast_fp16")]; + tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140625792)))]; + tensor var_2524_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2524_cast_fp16")]; + tensor encoder_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140626880)))]; + tensor var_2526_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2526_cast_fp16")]; + tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_289_transpose_x_0 = const()[name = tensor("x_289_transpose_x_0"), val = tensor(false)]; + tensor x_289_transpose_y_0 = const()[name = tensor("x_289_transpose_y_0"), val = tensor(false)]; + tensor var_2528_to_fp16 = const()[name = tensor("op_2528_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140627968)))]; + tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2526_cast_fp16)[name = tensor("transpose_138")]; + tensor x_289_cast_fp16 = matmul(transpose_x = x_289_transpose_x_0, transpose_y = x_289_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2528_to_fp16)[name = tensor("x_289_cast_fp16")]; + tensor x_291_pad_0 = const()[name = tensor("x_291_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_291_mode_0 = const()[name = tensor("x_291_mode_0"), val = tensor("constant")]; + tensor const_161_to_fp16 = const()[name = tensor("const_161_to_fp16"), val = tensor(0x0p+0)]; + tensor x_291_cast_fp16 = pad(constant_val = const_161_to_fp16, mode = x_291_mode_0, pad = x_291_pad_0, x = x_289_cast_fp16)[name = tensor("x_291_cast_fp16")]; + tensor var_2536 = const()[name = tensor("op_2536"), val = tensor([1, 8, -1, 16])]; + tensor x_293_cast_fp16 = reshape(shape = var_2536, x = x_291_cast_fp16)[name = tensor("x_293_cast_fp16")]; + tensor var_2540_begin_0 = const()[name = tensor("op_2540_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2540_end_0 = const()[name = tensor("op_2540_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_2540_end_mask_0 = const()[name = tensor("op_2540_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2540_cast_fp16 = slice_by_index(begin = var_2540_begin_0, end = var_2540_end_0, end_mask = var_2540_end_mask_0, x = x_293_cast_fp16)[name = tensor("op_2540_cast_fp16")]; + tensor var_2541 = const()[name = tensor("op_2541"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2541, x = var_2540_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; + tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; + tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_136")]; + tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2524_cast_fp16)[name = tensor("transpose_137")]; + tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; + tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; + tensor var_2550_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2550_cast_fp16")]; + tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_45_cast_fp16 = mul(x = var_2550_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; + tensor scores_47_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_3)[name = tensor("scores_47_cast_fp16")]; + tensor var_2556_cast_fp16 = softmax(axis = var_30, x = scores_47_cast_fp16)[name = tensor("op_2556_cast_fp16")]; + tensor input_589_cast_fp16 = select(a = var_23_to_fp16, b = var_2556_cast_fp16, cond = mask_3)[name = tensor("input_589_cast_fp16")]; + tensor x_295_transpose_x_0 = const()[name = tensor("x_295_transpose_x_0"), val = tensor(false)]; + tensor x_295_transpose_y_0 = const()[name = tensor("x_295_transpose_y_0"), val = tensor(false)]; + tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_139")]; + tensor x_295_cast_fp16 = matmul(transpose_x = x_295_transpose_x_0, transpose_y = x_295_transpose_y_0, x = input_589_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_295_cast_fp16")]; + tensor var_2560_perm_0 = const()[name = tensor("op_2560_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2561 = const()[name = tensor("op_2561"), val = tensor([1, -1, 512])]; + tensor var_2560_cast_fp16 = transpose(perm = var_2560_perm_0, x = x_295_cast_fp16)[name = tensor("transpose_135")]; + tensor input_591_cast_fp16 = reshape(shape = var_2561, x = var_2560_cast_fp16)[name = tensor("input_591_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140803136)))]; + tensor linear_105_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_out_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("linear_105_cast_fp16")]; + tensor input_595_cast_fp16 = add(x = input_585_cast_fp16, y = linear_105_cast_fp16)[name = tensor("input_595_cast_fp16")]; + tensor x_299_axes_0 = const()[name = tensor("x_299_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141327488)))]; + tensor encoder_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141328576)))]; + tensor x_299_cast_fp16 = layer_norm(axes = x_299_axes_0, beta = encoder_layers_11_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_11_norm_conv_weight_to_fp16, x = input_595_cast_fp16)[name = tensor("x_299_cast_fp16")]; + tensor input_597_perm_0 = const()[name = tensor("input_597_perm_0"), val = tensor([0, 2, 1])]; + tensor input_599_pad_type_0 = const()[name = tensor("input_599_pad_type_0"), val = tensor("valid")]; + tensor input_599_strides_0 = const()[name = tensor("input_599_strides_0"), val = tensor([1])]; + tensor input_599_pad_0 = const()[name = tensor("input_599_pad_0"), val = tensor([0, 0])]; + tensor input_599_dilations_0 = const()[name = tensor("input_599_dilations_0"), val = tensor([1])]; + tensor input_599_groups_0 = const()[name = tensor("input_599_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141329664)))]; + tensor input_597_cast_fp16 = transpose(perm = input_597_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_134")]; + tensor input_599_cast_fp16 = conv(dilations = input_599_dilations_0, groups = input_599_groups_0, pad = input_599_pad_0, pad_type = input_599_pad_type_0, strides = input_599_strides_0, weight = encoder_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("input_599_cast_fp16")]; + tensor x_301_split_num_splits_0 = const()[name = tensor("x_301_split_num_splits_0"), val = tensor(2)]; + tensor x_301_split_axis_0 = const()[name = tensor("x_301_split_axis_0"), val = tensor(1)]; + tensor x_301_split_cast_fp16_0, tensor x_301_split_cast_fp16_1 = split(axis = x_301_split_axis_0, num_splits = x_301_split_num_splits_0, x = input_599_cast_fp16)[name = tensor("x_301_split_cast_fp16")]; + tensor x_301_split_1_sigmoid_cast_fp16 = sigmoid(x = x_301_split_cast_fp16_1)[name = tensor("x_301_split_1_sigmoid_cast_fp16")]; + tensor x_301_cast_fp16 = mul(x = x_301_split_cast_fp16_0, y = x_301_split_1_sigmoid_cast_fp16)[name = tensor("x_301_cast_fp16")]; + tensor input_601_cast_fp16 = select(a = var_23_to_fp16, b = x_301_cast_fp16, cond = var_309)[name = tensor("input_601_cast_fp16")]; + tensor new_x_47_interleave_0 = const()[name = tensor("new_x_47_interleave_0"), val = tensor(false)]; + tensor new_x_47_cast_fp16 = concat(axis = var_30, interleave = new_x_47_interleave_0, values = (cache_47_cast_fp16, input_601_cast_fp16))[name = tensor("new_x_47_cast_fp16")]; + tensor next_cache_23_begin_0 = const()[name = tensor("next_cache_23_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_23_end_0 = const()[name = tensor("next_cache_23_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_23_end_mask_0 = const()[name = tensor("next_cache_23_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_23_cast_fp16 = slice_by_index(begin = next_cache_23_begin_0, end = next_cache_23_end_0, end_mask = next_cache_23_end_mask_0, x = new_x_47_cast_fp16)[name = tensor("next_cache_23_cast_fp16")]; + tensor var_2602_begin_0 = const()[name = tensor("op_2602_begin_0"), val = tensor([0, 0, 14])]; + tensor var_2602_end_0 = const()[name = tensor("op_2602_end_0"), val = tensor([1, 512, 22])]; + tensor var_2602_end_mask_0 = const()[name = tensor("op_2602_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2602_cast_fp16 = slice_by_index(begin = var_2602_begin_0, end = var_2602_end_0, end_mask = var_2602_end_mask_0, x = next_cache_23_cast_fp16)[name = tensor("op_2602_cast_fp16")]; + tensor x_303_pad_type_0 = const()[name = tensor("x_303_pad_type_0"), val = tensor("valid")]; + tensor x_303_groups_0 = const()[name = tensor("x_303_groups_0"), val = tensor(512)]; + tensor x_303_strides_0 = const()[name = tensor("x_303_strides_0"), val = tensor([1])]; + tensor x_303_pad_0 = const()[name = tensor("x_303_pad_0"), val = tensor([0, 0])]; + tensor x_303_dilations_0 = const()[name = tensor("x_303_dilations_0"), val = tensor([1])]; + tensor encoder_layers_11_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142378304)))]; + tensor x_303_cast_fp16 = conv(dilations = x_303_dilations_0, groups = x_303_groups_0, pad = x_303_pad_0, pad_type = x_303_pad_type_0, strides = x_303_strides_0, weight = encoder_layers_11_conv_depthwise_conv_weight_to_fp16, x = new_x_47_cast_fp16)[name = tensor("x_303_cast_fp16")]; + tensor input_603_perm_0 = const()[name = tensor("input_603_perm_0"), val = tensor([0, 2, 1])]; + tensor x_305_axes_0 = const()[name = tensor("x_305_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142387584)))]; + tensor encoder_layers_11_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142388672)))]; + tensor input_603_cast_fp16 = transpose(perm = input_603_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_133")]; + tensor x_305_cast_fp16 = layer_norm(axes = x_305_axes_0, beta = encoder_layers_11_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_11_conv_batch_norm_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("x_305_cast_fp16")]; + tensor input_605_perm_0 = const()[name = tensor("input_605_perm_0"), val = tensor([0, 2, 1])]; + tensor input_605_cast_fp16 = transpose(perm = input_605_perm_0, x = x_305_cast_fp16)[name = tensor("transpose_132")]; + tensor input_607_cast_fp16 = silu(x = input_605_cast_fp16)[name = tensor("input_607_cast_fp16")]; + tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; + tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; + tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; + tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; + tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142389760)))]; + tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = encoder_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_607_cast_fp16)[name = tensor("x_307_cast_fp16")]; + tensor input_609_perm_0 = const()[name = tensor("input_609_perm_0"), val = tensor([0, 2, 1])]; + tensor input_609_cast_fp16 = transpose(perm = input_609_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_131")]; + tensor input_611_cast_fp16 = add(x = input_595_cast_fp16, y = input_609_cast_fp16)[name = tensor("input_611_cast_fp16")]; + tensor input_613_axes_0 = const()[name = tensor("input_613_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142914112)))]; + tensor encoder_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142915200)))]; + tensor input_613_cast_fp16 = layer_norm(axes = input_613_axes_0, beta = encoder_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_11_norm_feed_forward2_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("input_613_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142916288)))]; + tensor linear_106_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_613_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_617_cast_fp16 = silu(x = linear_106_cast_fp16)[name = tensor("input_617_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145013504)))]; + tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_617_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor var_2643_to_fp16 = const()[name = tensor("op_2643_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2644_cast_fp16 = mul(x = linear_107_cast_fp16, y = var_2643_to_fp16)[name = tensor("op_2644_cast_fp16")]; + tensor input_623_cast_fp16 = add(x = input_611_cast_fp16, y = var_2644_cast_fp16)[name = tensor("input_623_cast_fp16")]; + tensor input_625_axes_0 = const()[name = tensor("input_625_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147110720)))]; + tensor encoder_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147111808)))]; + tensor input_625_cast_fp16 = layer_norm(axes = input_625_axes_0, beta = encoder_layers_11_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_11_norm_out_weight_to_fp16, x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; + tensor cache_49_begin_0 = const()[name = tensor("cache_49_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_49_end_0 = const()[name = tensor("cache_49_end_0"), val = tensor([13, 1, 70, 512])]; + tensor cache_49_end_mask_0 = const()[name = tensor("cache_49_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_49_squeeze_mask_0 = const()[name = tensor("cache_49_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_49_cast_fp16 = slice_by_index(begin = cache_49_begin_0, end = cache_49_end_0, end_mask = cache_49_end_mask_0, squeeze_mask = cache_49_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_49_cast_fp16")]; + tensor cache_51_begin_0 = const()[name = tensor("cache_51_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_51_end_0 = const()[name = tensor("cache_51_end_0"), val = tensor([13, 1, 512, 8])]; + tensor cache_51_end_mask_0 = const()[name = tensor("cache_51_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_51_squeeze_mask_0 = const()[name = tensor("cache_51_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_51_cast_fp16 = slice_by_index(begin = cache_51_begin_0, end = cache_51_end_0, end_mask = cache_51_end_mask_0, squeeze_mask = cache_51_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_51_cast_fp16")]; + tensor input_627_axes_0 = const()[name = tensor("input_627_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147112896)))]; + tensor encoder_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147113984)))]; + tensor input_627_cast_fp16 = layer_norm(axes = input_627_axes_0, beta = encoder_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_12_norm_feed_forward1_weight_to_fp16, x = input_625_cast_fp16)[name = tensor("input_627_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147115072)))]; + tensor linear_108_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_627_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor input_631_cast_fp16 = silu(x = linear_108_cast_fp16)[name = tensor("input_631_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149212288)))]; + tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor var_2678_to_fp16 = const()[name = tensor("op_2678_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2679_cast_fp16 = mul(x = linear_109_cast_fp16, y = var_2678_to_fp16)[name = tensor("op_2679_cast_fp16")]; + tensor input_637_cast_fp16 = add(x = input_625_cast_fp16, y = var_2679_cast_fp16)[name = tensor("input_637_cast_fp16")]; + tensor key_25_axes_0 = const()[name = tensor("key_25_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151309504)))]; + tensor encoder_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151310592)))]; + tensor key_25_cast_fp16 = layer_norm(axes = key_25_axes_0, beta = encoder_layers_12_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_12_norm_self_att_weight_to_fp16, x = input_637_cast_fp16)[name = tensor("key_25_cast_fp16")]; + tensor input_639_interleave_0 = const()[name = tensor("input_639_interleave_0"), val = tensor(false)]; + tensor input_639_cast_fp16 = concat(axis = var_44, interleave = input_639_interleave_0, values = (cache_49_cast_fp16, key_25_cast_fp16))[name = tensor("input_639_cast_fp16")]; + tensor var_2701_begin_0 = const()[name = tensor("op_2701_begin_0"), val = tensor([0, 14, 0])]; + tensor var_2701_end_0 = const()[name = tensor("op_2701_end_0"), val = tensor([1, 70, 512])]; + tensor var_2701_end_mask_0 = const()[name = tensor("op_2701_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2701_cast_fp16 = slice_by_index(begin = var_2701_begin_0, end = var_2701_end_0, end_mask = var_2701_end_mask_0, x = cache_49_cast_fp16)[name = tensor("op_2701_cast_fp16")]; + tensor var_2704_begin_0 = const()[name = tensor("op_2704_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2704_end_0 = const()[name = tensor("op_2704_end_0"), val = tensor([1, 14, 512])]; + tensor var_2704_end_mask_0 = const()[name = tensor("op_2704_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2704_cast_fp16 = slice_by_index(begin = var_2704_begin_0, end = var_2704_end_0, end_mask = var_2704_end_mask_0, x = key_25_cast_fp16)[name = tensor("op_2704_cast_fp16")]; + tensor var_2707_interleave_0 = const()[name = tensor("op_2707_interleave_0"), val = tensor(false)]; + tensor var_2707_cast_fp16 = concat(axis = var_44, interleave = var_2707_interleave_0, values = (var_2701_cast_fp16, var_2704_cast_fp16))[name = tensor("op_2707_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151311680)))]; + tensor linear_110_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_q_weight_to_fp16, x = key_25_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_2711 = const()[name = tensor("op_2711"), val = tensor([1, -1, 8, 64])]; + tensor q_73_cast_fp16 = reshape(shape = var_2711, x = linear_110_cast_fp16)[name = tensor("q_73_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151836032)))]; + tensor linear_111_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_k_weight_to_fp16, x = input_639_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor var_2715 = const()[name = tensor("op_2715"), val = tensor([1, -1, 8, 64])]; + tensor k_49_cast_fp16 = reshape(shape = var_2715, x = linear_111_cast_fp16)[name = tensor("k_49_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152360384)))]; + tensor linear_112_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_v_weight_to_fp16, x = input_639_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor var_2719 = const()[name = tensor("op_2719"), val = tensor([1, -1, 8, 64])]; + tensor v_25_cast_fp16 = reshape(shape = var_2719, x = linear_112_cast_fp16)[name = tensor("v_25_cast_fp16")]; + tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152884736)))]; + tensor var_2731_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2731_cast_fp16")]; + tensor encoder_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152885824)))]; + tensor var_2733_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2733_cast_fp16")]; + tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; + tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; + tensor var_2735_to_fp16 = const()[name = tensor("op_2735_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152886912)))]; + tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2733_cast_fp16)[name = tensor("transpose_129")]; + tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2735_to_fp16)[name = tensor("x_315_cast_fp16")]; + tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; + tensor const_174_to_fp16 = const()[name = tensor("const_174_to_fp16"), val = tensor(0x0p+0)]; + tensor x_317_cast_fp16 = pad(constant_val = const_174_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; + tensor var_2743 = const()[name = tensor("op_2743"), val = tensor([1, 8, -1, 16])]; + tensor x_319_cast_fp16 = reshape(shape = var_2743, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; + tensor var_2747_begin_0 = const()[name = tensor("op_2747_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2747_end_0 = const()[name = tensor("op_2747_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_2747_end_mask_0 = const()[name = tensor("op_2747_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2747_cast_fp16 = slice_by_index(begin = var_2747_begin_0, end = var_2747_end_0, end_mask = var_2747_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2747_cast_fp16")]; + tensor var_2748 = const()[name = tensor("op_2748"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2748, x = var_2747_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; + tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; + tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_127")]; + tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2731_cast_fp16)[name = tensor("transpose_128")]; + tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; + tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; + tensor var_2757_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2757_cast_fp16")]; + tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_49_cast_fp16 = mul(x = var_2757_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; + tensor scores_51_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_3)[name = tensor("scores_51_cast_fp16")]; + tensor var_2763_cast_fp16 = softmax(axis = var_30, x = scores_51_cast_fp16)[name = tensor("op_2763_cast_fp16")]; + tensor input_641_cast_fp16 = select(a = var_23_to_fp16, b = var_2763_cast_fp16, cond = mask_3)[name = tensor("input_641_cast_fp16")]; + tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; + tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; + tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_130")]; + tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_641_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_321_cast_fp16")]; + tensor var_2767_perm_0 = const()[name = tensor("op_2767_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2768 = const()[name = tensor("op_2768"), val = tensor([1, -1, 512])]; + tensor var_2767_cast_fp16 = transpose(perm = var_2767_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_126")]; + tensor input_643_cast_fp16 = reshape(shape = var_2768, x = var_2767_cast_fp16)[name = tensor("input_643_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153062080)))]; + tensor linear_114_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_out_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("linear_114_cast_fp16")]; + tensor input_647_cast_fp16 = add(x = input_637_cast_fp16, y = linear_114_cast_fp16)[name = tensor("input_647_cast_fp16")]; + tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153586432)))]; + tensor encoder_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153587520)))]; + tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = encoder_layers_12_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_12_norm_conv_weight_to_fp16, x = input_647_cast_fp16)[name = tensor("x_325_cast_fp16")]; + tensor input_649_perm_0 = const()[name = tensor("input_649_perm_0"), val = tensor([0, 2, 1])]; + tensor input_651_pad_type_0 = const()[name = tensor("input_651_pad_type_0"), val = tensor("valid")]; + tensor input_651_strides_0 = const()[name = tensor("input_651_strides_0"), val = tensor([1])]; + tensor input_651_pad_0 = const()[name = tensor("input_651_pad_0"), val = tensor([0, 0])]; + tensor input_651_dilations_0 = const()[name = tensor("input_651_dilations_0"), val = tensor([1])]; + tensor input_651_groups_0 = const()[name = tensor("input_651_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153588608)))]; + tensor input_649_cast_fp16 = transpose(perm = input_649_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_125")]; + tensor input_651_cast_fp16 = conv(dilations = input_651_dilations_0, groups = input_651_groups_0, pad = input_651_pad_0, pad_type = input_651_pad_type_0, strides = input_651_strides_0, weight = encoder_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("input_651_cast_fp16")]; + tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; + tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; + tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_651_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; + tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; + tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; + tensor input_653_cast_fp16 = select(a = var_23_to_fp16, b = x_327_cast_fp16, cond = var_309)[name = tensor("input_653_cast_fp16")]; + tensor new_x_51_interleave_0 = const()[name = tensor("new_x_51_interleave_0"), val = tensor(false)]; + tensor new_x_51_cast_fp16 = concat(axis = var_30, interleave = new_x_51_interleave_0, values = (cache_51_cast_fp16, input_653_cast_fp16))[name = tensor("new_x_51_cast_fp16")]; + tensor next_cache_25_begin_0 = const()[name = tensor("next_cache_25_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_25_end_0 = const()[name = tensor("next_cache_25_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_25_end_mask_0 = const()[name = tensor("next_cache_25_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_25_cast_fp16 = slice_by_index(begin = next_cache_25_begin_0, end = next_cache_25_end_0, end_mask = next_cache_25_end_mask_0, x = new_x_51_cast_fp16)[name = tensor("next_cache_25_cast_fp16")]; + tensor var_2809_begin_0 = const()[name = tensor("op_2809_begin_0"), val = tensor([0, 0, 14])]; + tensor var_2809_end_0 = const()[name = tensor("op_2809_end_0"), val = tensor([1, 512, 22])]; + tensor var_2809_end_mask_0 = const()[name = tensor("op_2809_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2809_cast_fp16 = slice_by_index(begin = var_2809_begin_0, end = var_2809_end_0, end_mask = var_2809_end_mask_0, x = next_cache_25_cast_fp16)[name = tensor("op_2809_cast_fp16")]; + tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; + tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(512)]; + tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; + tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; + tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; + tensor encoder_layers_12_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154637248)))]; + tensor x_329_cast_fp16 = conv(dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = encoder_layers_12_conv_depthwise_conv_weight_to_fp16, x = new_x_51_cast_fp16)[name = tensor("x_329_cast_fp16")]; + tensor input_655_perm_0 = const()[name = tensor("input_655_perm_0"), val = tensor([0, 2, 1])]; + tensor x_331_axes_0 = const()[name = tensor("x_331_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154646528)))]; + tensor encoder_layers_12_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154647616)))]; + tensor input_655_cast_fp16 = transpose(perm = input_655_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_124")]; + tensor x_331_cast_fp16 = layer_norm(axes = x_331_axes_0, beta = encoder_layers_12_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_12_conv_batch_norm_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("x_331_cast_fp16")]; + tensor input_657_perm_0 = const()[name = tensor("input_657_perm_0"), val = tensor([0, 2, 1])]; + tensor input_657_cast_fp16 = transpose(perm = input_657_perm_0, x = x_331_cast_fp16)[name = tensor("transpose_123")]; + tensor input_659_cast_fp16 = silu(x = input_657_cast_fp16)[name = tensor("input_659_cast_fp16")]; + tensor x_333_pad_type_0 = const()[name = tensor("x_333_pad_type_0"), val = tensor("valid")]; + tensor x_333_strides_0 = const()[name = tensor("x_333_strides_0"), val = tensor([1])]; + tensor x_333_pad_0 = const()[name = tensor("x_333_pad_0"), val = tensor([0, 0])]; + tensor x_333_dilations_0 = const()[name = tensor("x_333_dilations_0"), val = tensor([1])]; + tensor x_333_groups_0 = const()[name = tensor("x_333_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154648704)))]; + tensor x_333_cast_fp16 = conv(dilations = x_333_dilations_0, groups = x_333_groups_0, pad = x_333_pad_0, pad_type = x_333_pad_type_0, strides = x_333_strides_0, weight = encoder_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_659_cast_fp16)[name = tensor("x_333_cast_fp16")]; + tensor input_661_perm_0 = const()[name = tensor("input_661_perm_0"), val = tensor([0, 2, 1])]; + tensor input_661_cast_fp16 = transpose(perm = input_661_perm_0, x = x_333_cast_fp16)[name = tensor("transpose_122")]; + tensor input_663_cast_fp16 = add(x = input_647_cast_fp16, y = input_661_cast_fp16)[name = tensor("input_663_cast_fp16")]; + tensor input_665_axes_0 = const()[name = tensor("input_665_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155173056)))]; + tensor encoder_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155174144)))]; + tensor input_665_cast_fp16 = layer_norm(axes = input_665_axes_0, beta = encoder_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_12_norm_feed_forward2_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("input_665_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155175232)))]; + tensor linear_115_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_665_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor input_669_cast_fp16 = silu(x = linear_115_cast_fp16)[name = tensor("input_669_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157272448)))]; + tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_669_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor var_2850_to_fp16 = const()[name = tensor("op_2850_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2851_cast_fp16 = mul(x = linear_116_cast_fp16, y = var_2850_to_fp16)[name = tensor("op_2851_cast_fp16")]; + tensor input_675_cast_fp16 = add(x = input_663_cast_fp16, y = var_2851_cast_fp16)[name = tensor("input_675_cast_fp16")]; + tensor input_677_axes_0 = const()[name = tensor("input_677_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159369664)))]; + tensor encoder_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159370752)))]; + tensor input_677_cast_fp16 = layer_norm(axes = input_677_axes_0, beta = encoder_layers_12_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_12_norm_out_weight_to_fp16, x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; + tensor cache_53_begin_0 = const()[name = tensor("cache_53_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_53_end_0 = const()[name = tensor("cache_53_end_0"), val = tensor([14, 1, 70, 512])]; + tensor cache_53_end_mask_0 = const()[name = tensor("cache_53_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_53_squeeze_mask_0 = const()[name = tensor("cache_53_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_53_cast_fp16 = slice_by_index(begin = cache_53_begin_0, end = cache_53_end_0, end_mask = cache_53_end_mask_0, squeeze_mask = cache_53_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_53_cast_fp16")]; + tensor cache_55_begin_0 = const()[name = tensor("cache_55_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_55_end_0 = const()[name = tensor("cache_55_end_0"), val = tensor([14, 1, 512, 8])]; + tensor cache_55_end_mask_0 = const()[name = tensor("cache_55_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_55_squeeze_mask_0 = const()[name = tensor("cache_55_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_55_cast_fp16 = slice_by_index(begin = cache_55_begin_0, end = cache_55_end_0, end_mask = cache_55_end_mask_0, squeeze_mask = cache_55_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_55_cast_fp16")]; + tensor input_679_axes_0 = const()[name = tensor("input_679_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159371840)))]; + tensor encoder_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159372928)))]; + tensor input_679_cast_fp16 = layer_norm(axes = input_679_axes_0, beta = encoder_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_13_norm_feed_forward1_weight_to_fp16, x = input_677_cast_fp16)[name = tensor("input_679_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159374016)))]; + tensor linear_117_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_679_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor input_683_cast_fp16 = silu(x = linear_117_cast_fp16)[name = tensor("input_683_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161471232)))]; + tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor var_2885_to_fp16 = const()[name = tensor("op_2885_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2886_cast_fp16 = mul(x = linear_118_cast_fp16, y = var_2885_to_fp16)[name = tensor("op_2886_cast_fp16")]; + tensor input_689_cast_fp16 = add(x = input_677_cast_fp16, y = var_2886_cast_fp16)[name = tensor("input_689_cast_fp16")]; + tensor key_27_axes_0 = const()[name = tensor("key_27_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163568448)))]; + tensor encoder_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163569536)))]; + tensor key_27_cast_fp16 = layer_norm(axes = key_27_axes_0, beta = encoder_layers_13_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_13_norm_self_att_weight_to_fp16, x = input_689_cast_fp16)[name = tensor("key_27_cast_fp16")]; + tensor input_691_interleave_0 = const()[name = tensor("input_691_interleave_0"), val = tensor(false)]; + tensor input_691_cast_fp16 = concat(axis = var_44, interleave = input_691_interleave_0, values = (cache_53_cast_fp16, key_27_cast_fp16))[name = tensor("input_691_cast_fp16")]; + tensor var_2908_begin_0 = const()[name = tensor("op_2908_begin_0"), val = tensor([0, 14, 0])]; + tensor var_2908_end_0 = const()[name = tensor("op_2908_end_0"), val = tensor([1, 70, 512])]; + tensor var_2908_end_mask_0 = const()[name = tensor("op_2908_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2908_cast_fp16 = slice_by_index(begin = var_2908_begin_0, end = var_2908_end_0, end_mask = var_2908_end_mask_0, x = cache_53_cast_fp16)[name = tensor("op_2908_cast_fp16")]; + tensor var_2911_begin_0 = const()[name = tensor("op_2911_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2911_end_0 = const()[name = tensor("op_2911_end_0"), val = tensor([1, 14, 512])]; + tensor var_2911_end_mask_0 = const()[name = tensor("op_2911_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2911_cast_fp16 = slice_by_index(begin = var_2911_begin_0, end = var_2911_end_0, end_mask = var_2911_end_mask_0, x = key_27_cast_fp16)[name = tensor("op_2911_cast_fp16")]; + tensor var_2914_interleave_0 = const()[name = tensor("op_2914_interleave_0"), val = tensor(false)]; + tensor var_2914_cast_fp16 = concat(axis = var_44, interleave = var_2914_interleave_0, values = (var_2908_cast_fp16, var_2911_cast_fp16))[name = tensor("op_2914_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163570624)))]; + tensor linear_119_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_q_weight_to_fp16, x = key_27_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor var_2918 = const()[name = tensor("op_2918"), val = tensor([1, -1, 8, 64])]; + tensor q_79_cast_fp16 = reshape(shape = var_2918, x = linear_119_cast_fp16)[name = tensor("q_79_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164094976)))]; + tensor linear_120_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_k_weight_to_fp16, x = input_691_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_2922 = const()[name = tensor("op_2922"), val = tensor([1, -1, 8, 64])]; + tensor k_53_cast_fp16 = reshape(shape = var_2922, x = linear_120_cast_fp16)[name = tensor("k_53_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164619328)))]; + tensor linear_121_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_v_weight_to_fp16, x = input_691_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_2926 = const()[name = tensor("op_2926"), val = tensor([1, -1, 8, 64])]; + tensor v_27_cast_fp16 = reshape(shape = var_2926, x = linear_121_cast_fp16)[name = tensor("v_27_cast_fp16")]; + tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165143680)))]; + tensor var_2938_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2938_cast_fp16")]; + tensor encoder_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165144768)))]; + tensor var_2940_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2940_cast_fp16")]; + tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_341_transpose_x_0 = const()[name = tensor("x_341_transpose_x_0"), val = tensor(false)]; + tensor x_341_transpose_y_0 = const()[name = tensor("x_341_transpose_y_0"), val = tensor(false)]; + tensor var_2942_to_fp16 = const()[name = tensor("op_2942_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165145856)))]; + tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2940_cast_fp16)[name = tensor("transpose_120")]; + tensor x_341_cast_fp16 = matmul(transpose_x = x_341_transpose_x_0, transpose_y = x_341_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2942_to_fp16)[name = tensor("x_341_cast_fp16")]; + tensor x_343_pad_0 = const()[name = tensor("x_343_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_343_mode_0 = const()[name = tensor("x_343_mode_0"), val = tensor("constant")]; + tensor const_187_to_fp16 = const()[name = tensor("const_187_to_fp16"), val = tensor(0x0p+0)]; + tensor x_343_cast_fp16 = pad(constant_val = const_187_to_fp16, mode = x_343_mode_0, pad = x_343_pad_0, x = x_341_cast_fp16)[name = tensor("x_343_cast_fp16")]; + tensor var_2950 = const()[name = tensor("op_2950"), val = tensor([1, 8, -1, 16])]; + tensor x_345_cast_fp16 = reshape(shape = var_2950, x = x_343_cast_fp16)[name = tensor("x_345_cast_fp16")]; + tensor var_2954_begin_0 = const()[name = tensor("op_2954_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2954_end_0 = const()[name = tensor("op_2954_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_2954_end_mask_0 = const()[name = tensor("op_2954_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2954_cast_fp16 = slice_by_index(begin = var_2954_begin_0, end = var_2954_end_0, end_mask = var_2954_end_mask_0, x = x_345_cast_fp16)[name = tensor("op_2954_cast_fp16")]; + tensor var_2955 = const()[name = tensor("op_2955"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2955, x = var_2954_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; + tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; + tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_118")]; + tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_2938_cast_fp16)[name = tensor("transpose_119")]; + tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; + tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; + tensor var_2964_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2964_cast_fp16")]; + tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_53_cast_fp16 = mul(x = var_2964_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; + tensor scores_55_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_3)[name = tensor("scores_55_cast_fp16")]; + tensor var_2970_cast_fp16 = softmax(axis = var_30, x = scores_55_cast_fp16)[name = tensor("op_2970_cast_fp16")]; + tensor input_693_cast_fp16 = select(a = var_23_to_fp16, b = var_2970_cast_fp16, cond = mask_3)[name = tensor("input_693_cast_fp16")]; + tensor x_347_transpose_x_0 = const()[name = tensor("x_347_transpose_x_0"), val = tensor(false)]; + tensor x_347_transpose_y_0 = const()[name = tensor("x_347_transpose_y_0"), val = tensor(false)]; + tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_121")]; + tensor x_347_cast_fp16 = matmul(transpose_x = x_347_transpose_x_0, transpose_y = x_347_transpose_y_0, x = input_693_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_347_cast_fp16")]; + tensor var_2974_perm_0 = const()[name = tensor("op_2974_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2975 = const()[name = tensor("op_2975"), val = tensor([1, -1, 512])]; + tensor var_2974_cast_fp16 = transpose(perm = var_2974_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_117")]; + tensor input_695_cast_fp16 = reshape(shape = var_2975, x = var_2974_cast_fp16)[name = tensor("input_695_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165321024)))]; + tensor linear_123_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_out_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("linear_123_cast_fp16")]; + tensor input_699_cast_fp16 = add(x = input_689_cast_fp16, y = linear_123_cast_fp16)[name = tensor("input_699_cast_fp16")]; + tensor x_351_axes_0 = const()[name = tensor("x_351_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165845376)))]; + tensor encoder_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165846464)))]; + tensor x_351_cast_fp16 = layer_norm(axes = x_351_axes_0, beta = encoder_layers_13_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_13_norm_conv_weight_to_fp16, x = input_699_cast_fp16)[name = tensor("x_351_cast_fp16")]; + tensor input_701_perm_0 = const()[name = tensor("input_701_perm_0"), val = tensor([0, 2, 1])]; + tensor input_703_pad_type_0 = const()[name = tensor("input_703_pad_type_0"), val = tensor("valid")]; + tensor input_703_strides_0 = const()[name = tensor("input_703_strides_0"), val = tensor([1])]; + tensor input_703_pad_0 = const()[name = tensor("input_703_pad_0"), val = tensor([0, 0])]; + tensor input_703_dilations_0 = const()[name = tensor("input_703_dilations_0"), val = tensor([1])]; + tensor input_703_groups_0 = const()[name = tensor("input_703_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165847552)))]; + tensor input_701_cast_fp16 = transpose(perm = input_701_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_116")]; + tensor input_703_cast_fp16 = conv(dilations = input_703_dilations_0, groups = input_703_groups_0, pad = input_703_pad_0, pad_type = input_703_pad_type_0, strides = input_703_strides_0, weight = encoder_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("input_703_cast_fp16")]; + tensor x_353_split_num_splits_0 = const()[name = tensor("x_353_split_num_splits_0"), val = tensor(2)]; + tensor x_353_split_axis_0 = const()[name = tensor("x_353_split_axis_0"), val = tensor(1)]; + tensor x_353_split_cast_fp16_0, tensor x_353_split_cast_fp16_1 = split(axis = x_353_split_axis_0, num_splits = x_353_split_num_splits_0, x = input_703_cast_fp16)[name = tensor("x_353_split_cast_fp16")]; + tensor x_353_split_1_sigmoid_cast_fp16 = sigmoid(x = x_353_split_cast_fp16_1)[name = tensor("x_353_split_1_sigmoid_cast_fp16")]; + tensor x_353_cast_fp16 = mul(x = x_353_split_cast_fp16_0, y = x_353_split_1_sigmoid_cast_fp16)[name = tensor("x_353_cast_fp16")]; + tensor input_705_cast_fp16 = select(a = var_23_to_fp16, b = x_353_cast_fp16, cond = var_309)[name = tensor("input_705_cast_fp16")]; + tensor new_x_55_interleave_0 = const()[name = tensor("new_x_55_interleave_0"), val = tensor(false)]; + tensor new_x_55_cast_fp16 = concat(axis = var_30, interleave = new_x_55_interleave_0, values = (cache_55_cast_fp16, input_705_cast_fp16))[name = tensor("new_x_55_cast_fp16")]; + tensor next_cache_27_begin_0 = const()[name = tensor("next_cache_27_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_27_end_0 = const()[name = tensor("next_cache_27_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_27_end_mask_0 = const()[name = tensor("next_cache_27_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_27_cast_fp16 = slice_by_index(begin = next_cache_27_begin_0, end = next_cache_27_end_0, end_mask = next_cache_27_end_mask_0, x = new_x_55_cast_fp16)[name = tensor("next_cache_27_cast_fp16")]; + tensor var_3016_begin_0 = const()[name = tensor("op_3016_begin_0"), val = tensor([0, 0, 14])]; + tensor var_3016_end_0 = const()[name = tensor("op_3016_end_0"), val = tensor([1, 512, 22])]; + tensor var_3016_end_mask_0 = const()[name = tensor("op_3016_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3016_cast_fp16 = slice_by_index(begin = var_3016_begin_0, end = var_3016_end_0, end_mask = var_3016_end_mask_0, x = next_cache_27_cast_fp16)[name = tensor("op_3016_cast_fp16")]; + tensor x_355_pad_type_0 = const()[name = tensor("x_355_pad_type_0"), val = tensor("valid")]; + tensor x_355_groups_0 = const()[name = tensor("x_355_groups_0"), val = tensor(512)]; + tensor x_355_strides_0 = const()[name = tensor("x_355_strides_0"), val = tensor([1])]; + tensor x_355_pad_0 = const()[name = tensor("x_355_pad_0"), val = tensor([0, 0])]; + tensor x_355_dilations_0 = const()[name = tensor("x_355_dilations_0"), val = tensor([1])]; + tensor encoder_layers_13_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166896192)))]; + tensor x_355_cast_fp16 = conv(dilations = x_355_dilations_0, groups = x_355_groups_0, pad = x_355_pad_0, pad_type = x_355_pad_type_0, strides = x_355_strides_0, weight = encoder_layers_13_conv_depthwise_conv_weight_to_fp16, x = new_x_55_cast_fp16)[name = tensor("x_355_cast_fp16")]; + tensor input_707_perm_0 = const()[name = tensor("input_707_perm_0"), val = tensor([0, 2, 1])]; + tensor x_357_axes_0 = const()[name = tensor("x_357_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166905472)))]; + tensor encoder_layers_13_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166906560)))]; + tensor input_707_cast_fp16 = transpose(perm = input_707_perm_0, x = x_355_cast_fp16)[name = tensor("transpose_115")]; + tensor x_357_cast_fp16 = layer_norm(axes = x_357_axes_0, beta = encoder_layers_13_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_13_conv_batch_norm_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("x_357_cast_fp16")]; + tensor input_709_perm_0 = const()[name = tensor("input_709_perm_0"), val = tensor([0, 2, 1])]; + tensor input_709_cast_fp16 = transpose(perm = input_709_perm_0, x = x_357_cast_fp16)[name = tensor("transpose_114")]; + tensor input_711_cast_fp16 = silu(x = input_709_cast_fp16)[name = tensor("input_711_cast_fp16")]; + tensor x_359_pad_type_0 = const()[name = tensor("x_359_pad_type_0"), val = tensor("valid")]; + tensor x_359_strides_0 = const()[name = tensor("x_359_strides_0"), val = tensor([1])]; + tensor x_359_pad_0 = const()[name = tensor("x_359_pad_0"), val = tensor([0, 0])]; + tensor x_359_dilations_0 = const()[name = tensor("x_359_dilations_0"), val = tensor([1])]; + tensor x_359_groups_0 = const()[name = tensor("x_359_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166907648)))]; + tensor x_359_cast_fp16 = conv(dilations = x_359_dilations_0, groups = x_359_groups_0, pad = x_359_pad_0, pad_type = x_359_pad_type_0, strides = x_359_strides_0, weight = encoder_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_711_cast_fp16)[name = tensor("x_359_cast_fp16")]; + tensor input_713_perm_0 = const()[name = tensor("input_713_perm_0"), val = tensor([0, 2, 1])]; + tensor input_713_cast_fp16 = transpose(perm = input_713_perm_0, x = x_359_cast_fp16)[name = tensor("transpose_113")]; + tensor input_715_cast_fp16 = add(x = input_699_cast_fp16, y = input_713_cast_fp16)[name = tensor("input_715_cast_fp16")]; + tensor input_717_axes_0 = const()[name = tensor("input_717_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167432000)))]; + tensor encoder_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167433088)))]; + tensor input_717_cast_fp16 = layer_norm(axes = input_717_axes_0, beta = encoder_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_13_norm_feed_forward2_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("input_717_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167434176)))]; + tensor linear_124_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_717_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_721_cast_fp16 = silu(x = linear_124_cast_fp16)[name = tensor("input_721_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169531392)))]; + tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_721_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor var_3057_to_fp16 = const()[name = tensor("op_3057_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3058_cast_fp16 = mul(x = linear_125_cast_fp16, y = var_3057_to_fp16)[name = tensor("op_3058_cast_fp16")]; + tensor input_727_cast_fp16 = add(x = input_715_cast_fp16, y = var_3058_cast_fp16)[name = tensor("input_727_cast_fp16")]; + tensor input_729_axes_0 = const()[name = tensor("input_729_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171628608)))]; + tensor encoder_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171629696)))]; + tensor input_729_cast_fp16 = layer_norm(axes = input_729_axes_0, beta = encoder_layers_13_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_13_norm_out_weight_to_fp16, x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; + tensor cache_57_begin_0 = const()[name = tensor("cache_57_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_57_end_0 = const()[name = tensor("cache_57_end_0"), val = tensor([15, 1, 70, 512])]; + tensor cache_57_end_mask_0 = const()[name = tensor("cache_57_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_57_squeeze_mask_0 = const()[name = tensor("cache_57_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_57_cast_fp16 = slice_by_index(begin = cache_57_begin_0, end = cache_57_end_0, end_mask = cache_57_end_mask_0, squeeze_mask = cache_57_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_57_cast_fp16")]; + tensor cache_59_begin_0 = const()[name = tensor("cache_59_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_59_end_0 = const()[name = tensor("cache_59_end_0"), val = tensor([15, 1, 512, 8])]; + tensor cache_59_end_mask_0 = const()[name = tensor("cache_59_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_59_squeeze_mask_0 = const()[name = tensor("cache_59_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_59_cast_fp16 = slice_by_index(begin = cache_59_begin_0, end = cache_59_end_0, end_mask = cache_59_end_mask_0, squeeze_mask = cache_59_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_59_cast_fp16")]; + tensor input_731_axes_0 = const()[name = tensor("input_731_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171630784)))]; + tensor encoder_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171631872)))]; + tensor input_731_cast_fp16 = layer_norm(axes = input_731_axes_0, beta = encoder_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_14_norm_feed_forward1_weight_to_fp16, x = input_729_cast_fp16)[name = tensor("input_731_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171632960)))]; + tensor linear_126_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_731_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor input_735_cast_fp16 = silu(x = linear_126_cast_fp16)[name = tensor("input_735_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173730176)))]; + tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor var_3092_to_fp16 = const()[name = tensor("op_3092_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3093_cast_fp16 = mul(x = linear_127_cast_fp16, y = var_3092_to_fp16)[name = tensor("op_3093_cast_fp16")]; + tensor input_741_cast_fp16 = add(x = input_729_cast_fp16, y = var_3093_cast_fp16)[name = tensor("input_741_cast_fp16")]; + tensor key_29_axes_0 = const()[name = tensor("key_29_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175827392)))]; + tensor encoder_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175828480)))]; + tensor key_29_cast_fp16 = layer_norm(axes = key_29_axes_0, beta = encoder_layers_14_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_14_norm_self_att_weight_to_fp16, x = input_741_cast_fp16)[name = tensor("key_29_cast_fp16")]; + tensor input_743_interleave_0 = const()[name = tensor("input_743_interleave_0"), val = tensor(false)]; + tensor input_743_cast_fp16 = concat(axis = var_44, interleave = input_743_interleave_0, values = (cache_57_cast_fp16, key_29_cast_fp16))[name = tensor("input_743_cast_fp16")]; + tensor var_3115_begin_0 = const()[name = tensor("op_3115_begin_0"), val = tensor([0, 14, 0])]; + tensor var_3115_end_0 = const()[name = tensor("op_3115_end_0"), val = tensor([1, 70, 512])]; + tensor var_3115_end_mask_0 = const()[name = tensor("op_3115_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3115_cast_fp16 = slice_by_index(begin = var_3115_begin_0, end = var_3115_end_0, end_mask = var_3115_end_mask_0, x = cache_57_cast_fp16)[name = tensor("op_3115_cast_fp16")]; + tensor var_3118_begin_0 = const()[name = tensor("op_3118_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3118_end_0 = const()[name = tensor("op_3118_end_0"), val = tensor([1, 14, 512])]; + tensor var_3118_end_mask_0 = const()[name = tensor("op_3118_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3118_cast_fp16 = slice_by_index(begin = var_3118_begin_0, end = var_3118_end_0, end_mask = var_3118_end_mask_0, x = key_29_cast_fp16)[name = tensor("op_3118_cast_fp16")]; + tensor var_3121_interleave_0 = const()[name = tensor("op_3121_interleave_0"), val = tensor(false)]; + tensor var_3121_cast_fp16 = concat(axis = var_44, interleave = var_3121_interleave_0, values = (var_3115_cast_fp16, var_3118_cast_fp16))[name = tensor("op_3121_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175829568)))]; + tensor linear_128_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_q_weight_to_fp16, x = key_29_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_3125 = const()[name = tensor("op_3125"), val = tensor([1, -1, 8, 64])]; + tensor q_85_cast_fp16 = reshape(shape = var_3125, x = linear_128_cast_fp16)[name = tensor("q_85_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176353920)))]; + tensor linear_129_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_k_weight_to_fp16, x = input_743_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor var_3129 = const()[name = tensor("op_3129"), val = tensor([1, -1, 8, 64])]; + tensor k_57_cast_fp16 = reshape(shape = var_3129, x = linear_129_cast_fp16)[name = tensor("k_57_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176878272)))]; + tensor linear_130_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_v_weight_to_fp16, x = input_743_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor var_3133 = const()[name = tensor("op_3133"), val = tensor([1, -1, 8, 64])]; + tensor v_29_cast_fp16 = reshape(shape = var_3133, x = linear_130_cast_fp16)[name = tensor("v_29_cast_fp16")]; + tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177402624)))]; + tensor var_3145_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3145_cast_fp16")]; + tensor encoder_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177403712)))]; + tensor var_3147_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3147_cast_fp16")]; + tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_367_transpose_x_0 = const()[name = tensor("x_367_transpose_x_0"), val = tensor(false)]; + tensor x_367_transpose_y_0 = const()[name = tensor("x_367_transpose_y_0"), val = tensor(false)]; + tensor var_3149_to_fp16 = const()[name = tensor("op_3149_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177404800)))]; + tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_3147_cast_fp16)[name = tensor("transpose_111")]; + tensor x_367_cast_fp16 = matmul(transpose_x = x_367_transpose_x_0, transpose_y = x_367_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_3149_to_fp16)[name = tensor("x_367_cast_fp16")]; + tensor x_369_pad_0 = const()[name = tensor("x_369_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_369_mode_0 = const()[name = tensor("x_369_mode_0"), val = tensor("constant")]; + tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(0x0p+0)]; + tensor x_369_cast_fp16 = pad(constant_val = const_200_to_fp16, mode = x_369_mode_0, pad = x_369_pad_0, x = x_367_cast_fp16)[name = tensor("x_369_cast_fp16")]; + tensor var_3157 = const()[name = tensor("op_3157"), val = tensor([1, 8, -1, 16])]; + tensor x_371_cast_fp16 = reshape(shape = var_3157, x = x_369_cast_fp16)[name = tensor("x_371_cast_fp16")]; + tensor var_3161_begin_0 = const()[name = tensor("op_3161_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3161_end_0 = const()[name = tensor("op_3161_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_3161_end_mask_0 = const()[name = tensor("op_3161_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3161_cast_fp16 = slice_by_index(begin = var_3161_begin_0, end = var_3161_end_0, end_mask = var_3161_end_mask_0, x = x_371_cast_fp16)[name = tensor("op_3161_cast_fp16")]; + tensor var_3162 = const()[name = tensor("op_3162"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_57_cast_fp16 = reshape(shape = var_3162, x = var_3161_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; + tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; + tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_109")]; + tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_3145_cast_fp16)[name = tensor("transpose_110")]; + tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; + tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; + tensor var_3171_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_3171_cast_fp16")]; + tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_57_cast_fp16 = mul(x = var_3171_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; + tensor scores_59_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_3)[name = tensor("scores_59_cast_fp16")]; + tensor var_3177_cast_fp16 = softmax(axis = var_30, x = scores_59_cast_fp16)[name = tensor("op_3177_cast_fp16")]; + tensor input_745_cast_fp16 = select(a = var_23_to_fp16, b = var_3177_cast_fp16, cond = mask_3)[name = tensor("input_745_cast_fp16")]; + tensor x_373_transpose_x_0 = const()[name = tensor("x_373_transpose_x_0"), val = tensor(false)]; + tensor x_373_transpose_y_0 = const()[name = tensor("x_373_transpose_y_0"), val = tensor(false)]; + tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_112")]; + tensor x_373_cast_fp16 = matmul(transpose_x = x_373_transpose_x_0, transpose_y = x_373_transpose_y_0, x = input_745_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_373_cast_fp16")]; + tensor var_3181_perm_0 = const()[name = tensor("op_3181_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3182 = const()[name = tensor("op_3182"), val = tensor([1, -1, 512])]; + tensor var_3181_cast_fp16 = transpose(perm = var_3181_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_108")]; + tensor input_747_cast_fp16 = reshape(shape = var_3182, x = var_3181_cast_fp16)[name = tensor("input_747_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177579968)))]; + tensor linear_132_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_out_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("linear_132_cast_fp16")]; + tensor input_751_cast_fp16 = add(x = input_741_cast_fp16, y = linear_132_cast_fp16)[name = tensor("input_751_cast_fp16")]; + tensor x_377_axes_0 = const()[name = tensor("x_377_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178104320)))]; + tensor encoder_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178105408)))]; + tensor x_377_cast_fp16 = layer_norm(axes = x_377_axes_0, beta = encoder_layers_14_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_14_norm_conv_weight_to_fp16, x = input_751_cast_fp16)[name = tensor("x_377_cast_fp16")]; + tensor input_753_perm_0 = const()[name = tensor("input_753_perm_0"), val = tensor([0, 2, 1])]; + tensor input_755_pad_type_0 = const()[name = tensor("input_755_pad_type_0"), val = tensor("valid")]; + tensor input_755_strides_0 = const()[name = tensor("input_755_strides_0"), val = tensor([1])]; + tensor input_755_pad_0 = const()[name = tensor("input_755_pad_0"), val = tensor([0, 0])]; + tensor input_755_dilations_0 = const()[name = tensor("input_755_dilations_0"), val = tensor([1])]; + tensor input_755_groups_0 = const()[name = tensor("input_755_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178106496)))]; + tensor input_753_cast_fp16 = transpose(perm = input_753_perm_0, x = x_377_cast_fp16)[name = tensor("transpose_107")]; + tensor input_755_cast_fp16 = conv(dilations = input_755_dilations_0, groups = input_755_groups_0, pad = input_755_pad_0, pad_type = input_755_pad_type_0, strides = input_755_strides_0, weight = encoder_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("input_755_cast_fp16")]; + tensor x_379_split_num_splits_0 = const()[name = tensor("x_379_split_num_splits_0"), val = tensor(2)]; + tensor x_379_split_axis_0 = const()[name = tensor("x_379_split_axis_0"), val = tensor(1)]; + tensor x_379_split_cast_fp16_0, tensor x_379_split_cast_fp16_1 = split(axis = x_379_split_axis_0, num_splits = x_379_split_num_splits_0, x = input_755_cast_fp16)[name = tensor("x_379_split_cast_fp16")]; + tensor x_379_split_1_sigmoid_cast_fp16 = sigmoid(x = x_379_split_cast_fp16_1)[name = tensor("x_379_split_1_sigmoid_cast_fp16")]; + tensor x_379_cast_fp16 = mul(x = x_379_split_cast_fp16_0, y = x_379_split_1_sigmoid_cast_fp16)[name = tensor("x_379_cast_fp16")]; + tensor input_757_cast_fp16 = select(a = var_23_to_fp16, b = x_379_cast_fp16, cond = var_309)[name = tensor("input_757_cast_fp16")]; + tensor new_x_59_interleave_0 = const()[name = tensor("new_x_59_interleave_0"), val = tensor(false)]; + tensor new_x_59_cast_fp16 = concat(axis = var_30, interleave = new_x_59_interleave_0, values = (cache_59_cast_fp16, input_757_cast_fp16))[name = tensor("new_x_59_cast_fp16")]; + tensor next_cache_29_begin_0 = const()[name = tensor("next_cache_29_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_29_end_0 = const()[name = tensor("next_cache_29_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_29_end_mask_0 = const()[name = tensor("next_cache_29_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_29_cast_fp16 = slice_by_index(begin = next_cache_29_begin_0, end = next_cache_29_end_0, end_mask = next_cache_29_end_mask_0, x = new_x_59_cast_fp16)[name = tensor("next_cache_29_cast_fp16")]; + tensor var_3223_begin_0 = const()[name = tensor("op_3223_begin_0"), val = tensor([0, 0, 14])]; + tensor var_3223_end_0 = const()[name = tensor("op_3223_end_0"), val = tensor([1, 512, 22])]; + tensor var_3223_end_mask_0 = const()[name = tensor("op_3223_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3223_cast_fp16 = slice_by_index(begin = var_3223_begin_0, end = var_3223_end_0, end_mask = var_3223_end_mask_0, x = next_cache_29_cast_fp16)[name = tensor("op_3223_cast_fp16")]; + tensor x_381_pad_type_0 = const()[name = tensor("x_381_pad_type_0"), val = tensor("valid")]; + tensor x_381_groups_0 = const()[name = tensor("x_381_groups_0"), val = tensor(512)]; + tensor x_381_strides_0 = const()[name = tensor("x_381_strides_0"), val = tensor([1])]; + tensor x_381_pad_0 = const()[name = tensor("x_381_pad_0"), val = tensor([0, 0])]; + tensor x_381_dilations_0 = const()[name = tensor("x_381_dilations_0"), val = tensor([1])]; + tensor encoder_layers_14_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179155136)))]; + tensor x_381_cast_fp16 = conv(dilations = x_381_dilations_0, groups = x_381_groups_0, pad = x_381_pad_0, pad_type = x_381_pad_type_0, strides = x_381_strides_0, weight = encoder_layers_14_conv_depthwise_conv_weight_to_fp16, x = new_x_59_cast_fp16)[name = tensor("x_381_cast_fp16")]; + tensor input_759_perm_0 = const()[name = tensor("input_759_perm_0"), val = tensor([0, 2, 1])]; + tensor x_383_axes_0 = const()[name = tensor("x_383_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179164416)))]; + tensor encoder_layers_14_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179165504)))]; + tensor input_759_cast_fp16 = transpose(perm = input_759_perm_0, x = x_381_cast_fp16)[name = tensor("transpose_106")]; + tensor x_383_cast_fp16 = layer_norm(axes = x_383_axes_0, beta = encoder_layers_14_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_14_conv_batch_norm_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("x_383_cast_fp16")]; + tensor input_761_perm_0 = const()[name = tensor("input_761_perm_0"), val = tensor([0, 2, 1])]; + tensor input_761_cast_fp16 = transpose(perm = input_761_perm_0, x = x_383_cast_fp16)[name = tensor("transpose_105")]; + tensor input_763_cast_fp16 = silu(x = input_761_cast_fp16)[name = tensor("input_763_cast_fp16")]; + tensor x_385_pad_type_0 = const()[name = tensor("x_385_pad_type_0"), val = tensor("valid")]; + tensor x_385_strides_0 = const()[name = tensor("x_385_strides_0"), val = tensor([1])]; + tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0])]; + tensor x_385_dilations_0 = const()[name = tensor("x_385_dilations_0"), val = tensor([1])]; + tensor x_385_groups_0 = const()[name = tensor("x_385_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179166592)))]; + tensor x_385_cast_fp16 = conv(dilations = x_385_dilations_0, groups = x_385_groups_0, pad = x_385_pad_0, pad_type = x_385_pad_type_0, strides = x_385_strides_0, weight = encoder_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_763_cast_fp16)[name = tensor("x_385_cast_fp16")]; + tensor input_765_perm_0 = const()[name = tensor("input_765_perm_0"), val = tensor([0, 2, 1])]; + tensor input_765_cast_fp16 = transpose(perm = input_765_perm_0, x = x_385_cast_fp16)[name = tensor("transpose_104")]; + tensor input_767_cast_fp16 = add(x = input_751_cast_fp16, y = input_765_cast_fp16)[name = tensor("input_767_cast_fp16")]; + tensor input_769_axes_0 = const()[name = tensor("input_769_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179690944)))]; + tensor encoder_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179692032)))]; + tensor input_769_cast_fp16 = layer_norm(axes = input_769_axes_0, beta = encoder_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_14_norm_feed_forward2_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("input_769_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179693120)))]; + tensor linear_133_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_769_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor input_773_cast_fp16 = silu(x = linear_133_cast_fp16)[name = tensor("input_773_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181790336)))]; + tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_773_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor var_3264_to_fp16 = const()[name = tensor("op_3264_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3265_cast_fp16 = mul(x = linear_134_cast_fp16, y = var_3264_to_fp16)[name = tensor("op_3265_cast_fp16")]; + tensor input_779_cast_fp16 = add(x = input_767_cast_fp16, y = var_3265_cast_fp16)[name = tensor("input_779_cast_fp16")]; + tensor input_781_axes_0 = const()[name = tensor("input_781_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183887552)))]; + tensor encoder_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183888640)))]; + tensor input_781_cast_fp16 = layer_norm(axes = input_781_axes_0, beta = encoder_layers_14_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_14_norm_out_weight_to_fp16, x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; + tensor cache_61_begin_0 = const()[name = tensor("cache_61_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_61_end_0 = const()[name = tensor("cache_61_end_0"), val = tensor([16, 1, 70, 512])]; + tensor cache_61_end_mask_0 = const()[name = tensor("cache_61_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_61_squeeze_mask_0 = const()[name = tensor("cache_61_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_61_cast_fp16 = slice_by_index(begin = cache_61_begin_0, end = cache_61_end_0, end_mask = cache_61_end_mask_0, squeeze_mask = cache_61_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_61_cast_fp16")]; + tensor cache_63_begin_0 = const()[name = tensor("cache_63_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_63_end_0 = const()[name = tensor("cache_63_end_0"), val = tensor([16, 1, 512, 8])]; + tensor cache_63_end_mask_0 = const()[name = tensor("cache_63_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_63_squeeze_mask_0 = const()[name = tensor("cache_63_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_63_cast_fp16 = slice_by_index(begin = cache_63_begin_0, end = cache_63_end_0, end_mask = cache_63_end_mask_0, squeeze_mask = cache_63_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_63_cast_fp16")]; + tensor input_783_axes_0 = const()[name = tensor("input_783_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183889728)))]; + tensor encoder_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183890816)))]; + tensor input_783_cast_fp16 = layer_norm(axes = input_783_axes_0, beta = encoder_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_15_norm_feed_forward1_weight_to_fp16, x = input_781_cast_fp16)[name = tensor("input_783_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183891904)))]; + tensor linear_135_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_783_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor input_787_cast_fp16 = silu(x = linear_135_cast_fp16)[name = tensor("input_787_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185989120)))]; + tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor var_3299_to_fp16 = const()[name = tensor("op_3299_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3300_cast_fp16 = mul(x = linear_136_cast_fp16, y = var_3299_to_fp16)[name = tensor("op_3300_cast_fp16")]; + tensor input_793_cast_fp16 = add(x = input_781_cast_fp16, y = var_3300_cast_fp16)[name = tensor("input_793_cast_fp16")]; + tensor key_31_axes_0 = const()[name = tensor("key_31_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188086336)))]; + tensor encoder_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188087424)))]; + tensor key_31_cast_fp16 = layer_norm(axes = key_31_axes_0, beta = encoder_layers_15_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_15_norm_self_att_weight_to_fp16, x = input_793_cast_fp16)[name = tensor("key_31_cast_fp16")]; + tensor input_795_interleave_0 = const()[name = tensor("input_795_interleave_0"), val = tensor(false)]; + tensor input_795_cast_fp16 = concat(axis = var_44, interleave = input_795_interleave_0, values = (cache_61_cast_fp16, key_31_cast_fp16))[name = tensor("input_795_cast_fp16")]; + tensor var_3322_begin_0 = const()[name = tensor("op_3322_begin_0"), val = tensor([0, 14, 0])]; + tensor var_3322_end_0 = const()[name = tensor("op_3322_end_0"), val = tensor([1, 70, 512])]; + tensor var_3322_end_mask_0 = const()[name = tensor("op_3322_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3322_cast_fp16 = slice_by_index(begin = var_3322_begin_0, end = var_3322_end_0, end_mask = var_3322_end_mask_0, x = cache_61_cast_fp16)[name = tensor("op_3322_cast_fp16")]; + tensor var_3325_begin_0 = const()[name = tensor("op_3325_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3325_end_0 = const()[name = tensor("op_3325_end_0"), val = tensor([1, 14, 512])]; + tensor var_3325_end_mask_0 = const()[name = tensor("op_3325_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3325_cast_fp16 = slice_by_index(begin = var_3325_begin_0, end = var_3325_end_0, end_mask = var_3325_end_mask_0, x = key_31_cast_fp16)[name = tensor("op_3325_cast_fp16")]; + tensor var_3328_interleave_0 = const()[name = tensor("op_3328_interleave_0"), val = tensor(false)]; + tensor var_3328_cast_fp16 = concat(axis = var_44, interleave = var_3328_interleave_0, values = (var_3322_cast_fp16, var_3325_cast_fp16))[name = tensor("op_3328_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188088512)))]; + tensor linear_137_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_q_weight_to_fp16, x = key_31_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor var_3332 = const()[name = tensor("op_3332"), val = tensor([1, -1, 8, 64])]; + tensor q_91_cast_fp16 = reshape(shape = var_3332, x = linear_137_cast_fp16)[name = tensor("q_91_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188612864)))]; + tensor linear_138_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_k_weight_to_fp16, x = input_795_cast_fp16)[name = tensor("linear_138_cast_fp16")]; + tensor var_3336 = const()[name = tensor("op_3336"), val = tensor([1, -1, 8, 64])]; + tensor k_61_cast_fp16 = reshape(shape = var_3336, x = linear_138_cast_fp16)[name = tensor("k_61_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189137216)))]; + tensor linear_139_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_v_weight_to_fp16, x = input_795_cast_fp16)[name = tensor("linear_139_cast_fp16")]; + tensor var_3340 = const()[name = tensor("op_3340"), val = tensor([1, -1, 8, 64])]; + tensor v_31_cast_fp16 = reshape(shape = var_3340, x = linear_139_cast_fp16)[name = tensor("v_31_cast_fp16")]; + tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189661568)))]; + tensor var_3352_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3352_cast_fp16")]; + tensor encoder_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189662656)))]; + tensor var_3354_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3354_cast_fp16")]; + tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_393_transpose_x_0 = const()[name = tensor("x_393_transpose_x_0"), val = tensor(false)]; + tensor x_393_transpose_y_0 = const()[name = tensor("x_393_transpose_y_0"), val = tensor(false)]; + tensor var_3356_to_fp16 = const()[name = tensor("op_3356_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189663744)))]; + tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3354_cast_fp16)[name = tensor("transpose_102")]; + tensor x_393_cast_fp16 = matmul(transpose_x = x_393_transpose_x_0, transpose_y = x_393_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3356_to_fp16)[name = tensor("x_393_cast_fp16")]; + tensor x_395_pad_0 = const()[name = tensor("x_395_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_395_mode_0 = const()[name = tensor("x_395_mode_0"), val = tensor("constant")]; + tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(0x0p+0)]; + tensor x_395_cast_fp16 = pad(constant_val = const_213_to_fp16, mode = x_395_mode_0, pad = x_395_pad_0, x = x_393_cast_fp16)[name = tensor("x_395_cast_fp16")]; + tensor var_3364 = const()[name = tensor("op_3364"), val = tensor([1, 8, -1, 16])]; + tensor x_397_cast_fp16 = reshape(shape = var_3364, x = x_395_cast_fp16)[name = tensor("x_397_cast_fp16")]; + tensor var_3368_begin_0 = const()[name = tensor("op_3368_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3368_end_0 = const()[name = tensor("op_3368_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_3368_end_mask_0 = const()[name = tensor("op_3368_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3368_cast_fp16 = slice_by_index(begin = var_3368_begin_0, end = var_3368_end_0, end_mask = var_3368_end_mask_0, x = x_397_cast_fp16)[name = tensor("op_3368_cast_fp16")]; + tensor var_3369 = const()[name = tensor("op_3369"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3369, x = var_3368_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; + tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; + tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_100")]; + tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3352_cast_fp16)[name = tensor("transpose_101")]; + tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; + tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; + tensor var_3378_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3378_cast_fp16")]; + tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_61_cast_fp16 = mul(x = var_3378_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; + tensor scores_63_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_3)[name = tensor("scores_63_cast_fp16")]; + tensor var_3384_cast_fp16 = softmax(axis = var_30, x = scores_63_cast_fp16)[name = tensor("op_3384_cast_fp16")]; + tensor input_797_cast_fp16 = select(a = var_23_to_fp16, b = var_3384_cast_fp16, cond = mask_3)[name = tensor("input_797_cast_fp16")]; + tensor x_399_transpose_x_0 = const()[name = tensor("x_399_transpose_x_0"), val = tensor(false)]; + tensor x_399_transpose_y_0 = const()[name = tensor("x_399_transpose_y_0"), val = tensor(false)]; + tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_103")]; + tensor x_399_cast_fp16 = matmul(transpose_x = x_399_transpose_x_0, transpose_y = x_399_transpose_y_0, x = input_797_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_399_cast_fp16")]; + tensor var_3388_perm_0 = const()[name = tensor("op_3388_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3389 = const()[name = tensor("op_3389"), val = tensor([1, -1, 512])]; + tensor var_3388_cast_fp16 = transpose(perm = var_3388_perm_0, x = x_399_cast_fp16)[name = tensor("transpose_99")]; + tensor input_799_cast_fp16 = reshape(shape = var_3389, x = var_3388_cast_fp16)[name = tensor("input_799_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189838912)))]; + tensor linear_141_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_out_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("linear_141_cast_fp16")]; + tensor input_803_cast_fp16 = add(x = input_793_cast_fp16, y = linear_141_cast_fp16)[name = tensor("input_803_cast_fp16")]; + tensor x_403_axes_0 = const()[name = tensor("x_403_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190363264)))]; + tensor encoder_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190364352)))]; + tensor x_403_cast_fp16 = layer_norm(axes = x_403_axes_0, beta = encoder_layers_15_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_15_norm_conv_weight_to_fp16, x = input_803_cast_fp16)[name = tensor("x_403_cast_fp16")]; + tensor input_805_perm_0 = const()[name = tensor("input_805_perm_0"), val = tensor([0, 2, 1])]; + tensor input_807_pad_type_0 = const()[name = tensor("input_807_pad_type_0"), val = tensor("valid")]; + tensor input_807_strides_0 = const()[name = tensor("input_807_strides_0"), val = tensor([1])]; + tensor input_807_pad_0 = const()[name = tensor("input_807_pad_0"), val = tensor([0, 0])]; + tensor input_807_dilations_0 = const()[name = tensor("input_807_dilations_0"), val = tensor([1])]; + tensor input_807_groups_0 = const()[name = tensor("input_807_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190365440)))]; + tensor input_805_cast_fp16 = transpose(perm = input_805_perm_0, x = x_403_cast_fp16)[name = tensor("transpose_98")]; + tensor input_807_cast_fp16 = conv(dilations = input_807_dilations_0, groups = input_807_groups_0, pad = input_807_pad_0, pad_type = input_807_pad_type_0, strides = input_807_strides_0, weight = encoder_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("input_807_cast_fp16")]; + tensor x_405_split_num_splits_0 = const()[name = tensor("x_405_split_num_splits_0"), val = tensor(2)]; + tensor x_405_split_axis_0 = const()[name = tensor("x_405_split_axis_0"), val = tensor(1)]; + tensor x_405_split_cast_fp16_0, tensor x_405_split_cast_fp16_1 = split(axis = x_405_split_axis_0, num_splits = x_405_split_num_splits_0, x = input_807_cast_fp16)[name = tensor("x_405_split_cast_fp16")]; + tensor x_405_split_1_sigmoid_cast_fp16 = sigmoid(x = x_405_split_cast_fp16_1)[name = tensor("x_405_split_1_sigmoid_cast_fp16")]; + tensor x_405_cast_fp16 = mul(x = x_405_split_cast_fp16_0, y = x_405_split_1_sigmoid_cast_fp16)[name = tensor("x_405_cast_fp16")]; + tensor input_809_cast_fp16 = select(a = var_23_to_fp16, b = x_405_cast_fp16, cond = var_309)[name = tensor("input_809_cast_fp16")]; + tensor new_x_63_interleave_0 = const()[name = tensor("new_x_63_interleave_0"), val = tensor(false)]; + tensor new_x_63_cast_fp16 = concat(axis = var_30, interleave = new_x_63_interleave_0, values = (cache_63_cast_fp16, input_809_cast_fp16))[name = tensor("new_x_63_cast_fp16")]; + tensor next_cache_31_begin_0 = const()[name = tensor("next_cache_31_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_31_end_0 = const()[name = tensor("next_cache_31_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_31_end_mask_0 = const()[name = tensor("next_cache_31_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_31_cast_fp16 = slice_by_index(begin = next_cache_31_begin_0, end = next_cache_31_end_0, end_mask = next_cache_31_end_mask_0, x = new_x_63_cast_fp16)[name = tensor("next_cache_31_cast_fp16")]; + tensor var_3430_begin_0 = const()[name = tensor("op_3430_begin_0"), val = tensor([0, 0, 14])]; + tensor var_3430_end_0 = const()[name = tensor("op_3430_end_0"), val = tensor([1, 512, 22])]; + tensor var_3430_end_mask_0 = const()[name = tensor("op_3430_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3430_cast_fp16 = slice_by_index(begin = var_3430_begin_0, end = var_3430_end_0, end_mask = var_3430_end_mask_0, x = next_cache_31_cast_fp16)[name = tensor("op_3430_cast_fp16")]; + tensor x_407_pad_type_0 = const()[name = tensor("x_407_pad_type_0"), val = tensor("valid")]; + tensor x_407_groups_0 = const()[name = tensor("x_407_groups_0"), val = tensor(512)]; + tensor x_407_strides_0 = const()[name = tensor("x_407_strides_0"), val = tensor([1])]; + tensor x_407_pad_0 = const()[name = tensor("x_407_pad_0"), val = tensor([0, 0])]; + tensor x_407_dilations_0 = const()[name = tensor("x_407_dilations_0"), val = tensor([1])]; + tensor encoder_layers_15_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191414080)))]; + tensor x_407_cast_fp16 = conv(dilations = x_407_dilations_0, groups = x_407_groups_0, pad = x_407_pad_0, pad_type = x_407_pad_type_0, strides = x_407_strides_0, weight = encoder_layers_15_conv_depthwise_conv_weight_to_fp16, x = new_x_63_cast_fp16)[name = tensor("x_407_cast_fp16")]; + tensor input_811_perm_0 = const()[name = tensor("input_811_perm_0"), val = tensor([0, 2, 1])]; + tensor x_409_axes_0 = const()[name = tensor("x_409_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191423360)))]; + tensor encoder_layers_15_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191424448)))]; + tensor input_811_cast_fp16 = transpose(perm = input_811_perm_0, x = x_407_cast_fp16)[name = tensor("transpose_97")]; + tensor x_409_cast_fp16 = layer_norm(axes = x_409_axes_0, beta = encoder_layers_15_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_15_conv_batch_norm_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("x_409_cast_fp16")]; + tensor input_813_perm_0 = const()[name = tensor("input_813_perm_0"), val = tensor([0, 2, 1])]; + tensor input_813_cast_fp16 = transpose(perm = input_813_perm_0, x = x_409_cast_fp16)[name = tensor("transpose_96")]; + tensor input_815_cast_fp16 = silu(x = input_813_cast_fp16)[name = tensor("input_815_cast_fp16")]; + tensor x_411_pad_type_0 = const()[name = tensor("x_411_pad_type_0"), val = tensor("valid")]; + tensor x_411_strides_0 = const()[name = tensor("x_411_strides_0"), val = tensor([1])]; + tensor x_411_pad_0 = const()[name = tensor("x_411_pad_0"), val = tensor([0, 0])]; + tensor x_411_dilations_0 = const()[name = tensor("x_411_dilations_0"), val = tensor([1])]; + tensor x_411_groups_0 = const()[name = tensor("x_411_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191425536)))]; + tensor x_411_cast_fp16 = conv(dilations = x_411_dilations_0, groups = x_411_groups_0, pad = x_411_pad_0, pad_type = x_411_pad_type_0, strides = x_411_strides_0, weight = encoder_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_815_cast_fp16)[name = tensor("x_411_cast_fp16")]; + tensor input_817_perm_0 = const()[name = tensor("input_817_perm_0"), val = tensor([0, 2, 1])]; + tensor input_817_cast_fp16 = transpose(perm = input_817_perm_0, x = x_411_cast_fp16)[name = tensor("transpose_95")]; + tensor input_819_cast_fp16 = add(x = input_803_cast_fp16, y = input_817_cast_fp16)[name = tensor("input_819_cast_fp16")]; + tensor input_821_axes_0 = const()[name = tensor("input_821_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191949888)))]; + tensor encoder_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191950976)))]; + tensor input_821_cast_fp16 = layer_norm(axes = input_821_axes_0, beta = encoder_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_15_norm_feed_forward2_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("input_821_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191952064)))]; + tensor linear_142_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_821_cast_fp16)[name = tensor("linear_142_cast_fp16")]; + tensor input_825_cast_fp16 = silu(x = linear_142_cast_fp16)[name = tensor("input_825_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194049280)))]; + tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_825_cast_fp16)[name = tensor("linear_143_cast_fp16")]; + tensor var_3471_to_fp16 = const()[name = tensor("op_3471_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3472_cast_fp16 = mul(x = linear_143_cast_fp16, y = var_3471_to_fp16)[name = tensor("op_3472_cast_fp16")]; + tensor input_831_cast_fp16 = add(x = input_819_cast_fp16, y = var_3472_cast_fp16)[name = tensor("input_831_cast_fp16")]; + tensor input_833_axes_0 = const()[name = tensor("input_833_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196146496)))]; + tensor encoder_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196147584)))]; + tensor input_833_cast_fp16 = layer_norm(axes = input_833_axes_0, beta = encoder_layers_15_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_15_norm_out_weight_to_fp16, x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; + tensor cache_65_begin_0 = const()[name = tensor("cache_65_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_65_end_0 = const()[name = tensor("cache_65_end_0"), val = tensor([17, 1, 70, 512])]; + tensor cache_65_end_mask_0 = const()[name = tensor("cache_65_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_65_squeeze_mask_0 = const()[name = tensor("cache_65_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_65_cast_fp16 = slice_by_index(begin = cache_65_begin_0, end = cache_65_end_0, end_mask = cache_65_end_mask_0, squeeze_mask = cache_65_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_65_cast_fp16")]; + tensor cache_begin_0 = const()[name = tensor("cache_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_end_0 = const()[name = tensor("cache_end_0"), val = tensor([17, 1, 512, 8])]; + tensor cache_end_mask_0 = const()[name = tensor("cache_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_squeeze_mask_0 = const()[name = tensor("cache_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_cast_fp16 = slice_by_index(begin = cache_begin_0, end = cache_end_0, end_mask = cache_end_mask_0, squeeze_mask = cache_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_cast_fp16")]; + tensor input_835_axes_0 = const()[name = tensor("input_835_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196148672)))]; + tensor encoder_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196149760)))]; + tensor input_835_cast_fp16 = layer_norm(axes = input_835_axes_0, beta = encoder_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_16_norm_feed_forward1_weight_to_fp16, x = input_833_cast_fp16)[name = tensor("input_835_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196150848)))]; + tensor linear_144_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_835_cast_fp16)[name = tensor("linear_144_cast_fp16")]; + tensor input_839_cast_fp16 = silu(x = linear_144_cast_fp16)[name = tensor("input_839_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198248064)))]; + tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("linear_145_cast_fp16")]; + tensor var_3506_to_fp16 = const()[name = tensor("op_3506_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3507_cast_fp16 = mul(x = linear_145_cast_fp16, y = var_3506_to_fp16)[name = tensor("op_3507_cast_fp16")]; + tensor input_845_cast_fp16 = add(x = input_833_cast_fp16, y = var_3507_cast_fp16)[name = tensor("input_845_cast_fp16")]; + tensor key_axes_0 = const()[name = tensor("key_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200345280)))]; + tensor encoder_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200346368)))]; + tensor key_cast_fp16 = layer_norm(axes = key_axes_0, beta = encoder_layers_16_norm_self_att_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_16_norm_self_att_weight_to_fp16, x = input_845_cast_fp16)[name = tensor("key_cast_fp16")]; + tensor input_847_interleave_0 = const()[name = tensor("input_847_interleave_0"), val = tensor(false)]; + tensor input_847_cast_fp16 = concat(axis = var_44, interleave = input_847_interleave_0, values = (cache_65_cast_fp16, key_cast_fp16))[name = tensor("input_847_cast_fp16")]; + tensor var_3529_begin_0 = const()[name = tensor("op_3529_begin_0"), val = tensor([0, 14, 0])]; + tensor var_3529_end_0 = const()[name = tensor("op_3529_end_0"), val = tensor([1, 70, 512])]; + tensor var_3529_end_mask_0 = const()[name = tensor("op_3529_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3529_cast_fp16 = slice_by_index(begin = var_3529_begin_0, end = var_3529_end_0, end_mask = var_3529_end_mask_0, x = cache_65_cast_fp16)[name = tensor("op_3529_cast_fp16")]; + tensor var_3532_begin_0 = const()[name = tensor("op_3532_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3532_end_0 = const()[name = tensor("op_3532_end_0"), val = tensor([1, 14, 512])]; + tensor var_3532_end_mask_0 = const()[name = tensor("op_3532_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3532_cast_fp16 = slice_by_index(begin = var_3532_begin_0, end = var_3532_end_0, end_mask = var_3532_end_mask_0, x = key_cast_fp16)[name = tensor("op_3532_cast_fp16")]; + tensor cache_last_channel_cur_interleave_0 = const()[name = tensor("cache_last_channel_cur_interleave_0"), val = tensor(false)]; + tensor cache_last_channel_cur_cast_fp16 = concat(axis = var_44, interleave = cache_last_channel_cur_interleave_0, values = (var_3529_cast_fp16, var_3532_cast_fp16))[name = tensor("cache_last_channel_cur_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200347456)))]; + tensor linear_146_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_q_weight_to_fp16, x = key_cast_fp16)[name = tensor("linear_146_cast_fp16")]; + tensor var_3539 = const()[name = tensor("op_3539"), val = tensor([1, -1, 8, 64])]; + tensor q_97_cast_fp16 = reshape(shape = var_3539, x = linear_146_cast_fp16)[name = tensor("q_97_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200871808)))]; + tensor linear_147_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_k_weight_to_fp16, x = input_847_cast_fp16)[name = tensor("linear_147_cast_fp16")]; + tensor var_3543 = const()[name = tensor("op_3543"), val = tensor([1, -1, 8, 64])]; + tensor k_65_cast_fp16 = reshape(shape = var_3543, x = linear_147_cast_fp16)[name = tensor("k_65_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201396160)))]; + tensor linear_148_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_v_weight_to_fp16, x = input_847_cast_fp16)[name = tensor("linear_148_cast_fp16")]; + tensor var_3547 = const()[name = tensor("op_3547"), val = tensor([1, -1, 8, 64])]; + tensor v_cast_fp16 = reshape(shape = var_3547, x = linear_148_cast_fp16)[name = tensor("v_cast_fp16")]; + tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201920512)))]; + tensor var_3559_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3559_cast_fp16")]; + tensor encoder_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201921600)))]; + tensor var_3561_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3561_cast_fp16")]; + tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_419_transpose_x_0 = const()[name = tensor("x_419_transpose_x_0"), val = tensor(false)]; + tensor x_419_transpose_y_0 = const()[name = tensor("x_419_transpose_y_0"), val = tensor(false)]; + tensor var_3563_to_fp16 = const()[name = tensor("op_3563_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201922688)))]; + tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3561_cast_fp16)[name = tensor("transpose_93")]; + tensor x_419_cast_fp16 = matmul(transpose_x = x_419_transpose_x_0, transpose_y = x_419_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3563_to_fp16)[name = tensor("x_419_cast_fp16")]; + tensor x_421_pad_0 = const()[name = tensor("x_421_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_421_mode_0 = const()[name = tensor("x_421_mode_0"), val = tensor("constant")]; + tensor const_226_to_fp16 = const()[name = tensor("const_226_to_fp16"), val = tensor(0x0p+0)]; + tensor x_421_cast_fp16 = pad(constant_val = const_226_to_fp16, mode = x_421_mode_0, pad = x_421_pad_0, x = x_419_cast_fp16)[name = tensor("x_421_cast_fp16")]; + tensor var_3571 = const()[name = tensor("op_3571"), val = tensor([1, 8, -1, 16])]; + tensor x_423_cast_fp16 = reshape(shape = var_3571, x = x_421_cast_fp16)[name = tensor("x_423_cast_fp16")]; + tensor var_3575_begin_0 = const()[name = tensor("op_3575_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3575_end_0 = const()[name = tensor("op_3575_end_0"), val = tensor([1, 8, 172, 16])]; + tensor var_3575_end_mask_0 = const()[name = tensor("op_3575_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3575_cast_fp16 = slice_by_index(begin = var_3575_begin_0, end = var_3575_end_0, end_mask = var_3575_end_mask_0, x = x_423_cast_fp16)[name = tensor("op_3575_cast_fp16")]; + tensor var_3576 = const()[name = tensor("op_3576"), val = tensor([1, 8, 16, 171])]; + tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3576, x = var_3575_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; + tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; + tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_91")]; + tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3559_cast_fp16)[name = tensor("transpose_92")]; + tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; + tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 16, 86])]; + tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; + tensor var_3585_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3585_cast_fp16")]; + tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_65_cast_fp16 = mul(x = var_3585_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; + tensor scores_cast_fp16 = select(a = var_24_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_3)[name = tensor("scores_cast_fp16")]; + tensor var_3591_cast_fp16 = softmax(axis = var_30, x = scores_cast_fp16)[name = tensor("op_3591_cast_fp16")]; + tensor input_849_cast_fp16 = select(a = var_23_to_fp16, b = var_3591_cast_fp16, cond = mask_3)[name = tensor("input_849_cast_fp16")]; + tensor x_425_transpose_x_0 = const()[name = tensor("x_425_transpose_x_0"), val = tensor(false)]; + tensor x_425_transpose_y_0 = const()[name = tensor("x_425_transpose_y_0"), val = tensor(false)]; + tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_94")]; + tensor x_425_cast_fp16 = matmul(transpose_x = x_425_transpose_x_0, transpose_y = x_425_transpose_y_0, x = input_849_cast_fp16, y = value_cast_fp16)[name = tensor("x_425_cast_fp16")]; + tensor var_3595_perm_0 = const()[name = tensor("op_3595_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3596 = const()[name = tensor("op_3596"), val = tensor([1, -1, 512])]; + tensor var_3595_cast_fp16 = transpose(perm = var_3595_perm_0, x = x_425_cast_fp16)[name = tensor("transpose_90")]; + tensor input_851_cast_fp16 = reshape(shape = var_3596, x = var_3595_cast_fp16)[name = tensor("input_851_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202097856)))]; + tensor linear_150_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_out_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("linear_150_cast_fp16")]; + tensor input_855_cast_fp16 = add(x = input_845_cast_fp16, y = linear_150_cast_fp16)[name = tensor("input_855_cast_fp16")]; + tensor x_429_axes_0 = const()[name = tensor("x_429_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202622208)))]; + tensor encoder_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202623296)))]; + tensor x_429_cast_fp16 = layer_norm(axes = x_429_axes_0, beta = encoder_layers_16_norm_conv_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_16_norm_conv_weight_to_fp16, x = input_855_cast_fp16)[name = tensor("x_429_cast_fp16")]; + tensor input_857_perm_0 = const()[name = tensor("input_857_perm_0"), val = tensor([0, 2, 1])]; + tensor input_859_pad_type_0 = const()[name = tensor("input_859_pad_type_0"), val = tensor("valid")]; + tensor input_859_strides_0 = const()[name = tensor("input_859_strides_0"), val = tensor([1])]; + tensor input_859_pad_0 = const()[name = tensor("input_859_pad_0"), val = tensor([0, 0])]; + tensor input_859_dilations_0 = const()[name = tensor("input_859_dilations_0"), val = tensor([1])]; + tensor input_859_groups_0 = const()[name = tensor("input_859_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202624384)))]; + tensor input_857_cast_fp16 = transpose(perm = input_857_perm_0, x = x_429_cast_fp16)[name = tensor("transpose_89")]; + tensor input_859_cast_fp16 = conv(dilations = input_859_dilations_0, groups = input_859_groups_0, pad = input_859_pad_0, pad_type = input_859_pad_type_0, strides = input_859_strides_0, weight = encoder_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("input_859_cast_fp16")]; + tensor x_431_split_num_splits_0 = const()[name = tensor("x_431_split_num_splits_0"), val = tensor(2)]; + tensor x_431_split_axis_0 = const()[name = tensor("x_431_split_axis_0"), val = tensor(1)]; + tensor x_431_split_cast_fp16_0, tensor x_431_split_cast_fp16_1 = split(axis = x_431_split_axis_0, num_splits = x_431_split_num_splits_0, x = input_859_cast_fp16)[name = tensor("x_431_split_cast_fp16")]; + tensor x_431_split_1_sigmoid_cast_fp16 = sigmoid(x = x_431_split_cast_fp16_1)[name = tensor("x_431_split_1_sigmoid_cast_fp16")]; + tensor x_431_cast_fp16 = mul(x = x_431_split_cast_fp16_0, y = x_431_split_1_sigmoid_cast_fp16)[name = tensor("x_431_cast_fp16")]; + tensor input_861_cast_fp16 = select(a = var_23_to_fp16, b = x_431_cast_fp16, cond = var_309)[name = tensor("input_861_cast_fp16")]; + tensor new_x_interleave_0 = const()[name = tensor("new_x_interleave_0"), val = tensor(false)]; + tensor new_x_cast_fp16 = concat(axis = var_30, interleave = new_x_interleave_0, values = (cache_cast_fp16, input_861_cast_fp16))[name = tensor("new_x_cast_fp16")]; + tensor next_cache_begin_0 = const()[name = tensor("next_cache_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_end_0 = const()[name = tensor("next_cache_end_0"), val = tensor([1, 512, 22])]; + tensor next_cache_end_mask_0 = const()[name = tensor("next_cache_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_cast_fp16 = slice_by_index(begin = next_cache_begin_0, end = next_cache_end_0, end_mask = next_cache_end_mask_0, x = new_x_cast_fp16)[name = tensor("next_cache_cast_fp16")]; + tensor cache_last_time_cur_begin_0 = const()[name = tensor("cache_last_time_cur_begin_0"), val = tensor([0, 0, 14])]; + tensor cache_last_time_cur_end_0 = const()[name = tensor("cache_last_time_cur_end_0"), val = tensor([1, 512, 22])]; + tensor cache_last_time_cur_end_mask_0 = const()[name = tensor("cache_last_time_cur_end_mask_0"), val = tensor([true, true, true])]; + tensor cache_last_time_cur_cast_fp16 = slice_by_index(begin = cache_last_time_cur_begin_0, end = cache_last_time_cur_end_0, end_mask = cache_last_time_cur_end_mask_0, x = next_cache_cast_fp16)[name = tensor("cache_last_time_cur_cast_fp16")]; + tensor x_433_pad_type_0 = const()[name = tensor("x_433_pad_type_0"), val = tensor("valid")]; + tensor x_433_groups_0 = const()[name = tensor("x_433_groups_0"), val = tensor(512)]; + tensor x_433_strides_0 = const()[name = tensor("x_433_strides_0"), val = tensor([1])]; + tensor x_433_pad_0 = const()[name = tensor("x_433_pad_0"), val = tensor([0, 0])]; + tensor x_433_dilations_0 = const()[name = tensor("x_433_dilations_0"), val = tensor([1])]; + tensor encoder_layers_16_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203673024)))]; + tensor x_433_cast_fp16 = conv(dilations = x_433_dilations_0, groups = x_433_groups_0, pad = x_433_pad_0, pad_type = x_433_pad_type_0, strides = x_433_strides_0, weight = encoder_layers_16_conv_depthwise_conv_weight_to_fp16, x = new_x_cast_fp16)[name = tensor("x_433_cast_fp16")]; + tensor input_863_perm_0 = const()[name = tensor("input_863_perm_0"), val = tensor([0, 2, 1])]; + tensor x_435_axes_0 = const()[name = tensor("x_435_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203682304)))]; + tensor encoder_layers_16_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203683392)))]; + tensor input_863_cast_fp16 = transpose(perm = input_863_perm_0, x = x_433_cast_fp16)[name = tensor("transpose_88")]; + tensor x_435_cast_fp16 = layer_norm(axes = x_435_axes_0, beta = encoder_layers_16_conv_batch_norm_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_16_conv_batch_norm_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("x_435_cast_fp16")]; + tensor input_865_perm_0 = const()[name = tensor("input_865_perm_0"), val = tensor([0, 2, 1])]; + tensor input_865_cast_fp16 = transpose(perm = input_865_perm_0, x = x_435_cast_fp16)[name = tensor("transpose_87")]; + tensor input_867_cast_fp16 = silu(x = input_865_cast_fp16)[name = tensor("input_867_cast_fp16")]; + tensor x_437_pad_type_0 = const()[name = tensor("x_437_pad_type_0"), val = tensor("valid")]; + tensor x_437_strides_0 = const()[name = tensor("x_437_strides_0"), val = tensor([1])]; + tensor x_437_pad_0 = const()[name = tensor("x_437_pad_0"), val = tensor([0, 0])]; + tensor x_437_dilations_0 = const()[name = tensor("x_437_dilations_0"), val = tensor([1])]; + tensor x_437_groups_0 = const()[name = tensor("x_437_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203684480)))]; + tensor x_437_cast_fp16 = conv(dilations = x_437_dilations_0, groups = x_437_groups_0, pad = x_437_pad_0, pad_type = x_437_pad_type_0, strides = x_437_strides_0, weight = encoder_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_867_cast_fp16)[name = tensor("x_437_cast_fp16")]; + tensor input_869_perm_0 = const()[name = tensor("input_869_perm_0"), val = tensor([0, 2, 1])]; + tensor input_869_cast_fp16 = transpose(perm = input_869_perm_0, x = x_437_cast_fp16)[name = tensor("transpose_86")]; + tensor input_871_cast_fp16 = add(x = input_855_cast_fp16, y = input_869_cast_fp16)[name = tensor("input_871_cast_fp16")]; + tensor input_873_axes_0 = const()[name = tensor("input_873_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204208832)))]; + tensor encoder_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204209920)))]; + tensor input_873_cast_fp16 = layer_norm(axes = input_873_axes_0, beta = encoder_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_16_norm_feed_forward2_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("input_873_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204211008)))]; + tensor linear_151_cast_fp16 = linear(bias = linear_0_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_873_cast_fp16)[name = tensor("linear_151_cast_fp16")]; + tensor input_877_cast_fp16 = silu(x = linear_151_cast_fp16)[name = tensor("input_877_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206308224)))]; + tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_877_cast_fp16)[name = tensor("linear_152_cast_fp16")]; + tensor var_3678_to_fp16 = const()[name = tensor("op_3678_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3679_cast_fp16 = mul(x = linear_152_cast_fp16, y = var_3678_to_fp16)[name = tensor("op_3679_cast_fp16")]; + tensor input_cast_fp16 = add(x = input_871_cast_fp16, y = var_3679_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208405440)))]; + tensor encoder_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208406528)))]; + tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = encoder_layers_16_norm_out_bias_to_fp16, epsilon = var_21_to_fp16, gamma = encoder_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; + tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; + tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor cast_167_dtype_0 = const()[name = tensor("cast_167_dtype_0"), val = tensor("int32")]; + tensor obj_5_axis_0 = const()[name = tensor("obj_5_axis_0"), val = tensor(0)]; + tensor obj_5_cast_fp16 = stack(axis = obj_5_axis_0, values = (var_223_cast_fp16, var_430_cast_fp16, var_637_cast_fp16, var_844_cast_fp16, var_1051_cast_fp16, var_1258_cast_fp16, var_1465_cast_fp16, var_1672_cast_fp16, var_1879_cast_fp16, var_2086_cast_fp16, var_2293_cast_fp16, var_2500_cast_fp16, var_2707_cast_fp16, var_2914_cast_fp16, var_3121_cast_fp16, var_3328_cast_fp16, cache_last_channel_cur_cast_fp16))[name = tensor("obj_5_cast_fp16")]; + tensor obj_7_axis_0 = const()[name = tensor("obj_7_axis_0"), val = tensor(0)]; + tensor obj_7_cast_fp16 = stack(axis = obj_7_axis_0, values = (var_325_cast_fp16, var_532_cast_fp16, var_739_cast_fp16, var_946_cast_fp16, var_1153_cast_fp16, var_1360_cast_fp16, var_1567_cast_fp16, var_1774_cast_fp16, var_1981_cast_fp16, var_2188_cast_fp16, var_2395_cast_fp16, var_2602_cast_fp16, var_2809_cast_fp16, var_3016_cast_fp16, var_3223_cast_fp16, var_3430_cast_fp16, cache_last_time_cur_cast_fp16))[name = tensor("obj_7_cast_fp16")]; + tensor obj_7_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_7_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_3695 = add(x = cache_last_channel_len, y = cache_keep_size)[name = tensor("op_3695")]; + tensor var_3695_promoted_to_fp16_dtype_0 = const()[name = tensor("op_3695_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor const_232_to_fp16 = const()[name = tensor("const_232_to_fp16"), val = tensor(-inf)]; + tensor var_29_promoted_to_fp16 = const()[name = tensor("op_29_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor var_3695_to_fp16 = cast(dtype = var_3695_promoted_to_fp16_dtype_0, x = var_3695)[name = tensor("cast_171")]; + tensor clip_1_cast_fp16 = clip(alpha = const_232_to_fp16, beta = var_29_promoted_to_fp16, x = var_3695_to_fp16)[name = tensor("clip_1_cast_fp16")]; + tensor var_3722_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_3722_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor cast_168_dtype_0 = const()[name = tensor("cast_168_dtype_0"), val = tensor("int32")]; + tensor new_cache_len = cast(dtype = cast_168_dtype_0, x = clip_1_cast_fp16)[name = tensor("cast_169")]; + tensor new_cache_channel = cast(dtype = var_3722_cast_fp16_to_fp32_dtype_0, x = obj_5_cast_fp16)[name = tensor("cast_170")]; + tensor new_cache_time = cast(dtype = obj_7_cast_fp16_to_fp32_dtype_0, x = obj_7_cast_fp16)[name = tensor("cast_172")]; + tensor encoder_length = cast(dtype = cast_167_dtype_0, x = clip_0_cast_fp16)[name = tensor("cast_173")]; + tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; + tensor encoder = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_174")]; + } -> (encoder, encoder_length, new_cache_channel, new_cache_time, new_cache_len); +} \ No newline at end of file