program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] { func main(tensor audio_length, tensor audio_signal, tensor cache_last_channel, tensor cache_last_channel_len, tensor cache_last_time, tensor pre_cache) { tensor var_9 = const()[name = tensor("op_9"), val = tensor(2)]; tensor full_input_interleave_0 = const()[name = tensor("full_input_interleave_0"), val = tensor(false)]; tensor pre_cache_to_fp16_dtype_0 = const()[name = tensor("pre_cache_to_fp16_dtype_0"), val = tensor("fp16")]; tensor audio_signal_to_fp16_dtype_0 = const()[name = tensor("audio_signal_to_fp16_dtype_0"), val = tensor("fp16")]; tensor audio_signal_to_fp16 = cast(dtype = audio_signal_to_fp16_dtype_0, x = audio_signal)[name = tensor("cast_199")]; tensor pre_cache_to_fp16 = cast(dtype = pre_cache_to_fp16_dtype_0, x = pre_cache)[name = tensor("cast_200")]; tensor full_input_cast_fp16 = concat(axis = var_9, interleave = full_input_interleave_0, values = (pre_cache_to_fp16, audio_signal_to_fp16))[name = tensor("full_input_cast_fp16")]; tensor var_12 = const()[name = tensor("op_12"), val = tensor(9)]; tensor value_1 = add(x = audio_length, y = var_12)[name = tensor("value_1")]; tensor var_28_begin_0 = const()[name = tensor("op_28_begin_0"), val = tensor([0, 0, 64])]; tensor var_28_end_0 = const()[name = tensor("op_28_end_0"), val = tensor([1, 128, 73])]; tensor var_28_end_mask_0 = const()[name = tensor("op_28_end_mask_0"), val = tensor([true, true, true])]; tensor var_28_cast_fp16 = slice_by_index(begin = var_28_begin_0, end = var_28_end_0, end_mask = var_28_end_mask_0, x = full_input_cast_fp16)[name = tensor("op_28_cast_fp16")]; tensor var_28_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_28_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor var_56 = const()[name = tensor("op_56"), val = tensor(-1)]; tensor var_65 = const()[name = tensor("op_65"), val = tensor(1)]; tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = full_input_cast_fp16)[name = tensor("transpose_241")]; tensor tensor_1_cast_fp16 = expand_dims(axes = tensor_1_axes_0, x = x_1_cast_fp16)[name = tensor("tensor_1_cast_fp16")]; tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72]])]; tensor var_120_axes_0 = const()[name = tensor("op_120_axes_0"), val = tensor([1])]; tensor var_120 = expand_dims(axes = var_120_axes_0, x = value_1)[name = tensor("op_120")]; tensor time_mask_1 = less(x = expand_dims_0, y = var_120)[name = tensor("time_mask_1")]; tensor var_122_axes_0 = const()[name = tensor("op_122_axes_0"), val = tensor([-1])]; tensor var_122 = expand_dims(axes = var_122_axes_0, x = time_mask_1)[name = tensor("op_122")]; tensor var_124_reps_0 = const()[name = tensor("op_124_reps_0"), val = tensor([1, 1, 128])]; tensor var_124 = tile(reps = var_124_reps_0, x = var_122)[name = tensor("op_124")]; tensor var_130_axes_0 = const()[name = tensor("op_130_axes_0"), val = tensor([1])]; tensor cast_2_to_fp16_dtype_0 = const()[name = tensor("cast_2_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_124_to_fp16 = cast(dtype = cast_2_to_fp16_dtype_0, x = var_124)[name = tensor("cast_197")]; tensor var_130_cast_fp16 = expand_dims(axes = var_130_axes_0, x = var_124_to_fp16)[name = tensor("op_130_cast_fp16")]; tensor input_1_cast_fp16 = mul(x = tensor_1_cast_fp16, y = var_130_cast_fp16)[name = tensor("input_1_cast_fp16")]; tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; tensor input_3_mode_0 = const()[name = tensor("input_3_mode_0"), val = tensor("constant")]; tensor const_9_to_fp16 = const()[name = tensor("const_9_to_fp16"), val = tensor(0x0p+0)]; tensor input_3_cast_fp16 = pad(constant_val = const_9_to_fp16, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("valid")]; tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([0, 0, 0, 0])]; tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; tensor encoder_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor encoder_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; tensor tensor_3_cast_fp16 = conv(bias = encoder_pre_encode_conv_0_bias_to_fp16, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = encoder_pre_encode_conv_0_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("tensor_3_cast_fp16")]; tensor cast_0_to_fp16_dtype_0 = const()[name = tensor("cast_0_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_143_promoted_to_fp16 = const()[name = tensor("op_143_promoted_to_fp16"), val = tensor(0x1p+1)]; tensor value_1_to_fp16 = cast(dtype = cast_0_to_fp16_dtype_0, x = value_1)[name = tensor("cast_196")]; tensor var_144_cast_fp16 = add(x = value_1_to_fp16, y = var_143_promoted_to_fp16)[name = tensor("op_144_cast_fp16")]; tensor var_145_promoted_to_fp16 = const()[name = tensor("op_145_promoted_to_fp16"), val = tensor(0x1p+0)]; tensor var_146_cast_fp16 = add(x = var_144_cast_fp16, y = var_145_promoted_to_fp16)[name = tensor("op_146_cast_fp16")]; tensor var_147_promoted_to_fp16 = const()[name = tensor("op_147_promoted_to_fp16"), val = tensor(0x1.8p+1)]; tensor var_148_cast_fp16 = sub(x = var_146_cast_fp16, y = var_147_promoted_to_fp16)[name = tensor("op_148_cast_fp16")]; tensor var_53_promoted_to_fp16 = const()[name = tensor("op_53_promoted_to_fp16"), val = tensor(0x1p+1)]; tensor floor_div_0_cast_fp16 = floor_div(x = var_148_cast_fp16, y = var_53_promoted_to_fp16)[name = tensor("floor_div_0_cast_fp16")]; tensor var_150_promoted_to_fp16 = const()[name = tensor("op_150_promoted_to_fp16"), val = tensor(0x1p+0)]; tensor current_lengths_3_cast_fp16 = add(x = floor_div_0_cast_fp16, y = var_150_promoted_to_fp16)[name = tensor("current_lengths_3_cast_fp16")]; tensor cast_3_dtype_0 = const()[name = tensor("cast_3_dtype_0"), val = tensor("int32")]; tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36]])]; tensor var_159_axes_0 = const()[name = tensor("op_159_axes_0"), val = tensor([1])]; tensor current_lengths_3_cast_fp16_to_int32 = cast(dtype = cast_3_dtype_0, x = current_lengths_3_cast_fp16)[name = tensor("cast_195")]; tensor var_159 = expand_dims(axes = var_159_axes_0, x = current_lengths_3_cast_fp16_to_int32)[name = tensor("op_159")]; tensor time_mask_3 = less(x = expand_dims_1, y = var_159)[name = tensor("time_mask_3")]; tensor var_161_axes_0 = const()[name = tensor("op_161_axes_0"), val = tensor([-1])]; tensor var_161 = expand_dims(axes = var_161_axes_0, x = time_mask_3)[name = tensor("op_161")]; tensor var_163_reps_0 = const()[name = tensor("op_163_reps_0"), val = tensor([1, 1, 65])]; tensor var_163 = tile(reps = var_163_reps_0, x = var_161)[name = tensor("op_163")]; tensor var_169_axes_0 = const()[name = tensor("op_169_axes_0"), val = tensor([1])]; tensor cast_4_to_fp16_dtype_0 = const()[name = tensor("cast_4_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_163_to_fp16 = cast(dtype = cast_4_to_fp16_dtype_0, x = var_163)[name = tensor("cast_194")]; tensor var_169_cast_fp16 = expand_dims(axes = var_169_axes_0, x = var_163_to_fp16)[name = tensor("op_169_cast_fp16")]; tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; tensor expanded_mask_3_cast_fp16 = tile(reps = expanded_mask_3_reps_0, x = var_169_cast_fp16)[name = tensor("expanded_mask_3_cast_fp16")]; tensor input_5_cast_fp16 = mul(x = tensor_3_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; tensor tensor_5_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("tensor_5_cast_fp16")]; tensor input_7_cast_fp16 = mul(x = tensor_5_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_7_cast_fp16")]; tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; tensor input_9_mode_0 = const()[name = tensor("input_9_mode_0"), val = tensor("constant")]; tensor const_23_to_fp16 = const()[name = tensor("const_23_to_fp16"), val = tensor(0x0p+0)]; tensor input_9_cast_fp16 = pad(constant_val = const_23_to_fp16, mode = input_9_mode_0, pad = input_9_pad_0, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("valid")]; tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([0, 0, 0, 0])]; tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; tensor encoder_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; tensor encoder_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; tensor tensor_7_cast_fp16 = conv(bias = encoder_pre_encode_conv_2_bias_to_fp16, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = encoder_pre_encode_conv_2_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("tensor_7_cast_fp16")]; tensor var_191_promoted_to_fp16 = const()[name = tensor("op_191_promoted_to_fp16"), val = tensor(0x1p+1)]; tensor var_192_cast_fp16 = add(x = current_lengths_3_cast_fp16, y = var_191_promoted_to_fp16)[name = tensor("op_192_cast_fp16")]; tensor var_193_promoted_to_fp16 = const()[name = tensor("op_193_promoted_to_fp16"), val = tensor(0x1p+0)]; tensor var_194_cast_fp16 = add(x = var_192_cast_fp16, y = var_193_promoted_to_fp16)[name = tensor("op_194_cast_fp16")]; tensor var_195_promoted_to_fp16 = const()[name = tensor("op_195_promoted_to_fp16"), val = tensor(0x1.8p+1)]; tensor var_196_cast_fp16 = sub(x = var_194_cast_fp16, y = var_195_promoted_to_fp16)[name = tensor("op_196_cast_fp16")]; tensor var_53_promoted_1_to_fp16 = const()[name = tensor("op_53_promoted_1_to_fp16"), val = tensor(0x1p+1)]; tensor floor_div_1_cast_fp16 = floor_div(x = var_196_cast_fp16, y = var_53_promoted_1_to_fp16)[name = tensor("floor_div_1_cast_fp16")]; tensor var_198_promoted_to_fp16 = const()[name = tensor("op_198_promoted_to_fp16"), val = tensor(0x1p+0)]; tensor current_lengths_5_cast_fp16 = add(x = floor_div_1_cast_fp16, y = var_198_promoted_to_fp16)[name = tensor("current_lengths_5_cast_fp16")]; tensor cast_5_dtype_0 = const()[name = tensor("cast_5_dtype_0"), val = tensor("int32")]; tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18]])]; tensor var_207_axes_0 = const()[name = tensor("op_207_axes_0"), val = tensor([1])]; tensor current_lengths_5_cast_fp16_to_int32 = cast(dtype = cast_5_dtype_0, x = current_lengths_5_cast_fp16)[name = tensor("cast_193")]; tensor var_207 = expand_dims(axes = var_207_axes_0, x = current_lengths_5_cast_fp16_to_int32)[name = tensor("op_207")]; tensor time_mask_5 = less(x = expand_dims_2, y = var_207)[name = tensor("time_mask_5")]; tensor var_209_axes_0 = const()[name = tensor("op_209_axes_0"), val = tensor([-1])]; tensor var_209 = expand_dims(axes = var_209_axes_0, x = time_mask_5)[name = tensor("op_209")]; tensor var_211_reps_0 = const()[name = tensor("op_211_reps_0"), val = tensor([1, 1, 33])]; tensor var_211 = tile(reps = var_211_reps_0, x = var_209)[name = tensor("op_211")]; tensor var_217_axes_0 = const()[name = tensor("op_217_axes_0"), val = tensor([1])]; tensor cast_6_to_fp16_dtype_0 = const()[name = tensor("cast_6_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_211_to_fp16 = cast(dtype = cast_6_to_fp16_dtype_0, x = var_211)[name = tensor("cast_192")]; tensor var_217_cast_fp16 = expand_dims(axes = var_217_axes_0, x = var_211_to_fp16)[name = tensor("op_217_cast_fp16")]; tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; tensor expanded_mask_7_cast_fp16 = tile(reps = expanded_mask_7_reps_0, x = var_217_cast_fp16)[name = tensor("expanded_mask_7_cast_fp16")]; tensor input_11_cast_fp16 = mul(x = tensor_7_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_11_cast_fp16")]; tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; tensor encoder_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; tensor encoder_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; tensor tensor_9_cast_fp16 = conv(bias = encoder_pre_encode_conv_3_bias_to_fp16, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = encoder_pre_encode_conv_3_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("tensor_9_cast_fp16")]; tensor input_13_cast_fp16 = mul(x = tensor_9_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_13_cast_fp16")]; tensor tensor_11_cast_fp16 = relu(x = input_13_cast_fp16)[name = tensor("tensor_11_cast_fp16")]; tensor input_15_cast_fp16 = mul(x = tensor_11_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_15_cast_fp16")]; tensor input_17_pad_0 = const()[name = tensor("input_17_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("constant")]; tensor const_41_to_fp16 = const()[name = tensor("const_41_to_fp16"), val = tensor(0x0p+0)]; tensor input_17_cast_fp16 = pad(constant_val = const_41_to_fp16, mode = input_17_mode_0, pad = input_17_pad_0, x = input_15_cast_fp16)[name = tensor("input_17_cast_fp16")]; tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("valid")]; tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([0, 0, 0, 0])]; tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; tensor encoder_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; tensor encoder_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; tensor tensor_13_cast_fp16 = conv(bias = encoder_pre_encode_conv_5_bias_to_fp16, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = encoder_pre_encode_conv_5_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("tensor_13_cast_fp16")]; tensor var_254_promoted_to_fp16 = const()[name = tensor("op_254_promoted_to_fp16"), val = tensor(0x1p+1)]; tensor var_255_cast_fp16 = add(x = current_lengths_5_cast_fp16, y = var_254_promoted_to_fp16)[name = tensor("op_255_cast_fp16")]; tensor var_256_promoted_to_fp16 = const()[name = tensor("op_256_promoted_to_fp16"), val = tensor(0x1p+0)]; tensor var_257_cast_fp16 = add(x = var_255_cast_fp16, y = var_256_promoted_to_fp16)[name = tensor("op_257_cast_fp16")]; tensor var_258_promoted_to_fp16 = const()[name = tensor("op_258_promoted_to_fp16"), val = tensor(0x1.8p+1)]; tensor var_259_cast_fp16 = sub(x = var_257_cast_fp16, y = var_258_promoted_to_fp16)[name = tensor("op_259_cast_fp16")]; tensor var_53_promoted_2_to_fp16 = const()[name = tensor("op_53_promoted_2_to_fp16"), val = tensor(0x1p+1)]; tensor floor_div_2_cast_fp16 = floor_div(x = var_259_cast_fp16, y = var_53_promoted_2_to_fp16)[name = tensor("floor_div_2_cast_fp16")]; tensor var_261_promoted_to_fp16 = const()[name = tensor("op_261_promoted_to_fp16"), val = tensor(0x1p+0)]; tensor current_lengths_cast_fp16 = add(x = floor_div_2_cast_fp16, y = var_261_promoted_to_fp16)[name = tensor("current_lengths_cast_fp16")]; tensor cast_7_dtype_0 = const()[name = tensor("cast_7_dtype_0"), val = tensor("int32")]; tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9]])]; tensor var_270_axes_0 = const()[name = tensor("op_270_axes_0"), val = tensor([1])]; tensor current_lengths_cast_fp16_to_int32 = cast(dtype = cast_7_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_191")]; tensor var_270 = expand_dims(axes = var_270_axes_0, x = current_lengths_cast_fp16_to_int32)[name = tensor("op_270")]; tensor time_mask = less(x = expand_dims_3, y = var_270)[name = tensor("time_mask")]; tensor var_272_axes_0 = const()[name = tensor("op_272_axes_0"), val = tensor([-1])]; tensor var_272 = expand_dims(axes = var_272_axes_0, x = time_mask)[name = tensor("op_272")]; tensor var_274_reps_0 = const()[name = tensor("op_274_reps_0"), val = tensor([1, 1, 17])]; tensor var_274 = tile(reps = var_274_reps_0, x = var_272)[name = tensor("op_274")]; tensor var_280_axes_0 = const()[name = tensor("op_280_axes_0"), val = tensor([1])]; tensor cast_8_to_fp16_dtype_0 = const()[name = tensor("cast_8_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_274_to_fp16 = cast(dtype = cast_8_to_fp16_dtype_0, x = var_274)[name = tensor("cast_190")]; tensor var_280_cast_fp16 = expand_dims(axes = var_280_axes_0, x = var_274_to_fp16)[name = tensor("op_280_cast_fp16")]; tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; tensor expanded_mask_13_cast_fp16 = tile(reps = expanded_mask_13_reps_0, x = var_280_cast_fp16)[name = tensor("expanded_mask_13_cast_fp16")]; tensor input_19_cast_fp16 = mul(x = tensor_13_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_19_cast_fp16")]; tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; tensor encoder_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; tensor encoder_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; tensor tensor_15_cast_fp16 = conv(bias = encoder_pre_encode_conv_6_bias_to_fp16, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = encoder_pre_encode_conv_6_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("tensor_15_cast_fp16")]; tensor input_21_cast_fp16 = mul(x = tensor_15_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_21_cast_fp16")]; tensor tensor_cast_fp16 = relu(x = input_21_cast_fp16)[name = tensor("tensor_cast_fp16")]; tensor x_3_cast_fp16 = mul(x = tensor_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("x_3_cast_fp16")]; tensor var_314_perm_0 = const()[name = tensor("op_314_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_315 = const()[name = tensor("op_315"), val = tensor([1, 10, -1])]; tensor var_314_cast_fp16 = transpose(perm = var_314_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_240")]; tensor input_23_cast_fp16 = reshape(shape = var_315, x = var_314_cast_fp16)[name = tensor("input_23_cast_fp16")]; tensor encoder_pre_encode_out_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; tensor encoder_pre_encode_out_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4735744)))]; tensor linear_0_cast_fp16 = linear(bias = encoder_pre_encode_out_bias_to_fp16, weight = encoder_pre_encode_out_weight_to_fp16, x = input_23_cast_fp16)[name = tensor("linear_0_cast_fp16")]; tensor var_325_begin_0 = const()[name = tensor("op_325_begin_0"), val = tensor([0, 2, 0])]; tensor var_325_end_0 = const()[name = tensor("op_325_end_0"), val = tensor([1, 10, 512])]; tensor var_325_end_mask_0 = const()[name = tensor("op_325_end_mask_0"), val = tensor([true, true, true])]; tensor var_325_cast_fp16 = slice_by_index(begin = var_325_begin_0, end = var_325_end_0, end_mask = var_325_end_mask_0, x = linear_0_cast_fp16)[name = tensor("op_325_cast_fp16")]; tensor var_327 = const()[name = tensor("op_327"), val = tensor(2)]; tensor var_328 = sub(x = current_lengths_cast_fp16_to_int32, y = var_327)[name = tensor("op_328")]; tensor var_328_promoted_to_fp16_dtype_0 = const()[name = tensor("op_328_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_59_promoted_to_fp16 = const()[name = tensor("op_59_promoted_to_fp16"), val = tensor(0x0p+0)]; tensor const_61_to_fp16 = const()[name = tensor("const_61_to_fp16"), val = tensor(inf)]; tensor var_328_to_fp16 = cast(dtype = var_328_promoted_to_fp16_dtype_0, x = var_328)[name = tensor("cast_189")]; tensor clip_0_cast_fp16 = clip(alpha = var_59_promoted_to_fp16, beta = const_61_to_fp16, x = var_328_to_fp16)[name = tensor("clip_0_cast_fp16")]; tensor cache_keep_size = const()[name = tensor("cache_keep_size"), val = tensor([4])]; tensor var_344_promoted_to_fp16 = const()[name = tensor("op_344_promoted_to_fp16"), val = tensor(0x1.18p+6)]; tensor padding_length_cast_fp16 = add(x = clip_0_cast_fp16, y = var_344_promoted_to_fp16)[name = tensor("padding_length_cast_fp16")]; tensor const_63 = const()[name = tensor("const_63"), val = tensor(-1)]; tensor var_346 = mul(x = cache_last_channel_len, y = const_63)[name = tensor("op_346")]; tensor var_347 = const()[name = tensor("op_347"), val = tensor(70)]; tensor offset = add(x = var_346, y = var_347)[name = tensor("offset")]; tensor var_387_axes_0 = const()[name = tensor("op_387_axes_0"), val = tensor([-1])]; tensor var_387_cast_fp16 = expand_dims(axes = var_387_axes_0, x = padding_length_cast_fp16)[name = tensor("op_387_cast_fp16")]; tensor var_386_promoted_to_fp16 = const()[name = tensor("op_386_promoted_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736832)))]; tensor pad_mask_1_cast_fp16 = less(x = var_386_promoted_to_fp16, y = var_387_cast_fp16)[name = tensor("pad_mask_1_cast_fp16")]; tensor expand_dims_5 = const()[name = tensor("expand_dims_5"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77]])]; tensor var_393_axes_0 = const()[name = tensor("op_393_axes_0"), val = tensor([-1])]; tensor var_393 = expand_dims(axes = var_393_axes_0, x = offset)[name = tensor("op_393")]; tensor pad_mask_off = greater_equal(x = expand_dims_5, y = var_393)[name = tensor("pad_mask_off")]; tensor pad_mask_3 = logical_and(x = pad_mask_off, y = pad_mask_1_cast_fp16)[name = tensor("pad_mask_3")]; tensor var_396_axes_0 = const()[name = tensor("op_396_axes_0"), val = tensor([1])]; tensor var_396 = expand_dims(axes = var_396_axes_0, x = pad_mask_3)[name = tensor("op_396")]; tensor var_397 = const()[name = tensor("op_397"), val = tensor([1, 78, 1])]; tensor pad_mask_for_att_mask_1 = tile(reps = var_397, x = var_396)[name = tensor("pad_mask_for_att_mask_1")]; tensor var_399_perm_0 = const()[name = tensor("op_399_perm_0"), val = tensor([0, 2, 1])]; tensor var_399 = transpose(perm = var_399_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_239")]; tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_399)[name = tensor("pad_mask_for_att_mask")]; tensor const_71 = const()[name = tensor("const_71"), val = tensor([[[true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; tensor att_mask_9 = logical_and(x = pad_mask_for_att_mask, y = const_71)[name = tensor("att_mask_9")]; tensor att_mask = logical_not(x = att_mask_9)[name = tensor("att_mask")]; tensor pad_mask_5 = logical_not(x = pad_mask_3)[name = tensor("pad_mask_5")]; tensor pad_mask_begin_0 = const()[name = tensor("pad_mask_begin_0"), val = tensor([0, 70])]; tensor pad_mask_end_0 = const()[name = tensor("pad_mask_end_0"), val = tensor([1, 78])]; tensor pad_mask_end_mask_0 = const()[name = tensor("pad_mask_end_mask_0"), val = tensor([true, true])]; tensor pad_mask = slice_by_index(begin = pad_mask_begin_0, end = pad_mask_end_0, end_mask = pad_mask_end_mask_0, x = pad_mask_5)[name = tensor("pad_mask")]; tensor mask_9_begin_0 = const()[name = tensor("mask_9_begin_0"), val = tensor([0, 70, 0])]; tensor mask_9_end_0 = const()[name = tensor("mask_9_end_0"), val = tensor([1, 78, 78])]; tensor mask_9_end_mask_0 = const()[name = tensor("mask_9_end_mask_0"), val = tensor([true, true, true])]; tensor mask_9 = slice_by_index(begin = mask_9_begin_0, end = mask_9_end_0, end_mask = mask_9_end_mask_0, x = att_mask)[name = tensor("mask_9")]; tensor cache_1_begin_0 = const()[name = tensor("cache_1_begin_0"), val = tensor([0, 0, 0, 0])]; tensor cache_1_end_0 = const()[name = tensor("cache_1_end_0"), val = tensor([1, 1, 70, 512])]; tensor cache_1_end_mask_0 = const()[name = tensor("cache_1_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_1_squeeze_mask_0 = const()[name = tensor("cache_1_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_last_channel_to_fp16_dtype_0 = const()[name = tensor("cache_last_channel_to_fp16_dtype_0"), val = tensor("fp16")]; tensor cache_last_channel_to_fp16 = cast(dtype = cache_last_channel_to_fp16_dtype_0, x = cache_last_channel)[name = tensor("cast_188")]; tensor cache_1_cast_fp16 = slice_by_index(begin = cache_1_begin_0, end = cache_1_end_0, end_mask = cache_1_end_mask_0, squeeze_mask = cache_1_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_1_cast_fp16")]; tensor cache_3_begin_0 = const()[name = tensor("cache_3_begin_0"), val = tensor([0, 0, 0, 0])]; tensor cache_3_end_0 = const()[name = tensor("cache_3_end_0"), val = tensor([1, 1, 512, 8])]; tensor cache_3_end_mask_0 = const()[name = tensor("cache_3_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_3_squeeze_mask_0 = const()[name = tensor("cache_3_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_last_time_to_fp16_dtype_0 = const()[name = tensor("cache_last_time_to_fp16_dtype_0"), val = tensor("fp16")]; tensor cache_last_time_to_fp16 = cast(dtype = cache_last_time_to_fp16_dtype_0, x = cache_last_time)[name = tensor("cast_187")]; tensor cache_3_cast_fp16 = slice_by_index(begin = cache_3_begin_0, end = cache_3_end_0, end_mask = cache_3_end_mask_0, squeeze_mask = cache_3_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_3_cast_fp16")]; tensor input_27_axes_0 = const()[name = tensor("input_27_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4737088)))]; tensor encoder_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4738176)))]; tensor var_38_to_fp16 = const()[name = tensor("op_38_to_fp16"), val = tensor(0x1.5p-17)]; tensor input_27_cast_fp16 = layer_norm(axes = input_27_axes_0, beta = encoder_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward1_weight_to_fp16, x = var_325_cast_fp16)[name = tensor("input_27_cast_fp16")]; tensor encoder_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4739264)))]; tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6836480)))]; tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("linear_1_cast_fp16")]; tensor input_31_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_31_cast_fp16")]; tensor encoder_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6840640)))]; tensor linear_2_bias_0_to_fp16 = const()[name = tensor("linear_2_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8937856)))]; tensor linear_2_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("linear_2_cast_fp16")]; tensor var_436_to_fp16 = const()[name = tensor("op_436_to_fp16"), val = tensor(0x1p-1)]; tensor var_437_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_436_to_fp16)[name = tensor("op_437_cast_fp16")]; tensor input_37_cast_fp16 = add(x = var_325_cast_fp16, y = var_437_cast_fp16)[name = tensor("input_37_cast_fp16")]; tensor key_1_axes_0 = const()[name = tensor("key_1_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8938944)))]; tensor encoder_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8940032)))]; tensor key_1_cast_fp16 = layer_norm(axes = key_1_axes_0, beta = encoder_layers_0_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_self_att_weight_to_fp16, x = input_37_cast_fp16)[name = tensor("key_1_cast_fp16")]; tensor input_39_interleave_0 = const()[name = tensor("input_39_interleave_0"), val = tensor(false)]; tensor input_39_cast_fp16 = concat(axis = var_65, interleave = input_39_interleave_0, values = (cache_1_cast_fp16, key_1_cast_fp16))[name = tensor("input_39_cast_fp16")]; tensor var_459_begin_0 = const()[name = tensor("op_459_begin_0"), val = tensor([0, 4, 0])]; tensor var_459_end_0 = const()[name = tensor("op_459_end_0"), val = tensor([1, 70, 512])]; tensor var_459_end_mask_0 = const()[name = tensor("op_459_end_mask_0"), val = tensor([true, true, true])]; tensor var_459_cast_fp16 = slice_by_index(begin = var_459_begin_0, end = var_459_end_0, end_mask = var_459_end_mask_0, x = cache_1_cast_fp16)[name = tensor("op_459_cast_fp16")]; tensor var_462_begin_0 = const()[name = tensor("op_462_begin_0"), val = tensor([0, 0, 0])]; tensor var_462_end_0 = const()[name = tensor("op_462_end_0"), val = tensor([1, 4, 512])]; tensor var_462_end_mask_0 = const()[name = tensor("op_462_end_mask_0"), val = tensor([true, false, true])]; tensor var_462_cast_fp16 = slice_by_index(begin = var_462_begin_0, end = var_462_end_0, end_mask = var_462_end_mask_0, x = key_1_cast_fp16)[name = tensor("op_462_cast_fp16")]; tensor var_465_interleave_0 = const()[name = tensor("op_465_interleave_0"), val = tensor(false)]; tensor var_465_cast_fp16 = concat(axis = var_65, interleave = var_465_interleave_0, values = (var_459_cast_fp16, var_462_cast_fp16))[name = tensor("op_465_cast_fp16")]; tensor encoder_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8941120)))]; tensor linear_3_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_q_weight_to_fp16, x = key_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; tensor var_469 = const()[name = tensor("op_469"), val = tensor([1, -1, 8, 64])]; tensor q_1_cast_fp16 = reshape(shape = var_469, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; tensor encoder_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9465472)))]; tensor linear_4_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_k_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_4_cast_fp16")]; tensor var_473 = const()[name = tensor("op_473"), val = tensor([1, -1, 8, 64])]; tensor k_1_cast_fp16 = reshape(shape = var_473, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; tensor encoder_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9989824)))]; tensor linear_5_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_v_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_5_cast_fp16")]; tensor var_477 = const()[name = tensor("op_477"), val = tensor([1, -1, 8, 64])]; tensor v_1_cast_fp16 = reshape(shape = var_477, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10514176)))]; tensor var_489_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_489_cast_fp16")]; tensor encoder_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10515264)))]; tensor var_491_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_491_cast_fp16")]; tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; tensor var_493_to_fp16 = const()[name = tensor("op_493_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10516352)))]; tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_491_cast_fp16)[name = tensor("transpose_237")]; tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_493_to_fp16)[name = tensor("x_7_cast_fp16")]; tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; tensor const_79_to_fp16 = const()[name = tensor("const_79_to_fp16"), val = tensor(0x0p+0)]; tensor x_9_cast_fp16 = pad(constant_val = const_79_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; tensor var_501 = const()[name = tensor("op_501"), val = tensor([1, 8, -1, 8])]; tensor x_11_cast_fp16 = reshape(shape = var_501, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; tensor var_505_begin_0 = const()[name = tensor("op_505_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_505_end_0 = const()[name = tensor("op_505_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_505_end_mask_0 = const()[name = tensor("op_505_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_505_cast_fp16 = slice_by_index(begin = var_505_begin_0, end = var_505_end_0, end_mask = var_505_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_505_cast_fp16")]; tensor var_506 = const()[name = tensor("op_506"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_1_cast_fp16 = reshape(shape = var_506, x = var_505_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_235")]; tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_489_cast_fp16)[name = tensor("transpose_236")]; tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; tensor var_515_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_515_cast_fp16")]; tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_1_cast_fp16 = mul(x = var_515_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; tensor var_41_to_fp16 = const()[name = tensor("op_41_to_fp16"), val = tensor(-0x1.388p+13)]; tensor scores_3_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_11)[name = tensor("scores_3_cast_fp16")]; tensor var_521_cast_fp16 = softmax(axis = var_56, x = scores_3_cast_fp16)[name = tensor("op_521_cast_fp16")]; tensor var_40_to_fp16 = const()[name = tensor("op_40_to_fp16"), val = tensor(0x0p+0)]; tensor input_41_cast_fp16 = select(a = var_40_to_fp16, b = var_521_cast_fp16, cond = mask_11)[name = tensor("input_41_cast_fp16")]; tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_238")]; tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_41_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; tensor var_525_perm_0 = const()[name = tensor("op_525_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_526 = const()[name = tensor("op_526"), val = tensor([1, -1, 512])]; tensor var_525_cast_fp16 = transpose(perm = var_525_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_234")]; tensor input_43_cast_fp16 = reshape(shape = var_526, x = var_525_cast_fp16)[name = tensor("input_43_cast_fp16")]; tensor encoder_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10675136)))]; tensor linear_7_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_out_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("linear_7_cast_fp16")]; tensor input_47_cast_fp16 = add(x = input_37_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_47_cast_fp16")]; tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11199488)))]; tensor encoder_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11200576)))]; tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = encoder_layers_0_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_conv_weight_to_fp16, x = input_47_cast_fp16)[name = tensor("x_17_cast_fp16")]; tensor input_49_perm_0 = const()[name = tensor("input_49_perm_0"), val = tensor([0, 2, 1])]; tensor input_51_pad_type_0 = const()[name = tensor("input_51_pad_type_0"), val = tensor("valid")]; tensor input_51_strides_0 = const()[name = tensor("input_51_strides_0"), val = tensor([1])]; tensor input_51_pad_0 = const()[name = tensor("input_51_pad_0"), val = tensor([0, 0])]; tensor input_51_dilations_0 = const()[name = tensor("input_51_dilations_0"), val = tensor([1])]; tensor input_51_groups_0 = const()[name = tensor("input_51_groups_0"), val = tensor(1)]; tensor encoder_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11201664)))]; tensor input_49_cast_fp16 = transpose(perm = input_49_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_233")]; tensor input_51_cast_fp16 = conv(dilations = input_51_dilations_0, groups = input_51_groups_0, pad = input_51_pad_0, pad_type = input_51_pad_type_0, strides = input_51_strides_0, weight = encoder_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_49_cast_fp16)[name = tensor("input_51_cast_fp16")]; tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_51_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; tensor var_551_axes_0 = const()[name = tensor("op_551_axes_0"), val = tensor([1])]; tensor var_551 = expand_dims(axes = var_551_axes_0, x = pad_mask)[name = tensor("op_551")]; tensor input_53_cast_fp16 = select(a = var_40_to_fp16, b = x_19_cast_fp16, cond = var_551)[name = tensor("input_53_cast_fp16")]; tensor new_x_3_interleave_0 = const()[name = tensor("new_x_3_interleave_0"), val = tensor(false)]; tensor new_x_3_cast_fp16 = concat(axis = var_56, interleave = new_x_3_interleave_0, values = (cache_3_cast_fp16, input_53_cast_fp16))[name = tensor("new_x_3_cast_fp16")]; tensor next_cache_1_begin_0 = const()[name = tensor("next_cache_1_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_1_end_0 = const()[name = tensor("next_cache_1_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_1_end_mask_0 = const()[name = tensor("next_cache_1_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_1_cast_fp16 = slice_by_index(begin = next_cache_1_begin_0, end = next_cache_1_end_0, end_mask = next_cache_1_end_mask_0, x = new_x_3_cast_fp16)[name = tensor("next_cache_1_cast_fp16")]; tensor var_567_begin_0 = const()[name = tensor("op_567_begin_0"), val = tensor([0, 0, 4])]; tensor var_567_end_0 = const()[name = tensor("op_567_end_0"), val = tensor([1, 512, 12])]; tensor var_567_end_mask_0 = const()[name = tensor("op_567_end_mask_0"), val = tensor([true, true, true])]; tensor var_567_cast_fp16 = slice_by_index(begin = var_567_begin_0, end = var_567_end_0, end_mask = var_567_end_mask_0, x = next_cache_1_cast_fp16)[name = tensor("op_567_cast_fp16")]; tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(512)]; tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; tensor encoder_layers_0_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12250304)))]; tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = encoder_layers_0_conv_depthwise_conv_weight_to_fp16, x = new_x_3_cast_fp16)[name = tensor("x_21_cast_fp16")]; tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; tensor x_23_axes_0 = const()[name = tensor("x_23_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12259584)))]; tensor encoder_layers_0_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12260672)))]; tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_232")]; tensor x_23_cast_fp16 = layer_norm(axes = x_23_axes_0, beta = encoder_layers_0_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_conv_batch_norm_weight_to_fp16, x = input_55_cast_fp16)[name = tensor("x_23_cast_fp16")]; tensor input_57_perm_0 = const()[name = tensor("input_57_perm_0"), val = tensor([0, 2, 1])]; tensor input_57_cast_fp16 = transpose(perm = input_57_perm_0, x = x_23_cast_fp16)[name = tensor("transpose_231")]; tensor input_59_cast_fp16 = silu(x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; tensor x_25_pad_type_0 = const()[name = tensor("x_25_pad_type_0"), val = tensor("valid")]; tensor x_25_strides_0 = const()[name = tensor("x_25_strides_0"), val = tensor([1])]; tensor x_25_pad_0 = const()[name = tensor("x_25_pad_0"), val = tensor([0, 0])]; tensor x_25_dilations_0 = const()[name = tensor("x_25_dilations_0"), val = tensor([1])]; tensor x_25_groups_0 = const()[name = tensor("x_25_groups_0"), val = tensor(1)]; tensor encoder_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12261760)))]; tensor x_25_cast_fp16 = conv(dilations = x_25_dilations_0, groups = x_25_groups_0, pad = x_25_pad_0, pad_type = x_25_pad_type_0, strides = x_25_strides_0, weight = encoder_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("x_25_cast_fp16")]; tensor input_61_perm_0 = const()[name = tensor("input_61_perm_0"), val = tensor([0, 2, 1])]; tensor input_61_cast_fp16 = transpose(perm = input_61_perm_0, x = x_25_cast_fp16)[name = tensor("transpose_230")]; tensor input_63_cast_fp16 = add(x = input_47_cast_fp16, y = input_61_cast_fp16)[name = tensor("input_63_cast_fp16")]; tensor input_65_axes_0 = const()[name = tensor("input_65_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12786112)))]; tensor encoder_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12787200)))]; tensor input_65_cast_fp16 = layer_norm(axes = input_65_axes_0, beta = encoder_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("input_65_cast_fp16")]; tensor encoder_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12788288)))]; tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("linear_8_cast_fp16")]; tensor input_69_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_69_cast_fp16")]; tensor encoder_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14885504)))]; tensor linear_9_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("linear_9_cast_fp16")]; tensor var_608_to_fp16 = const()[name = tensor("op_608_to_fp16"), val = tensor(0x1p-1)]; tensor var_609_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_608_to_fp16)[name = tensor("op_609_cast_fp16")]; tensor input_75_cast_fp16 = add(x = input_63_cast_fp16, y = var_609_cast_fp16)[name = tensor("input_75_cast_fp16")]; tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16982720)))]; tensor encoder_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16983808)))]; tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = encoder_layers_0_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_out_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; tensor cache_5_begin_0 = const()[name = tensor("cache_5_begin_0"), val = tensor([1, 0, 0, 0])]; tensor cache_5_end_0 = const()[name = tensor("cache_5_end_0"), val = tensor([2, 1, 70, 512])]; tensor cache_5_end_mask_0 = const()[name = tensor("cache_5_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_5_squeeze_mask_0 = const()[name = tensor("cache_5_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_5_cast_fp16 = slice_by_index(begin = cache_5_begin_0, end = cache_5_end_0, end_mask = cache_5_end_mask_0, squeeze_mask = cache_5_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_5_cast_fp16")]; tensor cache_7_begin_0 = const()[name = tensor("cache_7_begin_0"), val = tensor([1, 0, 0, 0])]; tensor cache_7_end_0 = const()[name = tensor("cache_7_end_0"), val = tensor([2, 1, 512, 8])]; tensor cache_7_end_mask_0 = const()[name = tensor("cache_7_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_7_squeeze_mask_0 = const()[name = tensor("cache_7_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_7_cast_fp16 = slice_by_index(begin = cache_7_begin_0, end = cache_7_end_0, end_mask = cache_7_end_mask_0, squeeze_mask = cache_7_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_7_cast_fp16")]; tensor input_79_axes_0 = const()[name = tensor("input_79_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16984896)))]; tensor encoder_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16985984)))]; tensor input_79_cast_fp16 = layer_norm(axes = input_79_axes_0, beta = encoder_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward1_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("input_79_cast_fp16")]; tensor encoder_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16987072)))]; tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_79_cast_fp16)[name = tensor("linear_10_cast_fp16")]; tensor input_83_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_83_cast_fp16")]; tensor encoder_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19084288)))]; tensor linear_11_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("linear_11_cast_fp16")]; tensor var_643_to_fp16 = const()[name = tensor("op_643_to_fp16"), val = tensor(0x1p-1)]; tensor var_644_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_643_to_fp16)[name = tensor("op_644_cast_fp16")]; tensor input_89_cast_fp16 = add(x = input_77_cast_fp16, y = var_644_cast_fp16)[name = tensor("input_89_cast_fp16")]; tensor key_3_axes_0 = const()[name = tensor("key_3_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21181504)))]; tensor encoder_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21182592)))]; tensor key_3_cast_fp16 = layer_norm(axes = key_3_axes_0, beta = encoder_layers_1_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_self_att_weight_to_fp16, x = input_89_cast_fp16)[name = tensor("key_3_cast_fp16")]; tensor input_91_interleave_0 = const()[name = tensor("input_91_interleave_0"), val = tensor(false)]; tensor input_91_cast_fp16 = concat(axis = var_65, interleave = input_91_interleave_0, values = (cache_5_cast_fp16, key_3_cast_fp16))[name = tensor("input_91_cast_fp16")]; tensor var_666_begin_0 = const()[name = tensor("op_666_begin_0"), val = tensor([0, 4, 0])]; tensor var_666_end_0 = const()[name = tensor("op_666_end_0"), val = tensor([1, 70, 512])]; tensor var_666_end_mask_0 = const()[name = tensor("op_666_end_mask_0"), val = tensor([true, true, true])]; tensor var_666_cast_fp16 = slice_by_index(begin = var_666_begin_0, end = var_666_end_0, end_mask = var_666_end_mask_0, x = cache_5_cast_fp16)[name = tensor("op_666_cast_fp16")]; tensor var_669_begin_0 = const()[name = tensor("op_669_begin_0"), val = tensor([0, 0, 0])]; tensor var_669_end_0 = const()[name = tensor("op_669_end_0"), val = tensor([1, 4, 512])]; tensor var_669_end_mask_0 = const()[name = tensor("op_669_end_mask_0"), val = tensor([true, false, true])]; tensor var_669_cast_fp16 = slice_by_index(begin = var_669_begin_0, end = var_669_end_0, end_mask = var_669_end_mask_0, x = key_3_cast_fp16)[name = tensor("op_669_cast_fp16")]; tensor var_672_interleave_0 = const()[name = tensor("op_672_interleave_0"), val = tensor(false)]; tensor var_672_cast_fp16 = concat(axis = var_65, interleave = var_672_interleave_0, values = (var_666_cast_fp16, var_669_cast_fp16))[name = tensor("op_672_cast_fp16")]; tensor encoder_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21183680)))]; tensor linear_12_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_q_weight_to_fp16, x = key_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; tensor var_676 = const()[name = tensor("op_676"), val = tensor([1, -1, 8, 64])]; tensor q_7_cast_fp16 = reshape(shape = var_676, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; tensor encoder_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21708032)))]; tensor linear_13_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_k_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_13_cast_fp16")]; tensor var_680 = const()[name = tensor("op_680"), val = tensor([1, -1, 8, 64])]; tensor k_5_cast_fp16 = reshape(shape = var_680, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; tensor encoder_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22232384)))]; tensor linear_14_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_v_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_14_cast_fp16")]; tensor var_684 = const()[name = tensor("op_684"), val = tensor([1, -1, 8, 64])]; tensor v_3_cast_fp16 = reshape(shape = var_684, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22756736)))]; tensor var_696_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_696_cast_fp16")]; tensor encoder_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22757824)))]; tensor var_698_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_698_cast_fp16")]; tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_33_transpose_x_0 = const()[name = tensor("x_33_transpose_x_0"), val = tensor(false)]; tensor x_33_transpose_y_0 = const()[name = tensor("x_33_transpose_y_0"), val = tensor(false)]; tensor var_700_to_fp16 = const()[name = tensor("op_700_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22758912)))]; tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_698_cast_fp16)[name = tensor("transpose_228")]; tensor x_33_cast_fp16 = matmul(transpose_x = x_33_transpose_x_0, transpose_y = x_33_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_700_to_fp16)[name = tensor("x_33_cast_fp16")]; tensor x_35_pad_0 = const()[name = tensor("x_35_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_35_mode_0 = const()[name = tensor("x_35_mode_0"), val = tensor("constant")]; tensor const_92_to_fp16 = const()[name = tensor("const_92_to_fp16"), val = tensor(0x0p+0)]; tensor x_35_cast_fp16 = pad(constant_val = const_92_to_fp16, mode = x_35_mode_0, pad = x_35_pad_0, x = x_33_cast_fp16)[name = tensor("x_35_cast_fp16")]; tensor var_708 = const()[name = tensor("op_708"), val = tensor([1, 8, -1, 8])]; tensor x_37_cast_fp16 = reshape(shape = var_708, x = x_35_cast_fp16)[name = tensor("x_37_cast_fp16")]; tensor var_712_begin_0 = const()[name = tensor("op_712_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_712_end_0 = const()[name = tensor("op_712_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_712_end_mask_0 = const()[name = tensor("op_712_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_712_cast_fp16 = slice_by_index(begin = var_712_begin_0, end = var_712_end_0, end_mask = var_712_end_mask_0, x = x_37_cast_fp16)[name = tensor("op_712_cast_fp16")]; tensor var_713 = const()[name = tensor("op_713"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_5_cast_fp16 = reshape(shape = var_713, x = var_712_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_226")]; tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_696_cast_fp16)[name = tensor("transpose_227")]; tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; tensor var_722_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_722_cast_fp16")]; tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_5_cast_fp16 = mul(x = var_722_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; tensor scores_7_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_11)[name = tensor("scores_7_cast_fp16")]; tensor var_728_cast_fp16 = softmax(axis = var_56, x = scores_7_cast_fp16)[name = tensor("op_728_cast_fp16")]; tensor input_93_cast_fp16 = select(a = var_40_to_fp16, b = var_728_cast_fp16, cond = mask_11)[name = tensor("input_93_cast_fp16")]; tensor x_39_transpose_x_0 = const()[name = tensor("x_39_transpose_x_0"), val = tensor(false)]; tensor x_39_transpose_y_0 = const()[name = tensor("x_39_transpose_y_0"), val = tensor(false)]; tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_229")]; tensor x_39_cast_fp16 = matmul(transpose_x = x_39_transpose_x_0, transpose_y = x_39_transpose_y_0, x = input_93_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_39_cast_fp16")]; tensor var_732_perm_0 = const()[name = tensor("op_732_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_733 = const()[name = tensor("op_733"), val = tensor([1, -1, 512])]; tensor var_732_cast_fp16 = transpose(perm = var_732_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_225")]; tensor input_95_cast_fp16 = reshape(shape = var_733, x = var_732_cast_fp16)[name = tensor("input_95_cast_fp16")]; tensor encoder_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22917696)))]; tensor linear_16_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_out_weight_to_fp16, x = input_95_cast_fp16)[name = tensor("linear_16_cast_fp16")]; tensor input_99_cast_fp16 = add(x = input_89_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_99_cast_fp16")]; tensor x_43_axes_0 = const()[name = tensor("x_43_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23442048)))]; tensor encoder_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23443136)))]; tensor x_43_cast_fp16 = layer_norm(axes = x_43_axes_0, beta = encoder_layers_1_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_conv_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("x_43_cast_fp16")]; tensor input_101_perm_0 = const()[name = tensor("input_101_perm_0"), val = tensor([0, 2, 1])]; tensor input_103_pad_type_0 = const()[name = tensor("input_103_pad_type_0"), val = tensor("valid")]; tensor input_103_strides_0 = const()[name = tensor("input_103_strides_0"), val = tensor([1])]; tensor input_103_pad_0 = const()[name = tensor("input_103_pad_0"), val = tensor([0, 0])]; tensor input_103_dilations_0 = const()[name = tensor("input_103_dilations_0"), val = tensor([1])]; tensor input_103_groups_0 = const()[name = tensor("input_103_groups_0"), val = tensor(1)]; tensor encoder_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23444224)))]; tensor input_101_cast_fp16 = transpose(perm = input_101_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_224")]; tensor input_103_cast_fp16 = conv(dilations = input_103_dilations_0, groups = input_103_groups_0, pad = input_103_pad_0, pad_type = input_103_pad_type_0, strides = input_103_strides_0, weight = encoder_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; tensor x_45_split_num_splits_0 = const()[name = tensor("x_45_split_num_splits_0"), val = tensor(2)]; tensor x_45_split_axis_0 = const()[name = tensor("x_45_split_axis_0"), val = tensor(1)]; tensor x_45_split_cast_fp16_0, tensor x_45_split_cast_fp16_1 = split(axis = x_45_split_axis_0, num_splits = x_45_split_num_splits_0, x = input_103_cast_fp16)[name = tensor("x_45_split_cast_fp16")]; tensor x_45_split_1_sigmoid_cast_fp16 = sigmoid(x = x_45_split_cast_fp16_1)[name = tensor("x_45_split_1_sigmoid_cast_fp16")]; tensor x_45_cast_fp16 = mul(x = x_45_split_cast_fp16_0, y = x_45_split_1_sigmoid_cast_fp16)[name = tensor("x_45_cast_fp16")]; tensor input_105_cast_fp16 = select(a = var_40_to_fp16, b = x_45_cast_fp16, cond = var_551)[name = tensor("input_105_cast_fp16")]; tensor new_x_7_interleave_0 = const()[name = tensor("new_x_7_interleave_0"), val = tensor(false)]; tensor new_x_7_cast_fp16 = concat(axis = var_56, interleave = new_x_7_interleave_0, values = (cache_7_cast_fp16, input_105_cast_fp16))[name = tensor("new_x_7_cast_fp16")]; tensor next_cache_3_begin_0 = const()[name = tensor("next_cache_3_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_3_end_0 = const()[name = tensor("next_cache_3_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_3_end_mask_0 = const()[name = tensor("next_cache_3_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_3_cast_fp16 = slice_by_index(begin = next_cache_3_begin_0, end = next_cache_3_end_0, end_mask = next_cache_3_end_mask_0, x = new_x_7_cast_fp16)[name = tensor("next_cache_3_cast_fp16")]; tensor var_774_begin_0 = const()[name = tensor("op_774_begin_0"), val = tensor([0, 0, 4])]; tensor var_774_end_0 = const()[name = tensor("op_774_end_0"), val = tensor([1, 512, 12])]; tensor var_774_end_mask_0 = const()[name = tensor("op_774_end_mask_0"), val = tensor([true, true, true])]; tensor var_774_cast_fp16 = slice_by_index(begin = var_774_begin_0, end = var_774_end_0, end_mask = var_774_end_mask_0, x = next_cache_3_cast_fp16)[name = tensor("op_774_cast_fp16")]; tensor x_47_pad_type_0 = const()[name = tensor("x_47_pad_type_0"), val = tensor("valid")]; tensor x_47_groups_0 = const()[name = tensor("x_47_groups_0"), val = tensor(512)]; tensor x_47_strides_0 = const()[name = tensor("x_47_strides_0"), val = tensor([1])]; tensor x_47_pad_0 = const()[name = tensor("x_47_pad_0"), val = tensor([0, 0])]; tensor x_47_dilations_0 = const()[name = tensor("x_47_dilations_0"), val = tensor([1])]; tensor encoder_layers_1_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24492864)))]; tensor x_47_cast_fp16 = conv(dilations = x_47_dilations_0, groups = x_47_groups_0, pad = x_47_pad_0, pad_type = x_47_pad_type_0, strides = x_47_strides_0, weight = encoder_layers_1_conv_depthwise_conv_weight_to_fp16, x = new_x_7_cast_fp16)[name = tensor("x_47_cast_fp16")]; tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; tensor x_49_axes_0 = const()[name = tensor("x_49_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24502144)))]; tensor encoder_layers_1_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24503232)))]; tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_47_cast_fp16)[name = tensor("transpose_223")]; tensor x_49_cast_fp16 = layer_norm(axes = x_49_axes_0, beta = encoder_layers_1_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_conv_batch_norm_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("x_49_cast_fp16")]; tensor input_109_perm_0 = const()[name = tensor("input_109_perm_0"), val = tensor([0, 2, 1])]; tensor input_109_cast_fp16 = transpose(perm = input_109_perm_0, x = x_49_cast_fp16)[name = tensor("transpose_222")]; tensor input_111_cast_fp16 = silu(x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; tensor x_51_pad_type_0 = const()[name = tensor("x_51_pad_type_0"), val = tensor("valid")]; tensor x_51_strides_0 = const()[name = tensor("x_51_strides_0"), val = tensor([1])]; tensor x_51_pad_0 = const()[name = tensor("x_51_pad_0"), val = tensor([0, 0])]; tensor x_51_dilations_0 = const()[name = tensor("x_51_dilations_0"), val = tensor([1])]; tensor x_51_groups_0 = const()[name = tensor("x_51_groups_0"), val = tensor(1)]; tensor encoder_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24504320)))]; tensor x_51_cast_fp16 = conv(dilations = x_51_dilations_0, groups = x_51_groups_0, pad = x_51_pad_0, pad_type = x_51_pad_type_0, strides = x_51_strides_0, weight = encoder_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("x_51_cast_fp16")]; tensor input_113_perm_0 = const()[name = tensor("input_113_perm_0"), val = tensor([0, 2, 1])]; tensor input_113_cast_fp16 = transpose(perm = input_113_perm_0, x = x_51_cast_fp16)[name = tensor("transpose_221")]; tensor input_115_cast_fp16 = add(x = input_99_cast_fp16, y = input_113_cast_fp16)[name = tensor("input_115_cast_fp16")]; tensor input_117_axes_0 = const()[name = tensor("input_117_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25028672)))]; tensor encoder_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25029760)))]; tensor input_117_cast_fp16 = layer_norm(axes = input_117_axes_0, beta = encoder_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("input_117_cast_fp16")]; tensor encoder_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25030848)))]; tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_117_cast_fp16)[name = tensor("linear_17_cast_fp16")]; tensor input_121_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_121_cast_fp16")]; tensor encoder_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27128064)))]; tensor linear_18_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("linear_18_cast_fp16")]; tensor var_815_to_fp16 = const()[name = tensor("op_815_to_fp16"), val = tensor(0x1p-1)]; tensor var_816_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_815_to_fp16)[name = tensor("op_816_cast_fp16")]; tensor input_127_cast_fp16 = add(x = input_115_cast_fp16, y = var_816_cast_fp16)[name = tensor("input_127_cast_fp16")]; tensor input_129_axes_0 = const()[name = tensor("input_129_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29225280)))]; tensor encoder_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29226368)))]; tensor input_129_cast_fp16 = layer_norm(axes = input_129_axes_0, beta = encoder_layers_1_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_out_weight_to_fp16, x = input_127_cast_fp16)[name = tensor("input_129_cast_fp16")]; tensor cache_9_begin_0 = const()[name = tensor("cache_9_begin_0"), val = tensor([2, 0, 0, 0])]; tensor cache_9_end_0 = const()[name = tensor("cache_9_end_0"), val = tensor([3, 1, 70, 512])]; tensor cache_9_end_mask_0 = const()[name = tensor("cache_9_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_9_squeeze_mask_0 = const()[name = tensor("cache_9_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_9_cast_fp16 = slice_by_index(begin = cache_9_begin_0, end = cache_9_end_0, end_mask = cache_9_end_mask_0, squeeze_mask = cache_9_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_9_cast_fp16")]; tensor cache_11_begin_0 = const()[name = tensor("cache_11_begin_0"), val = tensor([2, 0, 0, 0])]; tensor cache_11_end_0 = const()[name = tensor("cache_11_end_0"), val = tensor([3, 1, 512, 8])]; tensor cache_11_end_mask_0 = const()[name = tensor("cache_11_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_11_squeeze_mask_0 = const()[name = tensor("cache_11_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_11_cast_fp16 = slice_by_index(begin = cache_11_begin_0, end = cache_11_end_0, end_mask = cache_11_end_mask_0, squeeze_mask = cache_11_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_11_cast_fp16")]; tensor input_131_axes_0 = const()[name = tensor("input_131_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29227456)))]; tensor encoder_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29228544)))]; tensor input_131_cast_fp16 = layer_norm(axes = input_131_axes_0, beta = encoder_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward1_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("input_131_cast_fp16")]; tensor encoder_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29229632)))]; tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("linear_19_cast_fp16")]; tensor input_135_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_135_cast_fp16")]; tensor encoder_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31326848)))]; tensor linear_20_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("linear_20_cast_fp16")]; tensor var_850_to_fp16 = const()[name = tensor("op_850_to_fp16"), val = tensor(0x1p-1)]; tensor var_851_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_850_to_fp16)[name = tensor("op_851_cast_fp16")]; tensor input_141_cast_fp16 = add(x = input_129_cast_fp16, y = var_851_cast_fp16)[name = tensor("input_141_cast_fp16")]; tensor key_5_axes_0 = const()[name = tensor("key_5_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33424064)))]; tensor encoder_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33425152)))]; tensor key_5_cast_fp16 = layer_norm(axes = key_5_axes_0, beta = encoder_layers_2_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_self_att_weight_to_fp16, x = input_141_cast_fp16)[name = tensor("key_5_cast_fp16")]; tensor input_143_interleave_0 = const()[name = tensor("input_143_interleave_0"), val = tensor(false)]; tensor input_143_cast_fp16 = concat(axis = var_65, interleave = input_143_interleave_0, values = (cache_9_cast_fp16, key_5_cast_fp16))[name = tensor("input_143_cast_fp16")]; tensor var_873_begin_0 = const()[name = tensor("op_873_begin_0"), val = tensor([0, 4, 0])]; tensor var_873_end_0 = const()[name = tensor("op_873_end_0"), val = tensor([1, 70, 512])]; tensor var_873_end_mask_0 = const()[name = tensor("op_873_end_mask_0"), val = tensor([true, true, true])]; tensor var_873_cast_fp16 = slice_by_index(begin = var_873_begin_0, end = var_873_end_0, end_mask = var_873_end_mask_0, x = cache_9_cast_fp16)[name = tensor("op_873_cast_fp16")]; tensor var_876_begin_0 = const()[name = tensor("op_876_begin_0"), val = tensor([0, 0, 0])]; tensor var_876_end_0 = const()[name = tensor("op_876_end_0"), val = tensor([1, 4, 512])]; tensor var_876_end_mask_0 = const()[name = tensor("op_876_end_mask_0"), val = tensor([true, false, true])]; tensor var_876_cast_fp16 = slice_by_index(begin = var_876_begin_0, end = var_876_end_0, end_mask = var_876_end_mask_0, x = key_5_cast_fp16)[name = tensor("op_876_cast_fp16")]; tensor var_879_interleave_0 = const()[name = tensor("op_879_interleave_0"), val = tensor(false)]; tensor var_879_cast_fp16 = concat(axis = var_65, interleave = var_879_interleave_0, values = (var_873_cast_fp16, var_876_cast_fp16))[name = tensor("op_879_cast_fp16")]; tensor encoder_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33426240)))]; tensor linear_21_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_q_weight_to_fp16, x = key_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; tensor var_883 = const()[name = tensor("op_883"), val = tensor([1, -1, 8, 64])]; tensor q_13_cast_fp16 = reshape(shape = var_883, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; tensor encoder_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33950592)))]; tensor linear_22_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_k_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_22_cast_fp16")]; tensor var_887 = const()[name = tensor("op_887"), val = tensor([1, -1, 8, 64])]; tensor k_9_cast_fp16 = reshape(shape = var_887, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; tensor encoder_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34474944)))]; tensor linear_23_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_v_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_23_cast_fp16")]; tensor var_891 = const()[name = tensor("op_891"), val = tensor([1, -1, 8, 64])]; tensor v_5_cast_fp16 = reshape(shape = var_891, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34999296)))]; tensor var_903_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_903_cast_fp16")]; tensor encoder_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35000384)))]; tensor var_905_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_905_cast_fp16")]; tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; tensor var_907_to_fp16 = const()[name = tensor("op_907_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35001472)))]; tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_905_cast_fp16)[name = tensor("transpose_219")]; tensor x_59_cast_fp16 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_907_to_fp16)[name = tensor("x_59_cast_fp16")]; tensor x_61_pad_0 = const()[name = tensor("x_61_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_61_mode_0 = const()[name = tensor("x_61_mode_0"), val = tensor("constant")]; tensor const_105_to_fp16 = const()[name = tensor("const_105_to_fp16"), val = tensor(0x0p+0)]; tensor x_61_cast_fp16 = pad(constant_val = const_105_to_fp16, mode = x_61_mode_0, pad = x_61_pad_0, x = x_59_cast_fp16)[name = tensor("x_61_cast_fp16")]; tensor var_915 = const()[name = tensor("op_915"), val = tensor([1, 8, -1, 8])]; tensor x_63_cast_fp16 = reshape(shape = var_915, x = x_61_cast_fp16)[name = tensor("x_63_cast_fp16")]; tensor var_919_begin_0 = const()[name = tensor("op_919_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_919_end_0 = const()[name = tensor("op_919_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_919_end_mask_0 = const()[name = tensor("op_919_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_919_cast_fp16 = slice_by_index(begin = var_919_begin_0, end = var_919_end_0, end_mask = var_919_end_mask_0, x = x_63_cast_fp16)[name = tensor("op_919_cast_fp16")]; tensor var_920 = const()[name = tensor("op_920"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_9_cast_fp16 = reshape(shape = var_920, x = var_919_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_217")]; tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_903_cast_fp16)[name = tensor("transpose_218")]; tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; tensor var_929_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_929_cast_fp16")]; tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_9_cast_fp16 = mul(x = var_929_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; tensor scores_11_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_11)[name = tensor("scores_11_cast_fp16")]; tensor var_935_cast_fp16 = softmax(axis = var_56, x = scores_11_cast_fp16)[name = tensor("op_935_cast_fp16")]; tensor input_145_cast_fp16 = select(a = var_40_to_fp16, b = var_935_cast_fp16, cond = mask_11)[name = tensor("input_145_cast_fp16")]; tensor x_65_transpose_x_0 = const()[name = tensor("x_65_transpose_x_0"), val = tensor(false)]; tensor x_65_transpose_y_0 = const()[name = tensor("x_65_transpose_y_0"), val = tensor(false)]; tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_220")]; tensor x_65_cast_fp16 = matmul(transpose_x = x_65_transpose_x_0, transpose_y = x_65_transpose_y_0, x = input_145_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_65_cast_fp16")]; tensor var_939_perm_0 = const()[name = tensor("op_939_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_940 = const()[name = tensor("op_940"), val = tensor([1, -1, 512])]; tensor var_939_cast_fp16 = transpose(perm = var_939_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_216")]; tensor input_147_cast_fp16 = reshape(shape = var_940, x = var_939_cast_fp16)[name = tensor("input_147_cast_fp16")]; tensor encoder_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35160256)))]; tensor linear_25_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_out_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("linear_25_cast_fp16")]; tensor input_151_cast_fp16 = add(x = input_141_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_151_cast_fp16")]; tensor x_69_axes_0 = const()[name = tensor("x_69_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35684608)))]; tensor encoder_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35685696)))]; tensor x_69_cast_fp16 = layer_norm(axes = x_69_axes_0, beta = encoder_layers_2_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_conv_weight_to_fp16, x = input_151_cast_fp16)[name = tensor("x_69_cast_fp16")]; tensor input_153_perm_0 = const()[name = tensor("input_153_perm_0"), val = tensor([0, 2, 1])]; tensor input_155_pad_type_0 = const()[name = tensor("input_155_pad_type_0"), val = tensor("valid")]; tensor input_155_strides_0 = const()[name = tensor("input_155_strides_0"), val = tensor([1])]; tensor input_155_pad_0 = const()[name = tensor("input_155_pad_0"), val = tensor([0, 0])]; tensor input_155_dilations_0 = const()[name = tensor("input_155_dilations_0"), val = tensor([1])]; tensor input_155_groups_0 = const()[name = tensor("input_155_groups_0"), val = tensor(1)]; tensor encoder_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35686784)))]; tensor input_153_cast_fp16 = transpose(perm = input_153_perm_0, x = x_69_cast_fp16)[name = tensor("transpose_215")]; tensor input_155_cast_fp16 = conv(dilations = input_155_dilations_0, groups = input_155_groups_0, pad = input_155_pad_0, pad_type = input_155_pad_type_0, strides = input_155_strides_0, weight = encoder_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_153_cast_fp16)[name = tensor("input_155_cast_fp16")]; tensor x_71_split_num_splits_0 = const()[name = tensor("x_71_split_num_splits_0"), val = tensor(2)]; tensor x_71_split_axis_0 = const()[name = tensor("x_71_split_axis_0"), val = tensor(1)]; tensor x_71_split_cast_fp16_0, tensor x_71_split_cast_fp16_1 = split(axis = x_71_split_axis_0, num_splits = x_71_split_num_splits_0, x = input_155_cast_fp16)[name = tensor("x_71_split_cast_fp16")]; tensor x_71_split_1_sigmoid_cast_fp16 = sigmoid(x = x_71_split_cast_fp16_1)[name = tensor("x_71_split_1_sigmoid_cast_fp16")]; tensor x_71_cast_fp16 = mul(x = x_71_split_cast_fp16_0, y = x_71_split_1_sigmoid_cast_fp16)[name = tensor("x_71_cast_fp16")]; tensor input_157_cast_fp16 = select(a = var_40_to_fp16, b = x_71_cast_fp16, cond = var_551)[name = tensor("input_157_cast_fp16")]; tensor new_x_11_interleave_0 = const()[name = tensor("new_x_11_interleave_0"), val = tensor(false)]; tensor new_x_11_cast_fp16 = concat(axis = var_56, interleave = new_x_11_interleave_0, values = (cache_11_cast_fp16, input_157_cast_fp16))[name = tensor("new_x_11_cast_fp16")]; tensor next_cache_5_begin_0 = const()[name = tensor("next_cache_5_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_5_end_0 = const()[name = tensor("next_cache_5_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_5_end_mask_0 = const()[name = tensor("next_cache_5_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_5_cast_fp16 = slice_by_index(begin = next_cache_5_begin_0, end = next_cache_5_end_0, end_mask = next_cache_5_end_mask_0, x = new_x_11_cast_fp16)[name = tensor("next_cache_5_cast_fp16")]; tensor var_981_begin_0 = const()[name = tensor("op_981_begin_0"), val = tensor([0, 0, 4])]; tensor var_981_end_0 = const()[name = tensor("op_981_end_0"), val = tensor([1, 512, 12])]; tensor var_981_end_mask_0 = const()[name = tensor("op_981_end_mask_0"), val = tensor([true, true, true])]; tensor var_981_cast_fp16 = slice_by_index(begin = var_981_begin_0, end = var_981_end_0, end_mask = var_981_end_mask_0, x = next_cache_5_cast_fp16)[name = tensor("op_981_cast_fp16")]; tensor x_73_pad_type_0 = const()[name = tensor("x_73_pad_type_0"), val = tensor("valid")]; tensor x_73_groups_0 = const()[name = tensor("x_73_groups_0"), val = tensor(512)]; tensor x_73_strides_0 = const()[name = tensor("x_73_strides_0"), val = tensor([1])]; tensor x_73_pad_0 = const()[name = tensor("x_73_pad_0"), val = tensor([0, 0])]; tensor x_73_dilations_0 = const()[name = tensor("x_73_dilations_0"), val = tensor([1])]; tensor encoder_layers_2_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36735424)))]; tensor x_73_cast_fp16 = conv(dilations = x_73_dilations_0, groups = x_73_groups_0, pad = x_73_pad_0, pad_type = x_73_pad_type_0, strides = x_73_strides_0, weight = encoder_layers_2_conv_depthwise_conv_weight_to_fp16, x = new_x_11_cast_fp16)[name = tensor("x_73_cast_fp16")]; tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; tensor x_75_axes_0 = const()[name = tensor("x_75_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36744704)))]; tensor encoder_layers_2_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36745792)))]; tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_73_cast_fp16)[name = tensor("transpose_214")]; tensor x_75_cast_fp16 = layer_norm(axes = x_75_axes_0, beta = encoder_layers_2_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_conv_batch_norm_weight_to_fp16, x = input_159_cast_fp16)[name = tensor("x_75_cast_fp16")]; tensor input_161_perm_0 = const()[name = tensor("input_161_perm_0"), val = tensor([0, 2, 1])]; tensor input_161_cast_fp16 = transpose(perm = input_161_perm_0, x = x_75_cast_fp16)[name = tensor("transpose_213")]; tensor input_163_cast_fp16 = silu(x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; tensor x_77_pad_type_0 = const()[name = tensor("x_77_pad_type_0"), val = tensor("valid")]; tensor x_77_strides_0 = const()[name = tensor("x_77_strides_0"), val = tensor([1])]; tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0])]; tensor x_77_dilations_0 = const()[name = tensor("x_77_dilations_0"), val = tensor([1])]; tensor x_77_groups_0 = const()[name = tensor("x_77_groups_0"), val = tensor(1)]; tensor encoder_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36746880)))]; tensor x_77_cast_fp16 = conv(dilations = x_77_dilations_0, groups = x_77_groups_0, pad = x_77_pad_0, pad_type = x_77_pad_type_0, strides = x_77_strides_0, weight = encoder_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("x_77_cast_fp16")]; tensor input_165_perm_0 = const()[name = tensor("input_165_perm_0"), val = tensor([0, 2, 1])]; tensor input_165_cast_fp16 = transpose(perm = input_165_perm_0, x = x_77_cast_fp16)[name = tensor("transpose_212")]; tensor input_167_cast_fp16 = add(x = input_151_cast_fp16, y = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; tensor input_169_axes_0 = const()[name = tensor("input_169_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37271232)))]; tensor encoder_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37272320)))]; tensor input_169_cast_fp16 = layer_norm(axes = input_169_axes_0, beta = encoder_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("input_169_cast_fp16")]; tensor encoder_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37273408)))]; tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("linear_26_cast_fp16")]; tensor input_173_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_173_cast_fp16")]; tensor encoder_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39370624)))]; tensor linear_27_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("linear_27_cast_fp16")]; tensor var_1022_to_fp16 = const()[name = tensor("op_1022_to_fp16"), val = tensor(0x1p-1)]; tensor var_1023_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_1022_to_fp16)[name = tensor("op_1023_cast_fp16")]; tensor input_179_cast_fp16 = add(x = input_167_cast_fp16, y = var_1023_cast_fp16)[name = tensor("input_179_cast_fp16")]; tensor input_181_axes_0 = const()[name = tensor("input_181_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41467840)))]; tensor encoder_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41468928)))]; tensor input_181_cast_fp16 = layer_norm(axes = input_181_axes_0, beta = encoder_layers_2_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_out_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("input_181_cast_fp16")]; tensor cache_13_begin_0 = const()[name = tensor("cache_13_begin_0"), val = tensor([3, 0, 0, 0])]; tensor cache_13_end_0 = const()[name = tensor("cache_13_end_0"), val = tensor([4, 1, 70, 512])]; tensor cache_13_end_mask_0 = const()[name = tensor("cache_13_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_13_squeeze_mask_0 = const()[name = tensor("cache_13_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_13_cast_fp16 = slice_by_index(begin = cache_13_begin_0, end = cache_13_end_0, end_mask = cache_13_end_mask_0, squeeze_mask = cache_13_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_13_cast_fp16")]; tensor cache_15_begin_0 = const()[name = tensor("cache_15_begin_0"), val = tensor([3, 0, 0, 0])]; tensor cache_15_end_0 = const()[name = tensor("cache_15_end_0"), val = tensor([4, 1, 512, 8])]; tensor cache_15_end_mask_0 = const()[name = tensor("cache_15_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_15_squeeze_mask_0 = const()[name = tensor("cache_15_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_15_cast_fp16 = slice_by_index(begin = cache_15_begin_0, end = cache_15_end_0, end_mask = cache_15_end_mask_0, squeeze_mask = cache_15_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_15_cast_fp16")]; tensor input_183_axes_0 = const()[name = tensor("input_183_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41470016)))]; tensor encoder_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41471104)))]; tensor input_183_cast_fp16 = layer_norm(axes = input_183_axes_0, beta = encoder_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward1_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; tensor encoder_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41472192)))]; tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_183_cast_fp16)[name = tensor("linear_28_cast_fp16")]; tensor input_187_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_187_cast_fp16")]; tensor encoder_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43569408)))]; tensor linear_29_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("linear_29_cast_fp16")]; tensor var_1057_to_fp16 = const()[name = tensor("op_1057_to_fp16"), val = tensor(0x1p-1)]; tensor var_1058_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_1057_to_fp16)[name = tensor("op_1058_cast_fp16")]; tensor input_193_cast_fp16 = add(x = input_181_cast_fp16, y = var_1058_cast_fp16)[name = tensor("input_193_cast_fp16")]; tensor key_7_axes_0 = const()[name = tensor("key_7_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45666624)))]; tensor encoder_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45667712)))]; tensor key_7_cast_fp16 = layer_norm(axes = key_7_axes_0, beta = encoder_layers_3_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_self_att_weight_to_fp16, x = input_193_cast_fp16)[name = tensor("key_7_cast_fp16")]; tensor input_195_interleave_0 = const()[name = tensor("input_195_interleave_0"), val = tensor(false)]; tensor input_195_cast_fp16 = concat(axis = var_65, interleave = input_195_interleave_0, values = (cache_13_cast_fp16, key_7_cast_fp16))[name = tensor("input_195_cast_fp16")]; tensor var_1080_begin_0 = const()[name = tensor("op_1080_begin_0"), val = tensor([0, 4, 0])]; tensor var_1080_end_0 = const()[name = tensor("op_1080_end_0"), val = tensor([1, 70, 512])]; tensor var_1080_end_mask_0 = const()[name = tensor("op_1080_end_mask_0"), val = tensor([true, true, true])]; tensor var_1080_cast_fp16 = slice_by_index(begin = var_1080_begin_0, end = var_1080_end_0, end_mask = var_1080_end_mask_0, x = cache_13_cast_fp16)[name = tensor("op_1080_cast_fp16")]; tensor var_1083_begin_0 = const()[name = tensor("op_1083_begin_0"), val = tensor([0, 0, 0])]; tensor var_1083_end_0 = const()[name = tensor("op_1083_end_0"), val = tensor([1, 4, 512])]; tensor var_1083_end_mask_0 = const()[name = tensor("op_1083_end_mask_0"), val = tensor([true, false, true])]; tensor var_1083_cast_fp16 = slice_by_index(begin = var_1083_begin_0, end = var_1083_end_0, end_mask = var_1083_end_mask_0, x = key_7_cast_fp16)[name = tensor("op_1083_cast_fp16")]; tensor var_1086_interleave_0 = const()[name = tensor("op_1086_interleave_0"), val = tensor(false)]; tensor var_1086_cast_fp16 = concat(axis = var_65, interleave = var_1086_interleave_0, values = (var_1080_cast_fp16, var_1083_cast_fp16))[name = tensor("op_1086_cast_fp16")]; tensor encoder_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45668800)))]; tensor linear_30_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_q_weight_to_fp16, x = key_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; tensor var_1090 = const()[name = tensor("op_1090"), val = tensor([1, -1, 8, 64])]; tensor q_19_cast_fp16 = reshape(shape = var_1090, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; tensor encoder_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46193152)))]; tensor linear_31_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_k_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_31_cast_fp16")]; tensor var_1094 = const()[name = tensor("op_1094"), val = tensor([1, -1, 8, 64])]; tensor k_13_cast_fp16 = reshape(shape = var_1094, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; tensor encoder_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46717504)))]; tensor linear_32_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_v_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_32_cast_fp16")]; tensor var_1098 = const()[name = tensor("op_1098"), val = tensor([1, -1, 8, 64])]; tensor v_7_cast_fp16 = reshape(shape = var_1098, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47241856)))]; tensor var_1110_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1110_cast_fp16")]; tensor encoder_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47242944)))]; tensor var_1112_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1112_cast_fp16")]; tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_85_transpose_x_0 = const()[name = tensor("x_85_transpose_x_0"), val = tensor(false)]; tensor x_85_transpose_y_0 = const()[name = tensor("x_85_transpose_y_0"), val = tensor(false)]; tensor var_1114_to_fp16 = const()[name = tensor("op_1114_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47244032)))]; tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_1112_cast_fp16)[name = tensor("transpose_210")]; tensor x_85_cast_fp16 = matmul(transpose_x = x_85_transpose_x_0, transpose_y = x_85_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_1114_to_fp16)[name = tensor("x_85_cast_fp16")]; tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_87_mode_0 = const()[name = tensor("x_87_mode_0"), val = tensor("constant")]; tensor const_118_to_fp16 = const()[name = tensor("const_118_to_fp16"), val = tensor(0x0p+0)]; tensor x_87_cast_fp16 = pad(constant_val = const_118_to_fp16, mode = x_87_mode_0, pad = x_87_pad_0, x = x_85_cast_fp16)[name = tensor("x_87_cast_fp16")]; tensor var_1122 = const()[name = tensor("op_1122"), val = tensor([1, 8, -1, 8])]; tensor x_89_cast_fp16 = reshape(shape = var_1122, x = x_87_cast_fp16)[name = tensor("x_89_cast_fp16")]; tensor var_1126_begin_0 = const()[name = tensor("op_1126_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1126_end_0 = const()[name = tensor("op_1126_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_1126_end_mask_0 = const()[name = tensor("op_1126_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1126_cast_fp16 = slice_by_index(begin = var_1126_begin_0, end = var_1126_end_0, end_mask = var_1126_end_mask_0, x = x_89_cast_fp16)[name = tensor("op_1126_cast_fp16")]; tensor var_1127 = const()[name = tensor("op_1127"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_13_cast_fp16 = reshape(shape = var_1127, x = var_1126_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_208")]; tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_1110_cast_fp16)[name = tensor("transpose_209")]; tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; tensor var_1136_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_1136_cast_fp16")]; tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_13_cast_fp16 = mul(x = var_1136_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; tensor scores_15_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_11)[name = tensor("scores_15_cast_fp16")]; tensor var_1142_cast_fp16 = softmax(axis = var_56, x = scores_15_cast_fp16)[name = tensor("op_1142_cast_fp16")]; tensor input_197_cast_fp16 = select(a = var_40_to_fp16, b = var_1142_cast_fp16, cond = mask_11)[name = tensor("input_197_cast_fp16")]; tensor x_91_transpose_x_0 = const()[name = tensor("x_91_transpose_x_0"), val = tensor(false)]; tensor x_91_transpose_y_0 = const()[name = tensor("x_91_transpose_y_0"), val = tensor(false)]; tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_211")]; tensor x_91_cast_fp16 = matmul(transpose_x = x_91_transpose_x_0, transpose_y = x_91_transpose_y_0, x = input_197_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_91_cast_fp16")]; tensor var_1146_perm_0 = const()[name = tensor("op_1146_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1147 = const()[name = tensor("op_1147"), val = tensor([1, -1, 512])]; tensor var_1146_cast_fp16 = transpose(perm = var_1146_perm_0, x = x_91_cast_fp16)[name = tensor("transpose_207")]; tensor input_199_cast_fp16 = reshape(shape = var_1147, x = var_1146_cast_fp16)[name = tensor("input_199_cast_fp16")]; tensor encoder_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47402816)))]; tensor linear_34_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_out_weight_to_fp16, x = input_199_cast_fp16)[name = tensor("linear_34_cast_fp16")]; tensor input_203_cast_fp16 = add(x = input_193_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_203_cast_fp16")]; tensor x_95_axes_0 = const()[name = tensor("x_95_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47927168)))]; tensor encoder_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47928256)))]; tensor x_95_cast_fp16 = layer_norm(axes = x_95_axes_0, beta = encoder_layers_3_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_conv_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("x_95_cast_fp16")]; tensor input_205_perm_0 = const()[name = tensor("input_205_perm_0"), val = tensor([0, 2, 1])]; tensor input_207_pad_type_0 = const()[name = tensor("input_207_pad_type_0"), val = tensor("valid")]; tensor input_207_strides_0 = const()[name = tensor("input_207_strides_0"), val = tensor([1])]; tensor input_207_pad_0 = const()[name = tensor("input_207_pad_0"), val = tensor([0, 0])]; tensor input_207_dilations_0 = const()[name = tensor("input_207_dilations_0"), val = tensor([1])]; tensor input_207_groups_0 = const()[name = tensor("input_207_groups_0"), val = tensor(1)]; tensor encoder_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47929344)))]; tensor input_205_cast_fp16 = transpose(perm = input_205_perm_0, x = x_95_cast_fp16)[name = tensor("transpose_206")]; tensor input_207_cast_fp16 = conv(dilations = input_207_dilations_0, groups = input_207_groups_0, pad = input_207_pad_0, pad_type = input_207_pad_type_0, strides = input_207_strides_0, weight = encoder_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_205_cast_fp16)[name = tensor("input_207_cast_fp16")]; tensor x_97_split_num_splits_0 = const()[name = tensor("x_97_split_num_splits_0"), val = tensor(2)]; tensor x_97_split_axis_0 = const()[name = tensor("x_97_split_axis_0"), val = tensor(1)]; tensor x_97_split_cast_fp16_0, tensor x_97_split_cast_fp16_1 = split(axis = x_97_split_axis_0, num_splits = x_97_split_num_splits_0, x = input_207_cast_fp16)[name = tensor("x_97_split_cast_fp16")]; tensor x_97_split_1_sigmoid_cast_fp16 = sigmoid(x = x_97_split_cast_fp16_1)[name = tensor("x_97_split_1_sigmoid_cast_fp16")]; tensor x_97_cast_fp16 = mul(x = x_97_split_cast_fp16_0, y = x_97_split_1_sigmoid_cast_fp16)[name = tensor("x_97_cast_fp16")]; tensor input_209_cast_fp16 = select(a = var_40_to_fp16, b = x_97_cast_fp16, cond = var_551)[name = tensor("input_209_cast_fp16")]; tensor new_x_15_interleave_0 = const()[name = tensor("new_x_15_interleave_0"), val = tensor(false)]; tensor new_x_15_cast_fp16 = concat(axis = var_56, interleave = new_x_15_interleave_0, values = (cache_15_cast_fp16, input_209_cast_fp16))[name = tensor("new_x_15_cast_fp16")]; tensor next_cache_7_begin_0 = const()[name = tensor("next_cache_7_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_7_end_0 = const()[name = tensor("next_cache_7_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_7_end_mask_0 = const()[name = tensor("next_cache_7_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_7_cast_fp16 = slice_by_index(begin = next_cache_7_begin_0, end = next_cache_7_end_0, end_mask = next_cache_7_end_mask_0, x = new_x_15_cast_fp16)[name = tensor("next_cache_7_cast_fp16")]; tensor var_1188_begin_0 = const()[name = tensor("op_1188_begin_0"), val = tensor([0, 0, 4])]; tensor var_1188_end_0 = const()[name = tensor("op_1188_end_0"), val = tensor([1, 512, 12])]; tensor var_1188_end_mask_0 = const()[name = tensor("op_1188_end_mask_0"), val = tensor([true, true, true])]; tensor var_1188_cast_fp16 = slice_by_index(begin = var_1188_begin_0, end = var_1188_end_0, end_mask = var_1188_end_mask_0, x = next_cache_7_cast_fp16)[name = tensor("op_1188_cast_fp16")]; tensor x_99_pad_type_0 = const()[name = tensor("x_99_pad_type_0"), val = tensor("valid")]; tensor x_99_groups_0 = const()[name = tensor("x_99_groups_0"), val = tensor(512)]; tensor x_99_strides_0 = const()[name = tensor("x_99_strides_0"), val = tensor([1])]; tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0])]; tensor x_99_dilations_0 = const()[name = tensor("x_99_dilations_0"), val = tensor([1])]; tensor encoder_layers_3_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48977984)))]; tensor x_99_cast_fp16 = conv(dilations = x_99_dilations_0, groups = x_99_groups_0, pad = x_99_pad_0, pad_type = x_99_pad_type_0, strides = x_99_strides_0, weight = encoder_layers_3_conv_depthwise_conv_weight_to_fp16, x = new_x_15_cast_fp16)[name = tensor("x_99_cast_fp16")]; tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; tensor x_101_axes_0 = const()[name = tensor("x_101_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48987264)))]; tensor encoder_layers_3_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48988352)))]; tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_99_cast_fp16)[name = tensor("transpose_205")]; tensor x_101_cast_fp16 = layer_norm(axes = x_101_axes_0, beta = encoder_layers_3_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_conv_batch_norm_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("x_101_cast_fp16")]; tensor input_213_perm_0 = const()[name = tensor("input_213_perm_0"), val = tensor([0, 2, 1])]; tensor input_213_cast_fp16 = transpose(perm = input_213_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_204")]; tensor input_215_cast_fp16 = silu(x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; tensor x_103_pad_type_0 = const()[name = tensor("x_103_pad_type_0"), val = tensor("valid")]; tensor x_103_strides_0 = const()[name = tensor("x_103_strides_0"), val = tensor([1])]; tensor x_103_pad_0 = const()[name = tensor("x_103_pad_0"), val = tensor([0, 0])]; tensor x_103_dilations_0 = const()[name = tensor("x_103_dilations_0"), val = tensor([1])]; tensor x_103_groups_0 = const()[name = tensor("x_103_groups_0"), val = tensor(1)]; tensor encoder_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48989440)))]; tensor x_103_cast_fp16 = conv(dilations = x_103_dilations_0, groups = x_103_groups_0, pad = x_103_pad_0, pad_type = x_103_pad_type_0, strides = x_103_strides_0, weight = encoder_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("x_103_cast_fp16")]; tensor input_217_perm_0 = const()[name = tensor("input_217_perm_0"), val = tensor([0, 2, 1])]; tensor input_217_cast_fp16 = transpose(perm = input_217_perm_0, x = x_103_cast_fp16)[name = tensor("transpose_203")]; tensor input_219_cast_fp16 = add(x = input_203_cast_fp16, y = input_217_cast_fp16)[name = tensor("input_219_cast_fp16")]; tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49513792)))]; tensor encoder_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49514880)))]; tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = encoder_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; tensor encoder_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49515968)))]; tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("linear_35_cast_fp16")]; tensor input_225_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_225_cast_fp16")]; tensor encoder_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51613184)))]; tensor linear_36_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("linear_36_cast_fp16")]; tensor var_1229_to_fp16 = const()[name = tensor("op_1229_to_fp16"), val = tensor(0x1p-1)]; tensor var_1230_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1229_to_fp16)[name = tensor("op_1230_cast_fp16")]; tensor input_231_cast_fp16 = add(x = input_219_cast_fp16, y = var_1230_cast_fp16)[name = tensor("input_231_cast_fp16")]; tensor input_233_axes_0 = const()[name = tensor("input_233_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53710400)))]; tensor encoder_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53711488)))]; tensor input_233_cast_fp16 = layer_norm(axes = input_233_axes_0, beta = encoder_layers_3_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_out_weight_to_fp16, x = input_231_cast_fp16)[name = tensor("input_233_cast_fp16")]; tensor cache_17_begin_0 = const()[name = tensor("cache_17_begin_0"), val = tensor([4, 0, 0, 0])]; tensor cache_17_end_0 = const()[name = tensor("cache_17_end_0"), val = tensor([5, 1, 70, 512])]; tensor cache_17_end_mask_0 = const()[name = tensor("cache_17_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_17_squeeze_mask_0 = const()[name = tensor("cache_17_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_17_cast_fp16 = slice_by_index(begin = cache_17_begin_0, end = cache_17_end_0, end_mask = cache_17_end_mask_0, squeeze_mask = cache_17_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_17_cast_fp16")]; tensor cache_19_begin_0 = const()[name = tensor("cache_19_begin_0"), val = tensor([4, 0, 0, 0])]; tensor cache_19_end_0 = const()[name = tensor("cache_19_end_0"), val = tensor([5, 1, 512, 8])]; tensor cache_19_end_mask_0 = const()[name = tensor("cache_19_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_19_squeeze_mask_0 = const()[name = tensor("cache_19_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_19_cast_fp16 = slice_by_index(begin = cache_19_begin_0, end = cache_19_end_0, end_mask = cache_19_end_mask_0, squeeze_mask = cache_19_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_19_cast_fp16")]; tensor input_235_axes_0 = const()[name = tensor("input_235_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53712576)))]; tensor encoder_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53713664)))]; tensor input_235_cast_fp16 = layer_norm(axes = input_235_axes_0, beta = encoder_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward1_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("input_235_cast_fp16")]; tensor encoder_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53714752)))]; tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("linear_37_cast_fp16")]; tensor input_239_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_239_cast_fp16")]; tensor encoder_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55811968)))]; tensor linear_38_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("linear_38_cast_fp16")]; tensor var_1264_to_fp16 = const()[name = tensor("op_1264_to_fp16"), val = tensor(0x1p-1)]; tensor var_1265_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1264_to_fp16)[name = tensor("op_1265_cast_fp16")]; tensor input_245_cast_fp16 = add(x = input_233_cast_fp16, y = var_1265_cast_fp16)[name = tensor("input_245_cast_fp16")]; tensor key_9_axes_0 = const()[name = tensor("key_9_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57909184)))]; tensor encoder_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57910272)))]; tensor key_9_cast_fp16 = layer_norm(axes = key_9_axes_0, beta = encoder_layers_4_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_self_att_weight_to_fp16, x = input_245_cast_fp16)[name = tensor("key_9_cast_fp16")]; tensor input_247_interleave_0 = const()[name = tensor("input_247_interleave_0"), val = tensor(false)]; tensor input_247_cast_fp16 = concat(axis = var_65, interleave = input_247_interleave_0, values = (cache_17_cast_fp16, key_9_cast_fp16))[name = tensor("input_247_cast_fp16")]; tensor var_1287_begin_0 = const()[name = tensor("op_1287_begin_0"), val = tensor([0, 4, 0])]; tensor var_1287_end_0 = const()[name = tensor("op_1287_end_0"), val = tensor([1, 70, 512])]; tensor var_1287_end_mask_0 = const()[name = tensor("op_1287_end_mask_0"), val = tensor([true, true, true])]; tensor var_1287_cast_fp16 = slice_by_index(begin = var_1287_begin_0, end = var_1287_end_0, end_mask = var_1287_end_mask_0, x = cache_17_cast_fp16)[name = tensor("op_1287_cast_fp16")]; tensor var_1290_begin_0 = const()[name = tensor("op_1290_begin_0"), val = tensor([0, 0, 0])]; tensor var_1290_end_0 = const()[name = tensor("op_1290_end_0"), val = tensor([1, 4, 512])]; tensor var_1290_end_mask_0 = const()[name = tensor("op_1290_end_mask_0"), val = tensor([true, false, true])]; tensor var_1290_cast_fp16 = slice_by_index(begin = var_1290_begin_0, end = var_1290_end_0, end_mask = var_1290_end_mask_0, x = key_9_cast_fp16)[name = tensor("op_1290_cast_fp16")]; tensor var_1293_interleave_0 = const()[name = tensor("op_1293_interleave_0"), val = tensor(false)]; tensor var_1293_cast_fp16 = concat(axis = var_65, interleave = var_1293_interleave_0, values = (var_1287_cast_fp16, var_1290_cast_fp16))[name = tensor("op_1293_cast_fp16")]; tensor encoder_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57911360)))]; tensor linear_39_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_q_weight_to_fp16, x = key_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; tensor var_1297 = const()[name = tensor("op_1297"), val = tensor([1, -1, 8, 64])]; tensor q_25_cast_fp16 = reshape(shape = var_1297, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; tensor encoder_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58435712)))]; tensor linear_40_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_k_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_40_cast_fp16")]; tensor var_1301 = const()[name = tensor("op_1301"), val = tensor([1, -1, 8, 64])]; tensor k_17_cast_fp16 = reshape(shape = var_1301, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; tensor encoder_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58960064)))]; tensor linear_41_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_v_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_41_cast_fp16")]; tensor var_1305 = const()[name = tensor("op_1305"), val = tensor([1, -1, 8, 64])]; tensor v_9_cast_fp16 = reshape(shape = var_1305, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59484416)))]; tensor var_1317_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1317_cast_fp16")]; tensor encoder_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59485504)))]; tensor var_1319_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1319_cast_fp16")]; tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_111_transpose_x_0 = const()[name = tensor("x_111_transpose_x_0"), val = tensor(false)]; tensor x_111_transpose_y_0 = const()[name = tensor("x_111_transpose_y_0"), val = tensor(false)]; tensor var_1321_to_fp16 = const()[name = tensor("op_1321_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59486592)))]; tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1319_cast_fp16)[name = tensor("transpose_201")]; tensor x_111_cast_fp16 = matmul(transpose_x = x_111_transpose_x_0, transpose_y = x_111_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1321_to_fp16)[name = tensor("x_111_cast_fp16")]; tensor x_113_pad_0 = const()[name = tensor("x_113_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_113_mode_0 = const()[name = tensor("x_113_mode_0"), val = tensor("constant")]; tensor const_131_to_fp16 = const()[name = tensor("const_131_to_fp16"), val = tensor(0x0p+0)]; tensor x_113_cast_fp16 = pad(constant_val = const_131_to_fp16, mode = x_113_mode_0, pad = x_113_pad_0, x = x_111_cast_fp16)[name = tensor("x_113_cast_fp16")]; tensor var_1329 = const()[name = tensor("op_1329"), val = tensor([1, 8, -1, 8])]; tensor x_115_cast_fp16 = reshape(shape = var_1329, x = x_113_cast_fp16)[name = tensor("x_115_cast_fp16")]; tensor var_1333_begin_0 = const()[name = tensor("op_1333_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1333_end_0 = const()[name = tensor("op_1333_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_1333_end_mask_0 = const()[name = tensor("op_1333_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1333_cast_fp16 = slice_by_index(begin = var_1333_begin_0, end = var_1333_end_0, end_mask = var_1333_end_mask_0, x = x_115_cast_fp16)[name = tensor("op_1333_cast_fp16")]; tensor var_1334 = const()[name = tensor("op_1334"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1334, x = var_1333_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_199")]; tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1317_cast_fp16)[name = tensor("transpose_200")]; tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; tensor var_1343_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1343_cast_fp16")]; tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_17_cast_fp16 = mul(x = var_1343_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; tensor scores_19_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_11)[name = tensor("scores_19_cast_fp16")]; tensor var_1349_cast_fp16 = softmax(axis = var_56, x = scores_19_cast_fp16)[name = tensor("op_1349_cast_fp16")]; tensor input_249_cast_fp16 = select(a = var_40_to_fp16, b = var_1349_cast_fp16, cond = mask_11)[name = tensor("input_249_cast_fp16")]; tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_202")]; tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = input_249_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_117_cast_fp16")]; tensor var_1353_perm_0 = const()[name = tensor("op_1353_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1354 = const()[name = tensor("op_1354"), val = tensor([1, -1, 512])]; tensor var_1353_cast_fp16 = transpose(perm = var_1353_perm_0, x = x_117_cast_fp16)[name = tensor("transpose_198")]; tensor input_251_cast_fp16 = reshape(shape = var_1354, x = var_1353_cast_fp16)[name = tensor("input_251_cast_fp16")]; tensor encoder_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59645376)))]; tensor linear_43_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_out_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("linear_43_cast_fp16")]; tensor input_255_cast_fp16 = add(x = input_245_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_255_cast_fp16")]; tensor x_121_axes_0 = const()[name = tensor("x_121_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60169728)))]; tensor encoder_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60170816)))]; tensor x_121_cast_fp16 = layer_norm(axes = x_121_axes_0, beta = encoder_layers_4_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_conv_weight_to_fp16, x = input_255_cast_fp16)[name = tensor("x_121_cast_fp16")]; tensor input_257_perm_0 = const()[name = tensor("input_257_perm_0"), val = tensor([0, 2, 1])]; tensor input_259_pad_type_0 = const()[name = tensor("input_259_pad_type_0"), val = tensor("valid")]; tensor input_259_strides_0 = const()[name = tensor("input_259_strides_0"), val = tensor([1])]; tensor input_259_pad_0 = const()[name = tensor("input_259_pad_0"), val = tensor([0, 0])]; tensor input_259_dilations_0 = const()[name = tensor("input_259_dilations_0"), val = tensor([1])]; tensor input_259_groups_0 = const()[name = tensor("input_259_groups_0"), val = tensor(1)]; tensor encoder_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60171904)))]; tensor input_257_cast_fp16 = transpose(perm = input_257_perm_0, x = x_121_cast_fp16)[name = tensor("transpose_197")]; tensor input_259_cast_fp16 = conv(dilations = input_259_dilations_0, groups = input_259_groups_0, pad = input_259_pad_0, pad_type = input_259_pad_type_0, strides = input_259_strides_0, weight = encoder_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_257_cast_fp16)[name = tensor("input_259_cast_fp16")]; tensor x_123_split_num_splits_0 = const()[name = tensor("x_123_split_num_splits_0"), val = tensor(2)]; tensor x_123_split_axis_0 = const()[name = tensor("x_123_split_axis_0"), val = tensor(1)]; tensor x_123_split_cast_fp16_0, tensor x_123_split_cast_fp16_1 = split(axis = x_123_split_axis_0, num_splits = x_123_split_num_splits_0, x = input_259_cast_fp16)[name = tensor("x_123_split_cast_fp16")]; tensor x_123_split_1_sigmoid_cast_fp16 = sigmoid(x = x_123_split_cast_fp16_1)[name = tensor("x_123_split_1_sigmoid_cast_fp16")]; tensor x_123_cast_fp16 = mul(x = x_123_split_cast_fp16_0, y = x_123_split_1_sigmoid_cast_fp16)[name = tensor("x_123_cast_fp16")]; tensor input_261_cast_fp16 = select(a = var_40_to_fp16, b = x_123_cast_fp16, cond = var_551)[name = tensor("input_261_cast_fp16")]; tensor new_x_19_interleave_0 = const()[name = tensor("new_x_19_interleave_0"), val = tensor(false)]; tensor new_x_19_cast_fp16 = concat(axis = var_56, interleave = new_x_19_interleave_0, values = (cache_19_cast_fp16, input_261_cast_fp16))[name = tensor("new_x_19_cast_fp16")]; tensor next_cache_9_begin_0 = const()[name = tensor("next_cache_9_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_9_end_0 = const()[name = tensor("next_cache_9_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_9_end_mask_0 = const()[name = tensor("next_cache_9_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_9_cast_fp16 = slice_by_index(begin = next_cache_9_begin_0, end = next_cache_9_end_0, end_mask = next_cache_9_end_mask_0, x = new_x_19_cast_fp16)[name = tensor("next_cache_9_cast_fp16")]; tensor var_1395_begin_0 = const()[name = tensor("op_1395_begin_0"), val = tensor([0, 0, 4])]; tensor var_1395_end_0 = const()[name = tensor("op_1395_end_0"), val = tensor([1, 512, 12])]; tensor var_1395_end_mask_0 = const()[name = tensor("op_1395_end_mask_0"), val = tensor([true, true, true])]; tensor var_1395_cast_fp16 = slice_by_index(begin = var_1395_begin_0, end = var_1395_end_0, end_mask = var_1395_end_mask_0, x = next_cache_9_cast_fp16)[name = tensor("op_1395_cast_fp16")]; tensor x_125_pad_type_0 = const()[name = tensor("x_125_pad_type_0"), val = tensor("valid")]; tensor x_125_groups_0 = const()[name = tensor("x_125_groups_0"), val = tensor(512)]; tensor x_125_strides_0 = const()[name = tensor("x_125_strides_0"), val = tensor([1])]; tensor x_125_pad_0 = const()[name = tensor("x_125_pad_0"), val = tensor([0, 0])]; tensor x_125_dilations_0 = const()[name = tensor("x_125_dilations_0"), val = tensor([1])]; tensor encoder_layers_4_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61220544)))]; tensor x_125_cast_fp16 = conv(dilations = x_125_dilations_0, groups = x_125_groups_0, pad = x_125_pad_0, pad_type = x_125_pad_type_0, strides = x_125_strides_0, weight = encoder_layers_4_conv_depthwise_conv_weight_to_fp16, x = new_x_19_cast_fp16)[name = tensor("x_125_cast_fp16")]; tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61229824)))]; tensor encoder_layers_4_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61230912)))]; tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_125_cast_fp16)[name = tensor("transpose_196")]; tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = encoder_layers_4_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_conv_batch_norm_weight_to_fp16, x = input_263_cast_fp16)[name = tensor("x_127_cast_fp16")]; tensor input_265_perm_0 = const()[name = tensor("input_265_perm_0"), val = tensor([0, 2, 1])]; tensor input_265_cast_fp16 = transpose(perm = input_265_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_195")]; tensor input_267_cast_fp16 = silu(x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; tensor x_129_pad_type_0 = const()[name = tensor("x_129_pad_type_0"), val = tensor("valid")]; tensor x_129_strides_0 = const()[name = tensor("x_129_strides_0"), val = tensor([1])]; tensor x_129_pad_0 = const()[name = tensor("x_129_pad_0"), val = tensor([0, 0])]; tensor x_129_dilations_0 = const()[name = tensor("x_129_dilations_0"), val = tensor([1])]; tensor x_129_groups_0 = const()[name = tensor("x_129_groups_0"), val = tensor(1)]; tensor encoder_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61232000)))]; tensor x_129_cast_fp16 = conv(dilations = x_129_dilations_0, groups = x_129_groups_0, pad = x_129_pad_0, pad_type = x_129_pad_type_0, strides = x_129_strides_0, weight = encoder_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("x_129_cast_fp16")]; tensor input_269_perm_0 = const()[name = tensor("input_269_perm_0"), val = tensor([0, 2, 1])]; tensor input_269_cast_fp16 = transpose(perm = input_269_perm_0, x = x_129_cast_fp16)[name = tensor("transpose_194")]; tensor input_271_cast_fp16 = add(x = input_255_cast_fp16, y = input_269_cast_fp16)[name = tensor("input_271_cast_fp16")]; tensor input_273_axes_0 = const()[name = tensor("input_273_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61756352)))]; tensor encoder_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61757440)))]; tensor input_273_cast_fp16 = layer_norm(axes = input_273_axes_0, beta = encoder_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("input_273_cast_fp16")]; tensor encoder_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61758528)))]; tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("linear_44_cast_fp16")]; tensor input_277_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_277_cast_fp16")]; tensor encoder_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63855744)))]; tensor linear_45_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("linear_45_cast_fp16")]; tensor var_1436_to_fp16 = const()[name = tensor("op_1436_to_fp16"), val = tensor(0x1p-1)]; tensor var_1437_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1436_to_fp16)[name = tensor("op_1437_cast_fp16")]; tensor input_283_cast_fp16 = add(x = input_271_cast_fp16, y = var_1437_cast_fp16)[name = tensor("input_283_cast_fp16")]; tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65952960)))]; tensor encoder_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65954048)))]; tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = encoder_layers_4_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_out_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; tensor cache_21_begin_0 = const()[name = tensor("cache_21_begin_0"), val = tensor([5, 0, 0, 0])]; tensor cache_21_end_0 = const()[name = tensor("cache_21_end_0"), val = tensor([6, 1, 70, 512])]; tensor cache_21_end_mask_0 = const()[name = tensor("cache_21_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_21_squeeze_mask_0 = const()[name = tensor("cache_21_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_21_cast_fp16 = slice_by_index(begin = cache_21_begin_0, end = cache_21_end_0, end_mask = cache_21_end_mask_0, squeeze_mask = cache_21_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_21_cast_fp16")]; tensor cache_23_begin_0 = const()[name = tensor("cache_23_begin_0"), val = tensor([5, 0, 0, 0])]; tensor cache_23_end_0 = const()[name = tensor("cache_23_end_0"), val = tensor([6, 1, 512, 8])]; tensor cache_23_end_mask_0 = const()[name = tensor("cache_23_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_23_squeeze_mask_0 = const()[name = tensor("cache_23_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_23_cast_fp16 = slice_by_index(begin = cache_23_begin_0, end = cache_23_end_0, end_mask = cache_23_end_mask_0, squeeze_mask = cache_23_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_23_cast_fp16")]; tensor input_287_axes_0 = const()[name = tensor("input_287_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65955136)))]; tensor encoder_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65956224)))]; tensor input_287_cast_fp16 = layer_norm(axes = input_287_axes_0, beta = encoder_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward1_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("input_287_cast_fp16")]; tensor encoder_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65957312)))]; tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_287_cast_fp16)[name = tensor("linear_46_cast_fp16")]; tensor input_291_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_291_cast_fp16")]; tensor encoder_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68054528)))]; tensor linear_47_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("linear_47_cast_fp16")]; tensor var_1471_to_fp16 = const()[name = tensor("op_1471_to_fp16"), val = tensor(0x1p-1)]; tensor var_1472_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1471_to_fp16)[name = tensor("op_1472_cast_fp16")]; tensor input_297_cast_fp16 = add(x = input_285_cast_fp16, y = var_1472_cast_fp16)[name = tensor("input_297_cast_fp16")]; tensor key_11_axes_0 = const()[name = tensor("key_11_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70151744)))]; tensor encoder_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70152832)))]; tensor key_11_cast_fp16 = layer_norm(axes = key_11_axes_0, beta = encoder_layers_5_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_self_att_weight_to_fp16, x = input_297_cast_fp16)[name = tensor("key_11_cast_fp16")]; tensor input_299_interleave_0 = const()[name = tensor("input_299_interleave_0"), val = tensor(false)]; tensor input_299_cast_fp16 = concat(axis = var_65, interleave = input_299_interleave_0, values = (cache_21_cast_fp16, key_11_cast_fp16))[name = tensor("input_299_cast_fp16")]; tensor var_1494_begin_0 = const()[name = tensor("op_1494_begin_0"), val = tensor([0, 4, 0])]; tensor var_1494_end_0 = const()[name = tensor("op_1494_end_0"), val = tensor([1, 70, 512])]; tensor var_1494_end_mask_0 = const()[name = tensor("op_1494_end_mask_0"), val = tensor([true, true, true])]; tensor var_1494_cast_fp16 = slice_by_index(begin = var_1494_begin_0, end = var_1494_end_0, end_mask = var_1494_end_mask_0, x = cache_21_cast_fp16)[name = tensor("op_1494_cast_fp16")]; tensor var_1497_begin_0 = const()[name = tensor("op_1497_begin_0"), val = tensor([0, 0, 0])]; tensor var_1497_end_0 = const()[name = tensor("op_1497_end_0"), val = tensor([1, 4, 512])]; tensor var_1497_end_mask_0 = const()[name = tensor("op_1497_end_mask_0"), val = tensor([true, false, true])]; tensor var_1497_cast_fp16 = slice_by_index(begin = var_1497_begin_0, end = var_1497_end_0, end_mask = var_1497_end_mask_0, x = key_11_cast_fp16)[name = tensor("op_1497_cast_fp16")]; tensor var_1500_interleave_0 = const()[name = tensor("op_1500_interleave_0"), val = tensor(false)]; tensor var_1500_cast_fp16 = concat(axis = var_65, interleave = var_1500_interleave_0, values = (var_1494_cast_fp16, var_1497_cast_fp16))[name = tensor("op_1500_cast_fp16")]; tensor encoder_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70153920)))]; tensor linear_48_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_q_weight_to_fp16, x = key_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; tensor var_1504 = const()[name = tensor("op_1504"), val = tensor([1, -1, 8, 64])]; tensor q_31_cast_fp16 = reshape(shape = var_1504, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; tensor encoder_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70678272)))]; tensor linear_49_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_k_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_49_cast_fp16")]; tensor var_1508 = const()[name = tensor("op_1508"), val = tensor([1, -1, 8, 64])]; tensor k_21_cast_fp16 = reshape(shape = var_1508, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; tensor encoder_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71202624)))]; tensor linear_50_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_v_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_50_cast_fp16")]; tensor var_1512 = const()[name = tensor("op_1512"), val = tensor([1, -1, 8, 64])]; tensor v_11_cast_fp16 = reshape(shape = var_1512, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71726976)))]; tensor var_1524_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1524_cast_fp16")]; tensor encoder_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71728064)))]; tensor var_1526_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1526_cast_fp16")]; tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_137_transpose_x_0 = const()[name = tensor("x_137_transpose_x_0"), val = tensor(false)]; tensor x_137_transpose_y_0 = const()[name = tensor("x_137_transpose_y_0"), val = tensor(false)]; tensor var_1528_to_fp16 = const()[name = tensor("op_1528_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71729152)))]; tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1526_cast_fp16)[name = tensor("transpose_192")]; tensor x_137_cast_fp16 = matmul(transpose_x = x_137_transpose_x_0, transpose_y = x_137_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1528_to_fp16)[name = tensor("x_137_cast_fp16")]; tensor x_139_pad_0 = const()[name = tensor("x_139_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_139_mode_0 = const()[name = tensor("x_139_mode_0"), val = tensor("constant")]; tensor const_144_to_fp16 = const()[name = tensor("const_144_to_fp16"), val = tensor(0x0p+0)]; tensor x_139_cast_fp16 = pad(constant_val = const_144_to_fp16, mode = x_139_mode_0, pad = x_139_pad_0, x = x_137_cast_fp16)[name = tensor("x_139_cast_fp16")]; tensor var_1536 = const()[name = tensor("op_1536"), val = tensor([1, 8, -1, 8])]; tensor x_141_cast_fp16 = reshape(shape = var_1536, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; tensor var_1540_begin_0 = const()[name = tensor("op_1540_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1540_end_0 = const()[name = tensor("op_1540_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_1540_end_mask_0 = const()[name = tensor("op_1540_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1540_cast_fp16 = slice_by_index(begin = var_1540_begin_0, end = var_1540_end_0, end_mask = var_1540_end_mask_0, x = x_141_cast_fp16)[name = tensor("op_1540_cast_fp16")]; tensor var_1541 = const()[name = tensor("op_1541"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1541, x = var_1540_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_190")]; tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1524_cast_fp16)[name = tensor("transpose_191")]; tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; tensor var_1550_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1550_cast_fp16")]; tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_21_cast_fp16 = mul(x = var_1550_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; tensor scores_23_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_11)[name = tensor("scores_23_cast_fp16")]; tensor var_1556_cast_fp16 = softmax(axis = var_56, x = scores_23_cast_fp16)[name = tensor("op_1556_cast_fp16")]; tensor input_301_cast_fp16 = select(a = var_40_to_fp16, b = var_1556_cast_fp16, cond = mask_11)[name = tensor("input_301_cast_fp16")]; tensor x_143_transpose_x_0 = const()[name = tensor("x_143_transpose_x_0"), val = tensor(false)]; tensor x_143_transpose_y_0 = const()[name = tensor("x_143_transpose_y_0"), val = tensor(false)]; tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_193")]; tensor x_143_cast_fp16 = matmul(transpose_x = x_143_transpose_x_0, transpose_y = x_143_transpose_y_0, x = input_301_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_143_cast_fp16")]; tensor var_1560_perm_0 = const()[name = tensor("op_1560_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1561 = const()[name = tensor("op_1561"), val = tensor([1, -1, 512])]; tensor var_1560_cast_fp16 = transpose(perm = var_1560_perm_0, x = x_143_cast_fp16)[name = tensor("transpose_189")]; tensor input_303_cast_fp16 = reshape(shape = var_1561, x = var_1560_cast_fp16)[name = tensor("input_303_cast_fp16")]; tensor encoder_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71887936)))]; tensor linear_52_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_out_weight_to_fp16, x = input_303_cast_fp16)[name = tensor("linear_52_cast_fp16")]; tensor input_307_cast_fp16 = add(x = input_297_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_307_cast_fp16")]; tensor x_147_axes_0 = const()[name = tensor("x_147_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72412288)))]; tensor encoder_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72413376)))]; tensor x_147_cast_fp16 = layer_norm(axes = x_147_axes_0, beta = encoder_layers_5_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_conv_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("x_147_cast_fp16")]; tensor input_309_perm_0 = const()[name = tensor("input_309_perm_0"), val = tensor([0, 2, 1])]; tensor input_311_pad_type_0 = const()[name = tensor("input_311_pad_type_0"), val = tensor("valid")]; tensor input_311_strides_0 = const()[name = tensor("input_311_strides_0"), val = tensor([1])]; tensor input_311_pad_0 = const()[name = tensor("input_311_pad_0"), val = tensor([0, 0])]; tensor input_311_dilations_0 = const()[name = tensor("input_311_dilations_0"), val = tensor([1])]; tensor input_311_groups_0 = const()[name = tensor("input_311_groups_0"), val = tensor(1)]; tensor encoder_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72414464)))]; tensor input_309_cast_fp16 = transpose(perm = input_309_perm_0, x = x_147_cast_fp16)[name = tensor("transpose_188")]; tensor input_311_cast_fp16 = conv(dilations = input_311_dilations_0, groups = input_311_groups_0, pad = input_311_pad_0, pad_type = input_311_pad_type_0, strides = input_311_strides_0, weight = encoder_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; tensor x_149_split_num_splits_0 = const()[name = tensor("x_149_split_num_splits_0"), val = tensor(2)]; tensor x_149_split_axis_0 = const()[name = tensor("x_149_split_axis_0"), val = tensor(1)]; tensor x_149_split_cast_fp16_0, tensor x_149_split_cast_fp16_1 = split(axis = x_149_split_axis_0, num_splits = x_149_split_num_splits_0, x = input_311_cast_fp16)[name = tensor("x_149_split_cast_fp16")]; tensor x_149_split_1_sigmoid_cast_fp16 = sigmoid(x = x_149_split_cast_fp16_1)[name = tensor("x_149_split_1_sigmoid_cast_fp16")]; tensor x_149_cast_fp16 = mul(x = x_149_split_cast_fp16_0, y = x_149_split_1_sigmoid_cast_fp16)[name = tensor("x_149_cast_fp16")]; tensor input_313_cast_fp16 = select(a = var_40_to_fp16, b = x_149_cast_fp16, cond = var_551)[name = tensor("input_313_cast_fp16")]; tensor new_x_23_interleave_0 = const()[name = tensor("new_x_23_interleave_0"), val = tensor(false)]; tensor new_x_23_cast_fp16 = concat(axis = var_56, interleave = new_x_23_interleave_0, values = (cache_23_cast_fp16, input_313_cast_fp16))[name = tensor("new_x_23_cast_fp16")]; tensor next_cache_11_begin_0 = const()[name = tensor("next_cache_11_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_11_end_0 = const()[name = tensor("next_cache_11_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_11_end_mask_0 = const()[name = tensor("next_cache_11_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_11_cast_fp16 = slice_by_index(begin = next_cache_11_begin_0, end = next_cache_11_end_0, end_mask = next_cache_11_end_mask_0, x = new_x_23_cast_fp16)[name = tensor("next_cache_11_cast_fp16")]; tensor var_1602_begin_0 = const()[name = tensor("op_1602_begin_0"), val = tensor([0, 0, 4])]; tensor var_1602_end_0 = const()[name = tensor("op_1602_end_0"), val = tensor([1, 512, 12])]; tensor var_1602_end_mask_0 = const()[name = tensor("op_1602_end_mask_0"), val = tensor([true, true, true])]; tensor var_1602_cast_fp16 = slice_by_index(begin = var_1602_begin_0, end = var_1602_end_0, end_mask = var_1602_end_mask_0, x = next_cache_11_cast_fp16)[name = tensor("op_1602_cast_fp16")]; tensor x_151_pad_type_0 = const()[name = tensor("x_151_pad_type_0"), val = tensor("valid")]; tensor x_151_groups_0 = const()[name = tensor("x_151_groups_0"), val = tensor(512)]; tensor x_151_strides_0 = const()[name = tensor("x_151_strides_0"), val = tensor([1])]; tensor x_151_pad_0 = const()[name = tensor("x_151_pad_0"), val = tensor([0, 0])]; tensor x_151_dilations_0 = const()[name = tensor("x_151_dilations_0"), val = tensor([1])]; tensor encoder_layers_5_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73463104)))]; tensor x_151_cast_fp16 = conv(dilations = x_151_dilations_0, groups = x_151_groups_0, pad = x_151_pad_0, pad_type = x_151_pad_type_0, strides = x_151_strides_0, weight = encoder_layers_5_conv_depthwise_conv_weight_to_fp16, x = new_x_23_cast_fp16)[name = tensor("x_151_cast_fp16")]; tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; tensor x_153_axes_0 = const()[name = tensor("x_153_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73472384)))]; tensor encoder_layers_5_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73473472)))]; tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_151_cast_fp16)[name = tensor("transpose_187")]; tensor x_153_cast_fp16 = layer_norm(axes = x_153_axes_0, beta = encoder_layers_5_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_conv_batch_norm_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("x_153_cast_fp16")]; tensor input_317_perm_0 = const()[name = tensor("input_317_perm_0"), val = tensor([0, 2, 1])]; tensor input_317_cast_fp16 = transpose(perm = input_317_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_186")]; tensor input_319_cast_fp16 = silu(x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; tensor encoder_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73474560)))]; tensor x_155_cast_fp16 = conv(dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = encoder_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("x_155_cast_fp16")]; tensor input_321_perm_0 = const()[name = tensor("input_321_perm_0"), val = tensor([0, 2, 1])]; tensor input_321_cast_fp16 = transpose(perm = input_321_perm_0, x = x_155_cast_fp16)[name = tensor("transpose_185")]; tensor input_323_cast_fp16 = add(x = input_307_cast_fp16, y = input_321_cast_fp16)[name = tensor("input_323_cast_fp16")]; tensor input_325_axes_0 = const()[name = tensor("input_325_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73998912)))]; tensor encoder_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74000000)))]; tensor input_325_cast_fp16 = layer_norm(axes = input_325_axes_0, beta = encoder_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("input_325_cast_fp16")]; tensor encoder_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74001088)))]; tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_325_cast_fp16)[name = tensor("linear_53_cast_fp16")]; tensor input_329_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_329_cast_fp16")]; tensor encoder_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76098304)))]; tensor linear_54_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("linear_54_cast_fp16")]; tensor var_1643_to_fp16 = const()[name = tensor("op_1643_to_fp16"), val = tensor(0x1p-1)]; tensor var_1644_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1643_to_fp16)[name = tensor("op_1644_cast_fp16")]; tensor input_335_cast_fp16 = add(x = input_323_cast_fp16, y = var_1644_cast_fp16)[name = tensor("input_335_cast_fp16")]; tensor input_337_axes_0 = const()[name = tensor("input_337_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78195520)))]; tensor encoder_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78196608)))]; tensor input_337_cast_fp16 = layer_norm(axes = input_337_axes_0, beta = encoder_layers_5_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_out_weight_to_fp16, x = input_335_cast_fp16)[name = tensor("input_337_cast_fp16")]; tensor cache_25_begin_0 = const()[name = tensor("cache_25_begin_0"), val = tensor([6, 0, 0, 0])]; tensor cache_25_end_0 = const()[name = tensor("cache_25_end_0"), val = tensor([7, 1, 70, 512])]; tensor cache_25_end_mask_0 = const()[name = tensor("cache_25_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_25_squeeze_mask_0 = const()[name = tensor("cache_25_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_25_cast_fp16 = slice_by_index(begin = cache_25_begin_0, end = cache_25_end_0, end_mask = cache_25_end_mask_0, squeeze_mask = cache_25_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_25_cast_fp16")]; tensor cache_27_begin_0 = const()[name = tensor("cache_27_begin_0"), val = tensor([6, 0, 0, 0])]; tensor cache_27_end_0 = const()[name = tensor("cache_27_end_0"), val = tensor([7, 1, 512, 8])]; tensor cache_27_end_mask_0 = const()[name = tensor("cache_27_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_27_squeeze_mask_0 = const()[name = tensor("cache_27_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_27_cast_fp16 = slice_by_index(begin = cache_27_begin_0, end = cache_27_end_0, end_mask = cache_27_end_mask_0, squeeze_mask = cache_27_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_27_cast_fp16")]; tensor input_339_axes_0 = const()[name = tensor("input_339_axes_0"), val = tensor([-1])]; tensor encoder_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78197696)))]; tensor encoder_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78198784)))]; tensor input_339_cast_fp16 = layer_norm(axes = input_339_axes_0, beta = encoder_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward1_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("input_339_cast_fp16")]; tensor encoder_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78199872)))]; tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("linear_55_cast_fp16")]; tensor input_343_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_343_cast_fp16")]; tensor encoder_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80297088)))]; tensor linear_56_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("linear_56_cast_fp16")]; tensor var_1678_to_fp16 = const()[name = tensor("op_1678_to_fp16"), val = tensor(0x1p-1)]; tensor var_1679_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1678_to_fp16)[name = tensor("op_1679_cast_fp16")]; tensor input_349_cast_fp16 = add(x = input_337_cast_fp16, y = var_1679_cast_fp16)[name = tensor("input_349_cast_fp16")]; tensor key_13_axes_0 = const()[name = tensor("key_13_axes_0"), val = tensor([-1])]; tensor encoder_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82394304)))]; tensor encoder_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82395392)))]; tensor key_13_cast_fp16 = layer_norm(axes = key_13_axes_0, beta = encoder_layers_6_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_self_att_weight_to_fp16, x = input_349_cast_fp16)[name = tensor("key_13_cast_fp16")]; tensor input_351_interleave_0 = const()[name = tensor("input_351_interleave_0"), val = tensor(false)]; tensor input_351_cast_fp16 = concat(axis = var_65, interleave = input_351_interleave_0, values = (cache_25_cast_fp16, key_13_cast_fp16))[name = tensor("input_351_cast_fp16")]; tensor var_1701_begin_0 = const()[name = tensor("op_1701_begin_0"), val = tensor([0, 4, 0])]; tensor var_1701_end_0 = const()[name = tensor("op_1701_end_0"), val = tensor([1, 70, 512])]; tensor var_1701_end_mask_0 = const()[name = tensor("op_1701_end_mask_0"), val = tensor([true, true, true])]; tensor var_1701_cast_fp16 = slice_by_index(begin = var_1701_begin_0, end = var_1701_end_0, end_mask = var_1701_end_mask_0, x = cache_25_cast_fp16)[name = tensor("op_1701_cast_fp16")]; tensor var_1704_begin_0 = const()[name = tensor("op_1704_begin_0"), val = tensor([0, 0, 0])]; tensor var_1704_end_0 = const()[name = tensor("op_1704_end_0"), val = tensor([1, 4, 512])]; tensor var_1704_end_mask_0 = const()[name = tensor("op_1704_end_mask_0"), val = tensor([true, false, true])]; tensor var_1704_cast_fp16 = slice_by_index(begin = var_1704_begin_0, end = var_1704_end_0, end_mask = var_1704_end_mask_0, x = key_13_cast_fp16)[name = tensor("op_1704_cast_fp16")]; tensor var_1707_interleave_0 = const()[name = tensor("op_1707_interleave_0"), val = tensor(false)]; tensor var_1707_cast_fp16 = concat(axis = var_65, interleave = var_1707_interleave_0, values = (var_1701_cast_fp16, var_1704_cast_fp16))[name = tensor("op_1707_cast_fp16")]; tensor encoder_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82396480)))]; tensor linear_57_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_q_weight_to_fp16, x = key_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; tensor var_1711 = const()[name = tensor("op_1711"), val = tensor([1, -1, 8, 64])]; tensor q_37_cast_fp16 = reshape(shape = var_1711, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; tensor encoder_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82920832)))]; tensor linear_58_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_k_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_58_cast_fp16")]; tensor var_1715 = const()[name = tensor("op_1715"), val = tensor([1, -1, 8, 64])]; tensor k_25_cast_fp16 = reshape(shape = var_1715, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; tensor encoder_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83445184)))]; tensor linear_59_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_v_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_59_cast_fp16")]; tensor var_1719 = const()[name = tensor("op_1719"), val = tensor([1, -1, 8, 64])]; tensor v_13_cast_fp16 = reshape(shape = var_1719, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83969536)))]; tensor var_1731_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1731_cast_fp16")]; tensor encoder_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83970624)))]; tensor var_1733_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1733_cast_fp16")]; tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; tensor var_1735_to_fp16 = const()[name = tensor("op_1735_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83971712)))]; tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1733_cast_fp16)[name = tensor("transpose_183")]; tensor x_163_cast_fp16 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1735_to_fp16)[name = tensor("x_163_cast_fp16")]; tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; tensor const_157_to_fp16 = const()[name = tensor("const_157_to_fp16"), val = tensor(0x0p+0)]; tensor x_165_cast_fp16 = pad(constant_val = const_157_to_fp16, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; tensor var_1743 = const()[name = tensor("op_1743"), val = tensor([1, 8, -1, 8])]; tensor x_167_cast_fp16 = reshape(shape = var_1743, x = x_165_cast_fp16)[name = tensor("x_167_cast_fp16")]; tensor var_1747_begin_0 = const()[name = tensor("op_1747_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1747_end_0 = const()[name = tensor("op_1747_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_1747_end_mask_0 = const()[name = tensor("op_1747_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1747_cast_fp16 = slice_by_index(begin = var_1747_begin_0, end = var_1747_end_0, end_mask = var_1747_end_mask_0, x = x_167_cast_fp16)[name = tensor("op_1747_cast_fp16")]; tensor var_1748 = const()[name = tensor("op_1748"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1748, x = var_1747_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_181")]; tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1731_cast_fp16)[name = tensor("transpose_182")]; tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; tensor var_1757_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1757_cast_fp16")]; tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_25_cast_fp16 = mul(x = var_1757_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; tensor scores_27_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_11)[name = tensor("scores_27_cast_fp16")]; tensor var_1763_cast_fp16 = softmax(axis = var_56, x = scores_27_cast_fp16)[name = tensor("op_1763_cast_fp16")]; tensor input_353_cast_fp16 = select(a = var_40_to_fp16, b = var_1763_cast_fp16, cond = mask_11)[name = tensor("input_353_cast_fp16")]; tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_184")]; tensor x_169_cast_fp16 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_353_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_169_cast_fp16")]; tensor var_1767_perm_0 = const()[name = tensor("op_1767_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1768 = const()[name = tensor("op_1768"), val = tensor([1, -1, 512])]; tensor var_1767_cast_fp16 = transpose(perm = var_1767_perm_0, x = x_169_cast_fp16)[name = tensor("transpose_180")]; tensor input_355_cast_fp16 = reshape(shape = var_1768, x = var_1767_cast_fp16)[name = tensor("input_355_cast_fp16")]; tensor encoder_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84130496)))]; tensor linear_61_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_out_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("linear_61_cast_fp16")]; tensor input_359_cast_fp16 = add(x = input_349_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_359_cast_fp16")]; tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; tensor encoder_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84654848)))]; tensor encoder_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84655936)))]; tensor x_173_cast_fp16 = layer_norm(axes = x_173_axes_0, beta = encoder_layers_6_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_conv_weight_to_fp16, x = input_359_cast_fp16)[name = tensor("x_173_cast_fp16")]; tensor input_361_perm_0 = const()[name = tensor("input_361_perm_0"), val = tensor([0, 2, 1])]; tensor input_363_pad_type_0 = const()[name = tensor("input_363_pad_type_0"), val = tensor("valid")]; tensor input_363_strides_0 = const()[name = tensor("input_363_strides_0"), val = tensor([1])]; tensor input_363_pad_0 = const()[name = tensor("input_363_pad_0"), val = tensor([0, 0])]; tensor input_363_dilations_0 = const()[name = tensor("input_363_dilations_0"), val = tensor([1])]; tensor input_363_groups_0 = const()[name = tensor("input_363_groups_0"), val = tensor(1)]; tensor encoder_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84657024)))]; tensor input_361_cast_fp16 = transpose(perm = input_361_perm_0, x = x_173_cast_fp16)[name = tensor("transpose_179")]; tensor input_363_cast_fp16 = conv(dilations = input_363_dilations_0, groups = input_363_groups_0, pad = input_363_pad_0, pad_type = input_363_pad_type_0, strides = input_363_strides_0, weight = encoder_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_361_cast_fp16)[name = tensor("input_363_cast_fp16")]; tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; tensor x_175_split_cast_fp16_0, tensor x_175_split_cast_fp16_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_363_cast_fp16)[name = tensor("x_175_split_cast_fp16")]; tensor x_175_split_1_sigmoid_cast_fp16 = sigmoid(x = x_175_split_cast_fp16_1)[name = tensor("x_175_split_1_sigmoid_cast_fp16")]; tensor x_175_cast_fp16 = mul(x = x_175_split_cast_fp16_0, y = x_175_split_1_sigmoid_cast_fp16)[name = tensor("x_175_cast_fp16")]; tensor input_365_cast_fp16 = select(a = var_40_to_fp16, b = x_175_cast_fp16, cond = var_551)[name = tensor("input_365_cast_fp16")]; tensor new_x_27_interleave_0 = const()[name = tensor("new_x_27_interleave_0"), val = tensor(false)]; tensor new_x_27_cast_fp16 = concat(axis = var_56, interleave = new_x_27_interleave_0, values = (cache_27_cast_fp16, input_365_cast_fp16))[name = tensor("new_x_27_cast_fp16")]; tensor next_cache_13_begin_0 = const()[name = tensor("next_cache_13_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_13_end_0 = const()[name = tensor("next_cache_13_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_13_end_mask_0 = const()[name = tensor("next_cache_13_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_13_cast_fp16 = slice_by_index(begin = next_cache_13_begin_0, end = next_cache_13_end_0, end_mask = next_cache_13_end_mask_0, x = new_x_27_cast_fp16)[name = tensor("next_cache_13_cast_fp16")]; tensor var_1809_begin_0 = const()[name = tensor("op_1809_begin_0"), val = tensor([0, 0, 4])]; tensor var_1809_end_0 = const()[name = tensor("op_1809_end_0"), val = tensor([1, 512, 12])]; tensor var_1809_end_mask_0 = const()[name = tensor("op_1809_end_mask_0"), val = tensor([true, true, true])]; tensor var_1809_cast_fp16 = slice_by_index(begin = var_1809_begin_0, end = var_1809_end_0, end_mask = var_1809_end_mask_0, x = next_cache_13_cast_fp16)[name = tensor("op_1809_cast_fp16")]; tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(512)]; tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; tensor encoder_layers_6_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85705664)))]; tensor x_177_cast_fp16 = conv(dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = encoder_layers_6_conv_depthwise_conv_weight_to_fp16, x = new_x_27_cast_fp16)[name = tensor("x_177_cast_fp16")]; tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; tensor x_179_axes_0 = const()[name = tensor("x_179_axes_0"), val = tensor([-1])]; tensor encoder_layers_6_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85714944)))]; tensor encoder_layers_6_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85716032)))]; tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_177_cast_fp16)[name = tensor("transpose_178")]; tensor x_179_cast_fp16 = layer_norm(axes = x_179_axes_0, beta = encoder_layers_6_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_conv_batch_norm_weight_to_fp16, x = input_367_cast_fp16)[name = tensor("x_179_cast_fp16")]; tensor input_369_perm_0 = const()[name = tensor("input_369_perm_0"), val = tensor([0, 2, 1])]; tensor input_369_cast_fp16 = transpose(perm = input_369_perm_0, x = x_179_cast_fp16)[name = tensor("transpose_177")]; tensor input_371_cast_fp16 = silu(x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; tensor x_181_pad_type_0 = const()[name = tensor("x_181_pad_type_0"), val = tensor("valid")]; tensor x_181_strides_0 = const()[name = tensor("x_181_strides_0"), val = tensor([1])]; tensor x_181_pad_0 = const()[name = tensor("x_181_pad_0"), val = tensor([0, 0])]; tensor x_181_dilations_0 = const()[name = tensor("x_181_dilations_0"), val = tensor([1])]; tensor x_181_groups_0 = const()[name = tensor("x_181_groups_0"), val = tensor(1)]; tensor encoder_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85717120)))]; tensor x_181_cast_fp16 = conv(dilations = x_181_dilations_0, groups = x_181_groups_0, pad = x_181_pad_0, pad_type = x_181_pad_type_0, strides = x_181_strides_0, weight = encoder_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("x_181_cast_fp16")]; tensor input_373_perm_0 = const()[name = tensor("input_373_perm_0"), val = tensor([0, 2, 1])]; tensor input_373_cast_fp16 = transpose(perm = input_373_perm_0, x = x_181_cast_fp16)[name = tensor("transpose_176")]; tensor input_375_cast_fp16 = add(x = input_359_cast_fp16, y = input_373_cast_fp16)[name = tensor("input_375_cast_fp16")]; tensor input_377_axes_0 = const()[name = tensor("input_377_axes_0"), val = tensor([-1])]; tensor encoder_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86241472)))]; tensor encoder_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86242560)))]; tensor input_377_cast_fp16 = layer_norm(axes = input_377_axes_0, beta = encoder_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("input_377_cast_fp16")]; tensor encoder_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86243648)))]; tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_377_cast_fp16)[name = tensor("linear_62_cast_fp16")]; tensor input_381_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_381_cast_fp16")]; tensor encoder_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88340864)))]; tensor linear_63_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("linear_63_cast_fp16")]; tensor var_1850_to_fp16 = const()[name = tensor("op_1850_to_fp16"), val = tensor(0x1p-1)]; tensor var_1851_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1850_to_fp16)[name = tensor("op_1851_cast_fp16")]; tensor input_387_cast_fp16 = add(x = input_375_cast_fp16, y = var_1851_cast_fp16)[name = tensor("input_387_cast_fp16")]; tensor input_389_axes_0 = const()[name = tensor("input_389_axes_0"), val = tensor([-1])]; tensor encoder_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90438080)))]; tensor encoder_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90439168)))]; tensor input_389_cast_fp16 = layer_norm(axes = input_389_axes_0, beta = encoder_layers_6_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_out_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("input_389_cast_fp16")]; tensor cache_29_begin_0 = const()[name = tensor("cache_29_begin_0"), val = tensor([7, 0, 0, 0])]; tensor cache_29_end_0 = const()[name = tensor("cache_29_end_0"), val = tensor([8, 1, 70, 512])]; tensor cache_29_end_mask_0 = const()[name = tensor("cache_29_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_29_squeeze_mask_0 = const()[name = tensor("cache_29_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_29_cast_fp16 = slice_by_index(begin = cache_29_begin_0, end = cache_29_end_0, end_mask = cache_29_end_mask_0, squeeze_mask = cache_29_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_29_cast_fp16")]; tensor cache_31_begin_0 = const()[name = tensor("cache_31_begin_0"), val = tensor([7, 0, 0, 0])]; tensor cache_31_end_0 = const()[name = tensor("cache_31_end_0"), val = tensor([8, 1, 512, 8])]; tensor cache_31_end_mask_0 = const()[name = tensor("cache_31_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_31_squeeze_mask_0 = const()[name = tensor("cache_31_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_31_cast_fp16 = slice_by_index(begin = cache_31_begin_0, end = cache_31_end_0, end_mask = cache_31_end_mask_0, squeeze_mask = cache_31_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_31_cast_fp16")]; tensor input_391_axes_0 = const()[name = tensor("input_391_axes_0"), val = tensor([-1])]; tensor encoder_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90440256)))]; tensor encoder_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90441344)))]; tensor input_391_cast_fp16 = layer_norm(axes = input_391_axes_0, beta = encoder_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward1_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; tensor encoder_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90442432)))]; tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_391_cast_fp16)[name = tensor("linear_64_cast_fp16")]; tensor input_395_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_395_cast_fp16")]; tensor encoder_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92539648)))]; tensor linear_65_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("linear_65_cast_fp16")]; tensor var_1885_to_fp16 = const()[name = tensor("op_1885_to_fp16"), val = tensor(0x1p-1)]; tensor var_1886_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1885_to_fp16)[name = tensor("op_1886_cast_fp16")]; tensor input_401_cast_fp16 = add(x = input_389_cast_fp16, y = var_1886_cast_fp16)[name = tensor("input_401_cast_fp16")]; tensor key_15_axes_0 = const()[name = tensor("key_15_axes_0"), val = tensor([-1])]; tensor encoder_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94636864)))]; tensor encoder_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94637952)))]; tensor key_15_cast_fp16 = layer_norm(axes = key_15_axes_0, beta = encoder_layers_7_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_self_att_weight_to_fp16, x = input_401_cast_fp16)[name = tensor("key_15_cast_fp16")]; tensor input_403_interleave_0 = const()[name = tensor("input_403_interleave_0"), val = tensor(false)]; tensor input_403_cast_fp16 = concat(axis = var_65, interleave = input_403_interleave_0, values = (cache_29_cast_fp16, key_15_cast_fp16))[name = tensor("input_403_cast_fp16")]; tensor var_1908_begin_0 = const()[name = tensor("op_1908_begin_0"), val = tensor([0, 4, 0])]; tensor var_1908_end_0 = const()[name = tensor("op_1908_end_0"), val = tensor([1, 70, 512])]; tensor var_1908_end_mask_0 = const()[name = tensor("op_1908_end_mask_0"), val = tensor([true, true, true])]; tensor var_1908_cast_fp16 = slice_by_index(begin = var_1908_begin_0, end = var_1908_end_0, end_mask = var_1908_end_mask_0, x = cache_29_cast_fp16)[name = tensor("op_1908_cast_fp16")]; tensor var_1911_begin_0 = const()[name = tensor("op_1911_begin_0"), val = tensor([0, 0, 0])]; tensor var_1911_end_0 = const()[name = tensor("op_1911_end_0"), val = tensor([1, 4, 512])]; tensor var_1911_end_mask_0 = const()[name = tensor("op_1911_end_mask_0"), val = tensor([true, false, true])]; tensor var_1911_cast_fp16 = slice_by_index(begin = var_1911_begin_0, end = var_1911_end_0, end_mask = var_1911_end_mask_0, x = key_15_cast_fp16)[name = tensor("op_1911_cast_fp16")]; tensor var_1914_interleave_0 = const()[name = tensor("op_1914_interleave_0"), val = tensor(false)]; tensor var_1914_cast_fp16 = concat(axis = var_65, interleave = var_1914_interleave_0, values = (var_1908_cast_fp16, var_1911_cast_fp16))[name = tensor("op_1914_cast_fp16")]; tensor encoder_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94639040)))]; tensor linear_66_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_q_weight_to_fp16, x = key_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; tensor var_1918 = const()[name = tensor("op_1918"), val = tensor([1, -1, 8, 64])]; tensor q_43_cast_fp16 = reshape(shape = var_1918, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; tensor encoder_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95163392)))]; tensor linear_67_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_k_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_67_cast_fp16")]; tensor var_1922 = const()[name = tensor("op_1922"), val = tensor([1, -1, 8, 64])]; tensor k_29_cast_fp16 = reshape(shape = var_1922, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; tensor encoder_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95687744)))]; tensor linear_68_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_v_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_68_cast_fp16")]; tensor var_1926 = const()[name = tensor("op_1926"), val = tensor([1, -1, 8, 64])]; tensor v_15_cast_fp16 = reshape(shape = var_1926, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96212096)))]; tensor var_1938_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1938_cast_fp16")]; tensor encoder_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96213184)))]; tensor var_1940_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1940_cast_fp16")]; tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; tensor var_1942_to_fp16 = const()[name = tensor("op_1942_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96214272)))]; tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1940_cast_fp16)[name = tensor("transpose_174")]; tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1942_to_fp16)[name = tensor("x_189_cast_fp16")]; tensor x_191_pad_0 = const()[name = tensor("x_191_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_191_mode_0 = const()[name = tensor("x_191_mode_0"), val = tensor("constant")]; tensor const_170_to_fp16 = const()[name = tensor("const_170_to_fp16"), val = tensor(0x0p+0)]; tensor x_191_cast_fp16 = pad(constant_val = const_170_to_fp16, mode = x_191_mode_0, pad = x_191_pad_0, x = x_189_cast_fp16)[name = tensor("x_191_cast_fp16")]; tensor var_1950 = const()[name = tensor("op_1950"), val = tensor([1, 8, -1, 8])]; tensor x_193_cast_fp16 = reshape(shape = var_1950, x = x_191_cast_fp16)[name = tensor("x_193_cast_fp16")]; tensor var_1954_begin_0 = const()[name = tensor("op_1954_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1954_end_0 = const()[name = tensor("op_1954_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_1954_end_mask_0 = const()[name = tensor("op_1954_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1954_cast_fp16 = slice_by_index(begin = var_1954_begin_0, end = var_1954_end_0, end_mask = var_1954_end_mask_0, x = x_193_cast_fp16)[name = tensor("op_1954_cast_fp16")]; tensor var_1955 = const()[name = tensor("op_1955"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1955, x = var_1954_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_172")]; tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1938_cast_fp16)[name = tensor("transpose_173")]; tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; tensor var_1964_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1964_cast_fp16")]; tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_29_cast_fp16 = mul(x = var_1964_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; tensor scores_31_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_11)[name = tensor("scores_31_cast_fp16")]; tensor var_1970_cast_fp16 = softmax(axis = var_56, x = scores_31_cast_fp16)[name = tensor("op_1970_cast_fp16")]; tensor input_405_cast_fp16 = select(a = var_40_to_fp16, b = var_1970_cast_fp16, cond = mask_11)[name = tensor("input_405_cast_fp16")]; tensor x_195_transpose_x_0 = const()[name = tensor("x_195_transpose_x_0"), val = tensor(false)]; tensor x_195_transpose_y_0 = const()[name = tensor("x_195_transpose_y_0"), val = tensor(false)]; tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_175")]; tensor x_195_cast_fp16 = matmul(transpose_x = x_195_transpose_x_0, transpose_y = x_195_transpose_y_0, x = input_405_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_195_cast_fp16")]; tensor var_1974_perm_0 = const()[name = tensor("op_1974_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1975 = const()[name = tensor("op_1975"), val = tensor([1, -1, 512])]; tensor var_1974_cast_fp16 = transpose(perm = var_1974_perm_0, x = x_195_cast_fp16)[name = tensor("transpose_171")]; tensor input_407_cast_fp16 = reshape(shape = var_1975, x = var_1974_cast_fp16)[name = tensor("input_407_cast_fp16")]; tensor encoder_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96373056)))]; tensor linear_70_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_out_weight_to_fp16, x = input_407_cast_fp16)[name = tensor("linear_70_cast_fp16")]; tensor input_411_cast_fp16 = add(x = input_401_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_411_cast_fp16")]; tensor x_199_axes_0 = const()[name = tensor("x_199_axes_0"), val = tensor([-1])]; tensor encoder_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96897408)))]; tensor encoder_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96898496)))]; tensor x_199_cast_fp16 = layer_norm(axes = x_199_axes_0, beta = encoder_layers_7_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_conv_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("x_199_cast_fp16")]; tensor input_413_perm_0 = const()[name = tensor("input_413_perm_0"), val = tensor([0, 2, 1])]; tensor input_415_pad_type_0 = const()[name = tensor("input_415_pad_type_0"), val = tensor("valid")]; tensor input_415_strides_0 = const()[name = tensor("input_415_strides_0"), val = tensor([1])]; tensor input_415_pad_0 = const()[name = tensor("input_415_pad_0"), val = tensor([0, 0])]; tensor input_415_dilations_0 = const()[name = tensor("input_415_dilations_0"), val = tensor([1])]; tensor input_415_groups_0 = const()[name = tensor("input_415_groups_0"), val = tensor(1)]; tensor encoder_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96899584)))]; tensor input_413_cast_fp16 = transpose(perm = input_413_perm_0, x = x_199_cast_fp16)[name = tensor("transpose_170")]; tensor input_415_cast_fp16 = conv(dilations = input_415_dilations_0, groups = input_415_groups_0, pad = input_415_pad_0, pad_type = input_415_pad_type_0, strides = input_415_strides_0, weight = encoder_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_413_cast_fp16)[name = tensor("input_415_cast_fp16")]; tensor x_201_split_num_splits_0 = const()[name = tensor("x_201_split_num_splits_0"), val = tensor(2)]; tensor x_201_split_axis_0 = const()[name = tensor("x_201_split_axis_0"), val = tensor(1)]; tensor x_201_split_cast_fp16_0, tensor x_201_split_cast_fp16_1 = split(axis = x_201_split_axis_0, num_splits = x_201_split_num_splits_0, x = input_415_cast_fp16)[name = tensor("x_201_split_cast_fp16")]; tensor x_201_split_1_sigmoid_cast_fp16 = sigmoid(x = x_201_split_cast_fp16_1)[name = tensor("x_201_split_1_sigmoid_cast_fp16")]; tensor x_201_cast_fp16 = mul(x = x_201_split_cast_fp16_0, y = x_201_split_1_sigmoid_cast_fp16)[name = tensor("x_201_cast_fp16")]; tensor input_417_cast_fp16 = select(a = var_40_to_fp16, b = x_201_cast_fp16, cond = var_551)[name = tensor("input_417_cast_fp16")]; tensor new_x_31_interleave_0 = const()[name = tensor("new_x_31_interleave_0"), val = tensor(false)]; tensor new_x_31_cast_fp16 = concat(axis = var_56, interleave = new_x_31_interleave_0, values = (cache_31_cast_fp16, input_417_cast_fp16))[name = tensor("new_x_31_cast_fp16")]; tensor next_cache_15_begin_0 = const()[name = tensor("next_cache_15_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_15_end_0 = const()[name = tensor("next_cache_15_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_15_end_mask_0 = const()[name = tensor("next_cache_15_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_15_cast_fp16 = slice_by_index(begin = next_cache_15_begin_0, end = next_cache_15_end_0, end_mask = next_cache_15_end_mask_0, x = new_x_31_cast_fp16)[name = tensor("next_cache_15_cast_fp16")]; tensor var_2016_begin_0 = const()[name = tensor("op_2016_begin_0"), val = tensor([0, 0, 4])]; tensor var_2016_end_0 = const()[name = tensor("op_2016_end_0"), val = tensor([1, 512, 12])]; tensor var_2016_end_mask_0 = const()[name = tensor("op_2016_end_mask_0"), val = tensor([true, true, true])]; tensor var_2016_cast_fp16 = slice_by_index(begin = var_2016_begin_0, end = var_2016_end_0, end_mask = var_2016_end_mask_0, x = next_cache_15_cast_fp16)[name = tensor("op_2016_cast_fp16")]; tensor x_203_pad_type_0 = const()[name = tensor("x_203_pad_type_0"), val = tensor("valid")]; tensor x_203_groups_0 = const()[name = tensor("x_203_groups_0"), val = tensor(512)]; tensor x_203_strides_0 = const()[name = tensor("x_203_strides_0"), val = tensor([1])]; tensor x_203_pad_0 = const()[name = tensor("x_203_pad_0"), val = tensor([0, 0])]; tensor x_203_dilations_0 = const()[name = tensor("x_203_dilations_0"), val = tensor([1])]; tensor encoder_layers_7_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97948224)))]; tensor x_203_cast_fp16 = conv(dilations = x_203_dilations_0, groups = x_203_groups_0, pad = x_203_pad_0, pad_type = x_203_pad_type_0, strides = x_203_strides_0, weight = encoder_layers_7_conv_depthwise_conv_weight_to_fp16, x = new_x_31_cast_fp16)[name = tensor("x_203_cast_fp16")]; tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; tensor x_205_axes_0 = const()[name = tensor("x_205_axes_0"), val = tensor([-1])]; tensor encoder_layers_7_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97957504)))]; tensor encoder_layers_7_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97958592)))]; tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_203_cast_fp16)[name = tensor("transpose_169")]; tensor x_205_cast_fp16 = layer_norm(axes = x_205_axes_0, beta = encoder_layers_7_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_conv_batch_norm_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("x_205_cast_fp16")]; tensor input_421_perm_0 = const()[name = tensor("input_421_perm_0"), val = tensor([0, 2, 1])]; tensor input_421_cast_fp16 = transpose(perm = input_421_perm_0, x = x_205_cast_fp16)[name = tensor("transpose_168")]; tensor input_423_cast_fp16 = silu(x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; tensor x_207_pad_type_0 = const()[name = tensor("x_207_pad_type_0"), val = tensor("valid")]; tensor x_207_strides_0 = const()[name = tensor("x_207_strides_0"), val = tensor([1])]; tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0])]; tensor x_207_dilations_0 = const()[name = tensor("x_207_dilations_0"), val = tensor([1])]; tensor x_207_groups_0 = const()[name = tensor("x_207_groups_0"), val = tensor(1)]; tensor encoder_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97959680)))]; tensor x_207_cast_fp16 = conv(dilations = x_207_dilations_0, groups = x_207_groups_0, pad = x_207_pad_0, pad_type = x_207_pad_type_0, strides = x_207_strides_0, weight = encoder_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("x_207_cast_fp16")]; tensor input_425_perm_0 = const()[name = tensor("input_425_perm_0"), val = tensor([0, 2, 1])]; tensor input_425_cast_fp16 = transpose(perm = input_425_perm_0, x = x_207_cast_fp16)[name = tensor("transpose_167")]; tensor input_427_cast_fp16 = add(x = input_411_cast_fp16, y = input_425_cast_fp16)[name = tensor("input_427_cast_fp16")]; tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; tensor encoder_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98484032)))]; tensor encoder_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98485120)))]; tensor input_429_cast_fp16 = layer_norm(axes = input_429_axes_0, beta = encoder_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; tensor encoder_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98486208)))]; tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_429_cast_fp16)[name = tensor("linear_71_cast_fp16")]; tensor input_433_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_433_cast_fp16")]; tensor encoder_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100583424)))]; tensor linear_72_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("linear_72_cast_fp16")]; tensor var_2057_to_fp16 = const()[name = tensor("op_2057_to_fp16"), val = tensor(0x1p-1)]; tensor var_2058_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_2057_to_fp16)[name = tensor("op_2058_cast_fp16")]; tensor input_439_cast_fp16 = add(x = input_427_cast_fp16, y = var_2058_cast_fp16)[name = tensor("input_439_cast_fp16")]; tensor input_441_axes_0 = const()[name = tensor("input_441_axes_0"), val = tensor([-1])]; tensor encoder_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102680640)))]; tensor encoder_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102681728)))]; tensor input_441_cast_fp16 = layer_norm(axes = input_441_axes_0, beta = encoder_layers_7_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_out_weight_to_fp16, x = input_439_cast_fp16)[name = tensor("input_441_cast_fp16")]; tensor cache_33_begin_0 = const()[name = tensor("cache_33_begin_0"), val = tensor([8, 0, 0, 0])]; tensor cache_33_end_0 = const()[name = tensor("cache_33_end_0"), val = tensor([9, 1, 70, 512])]; tensor cache_33_end_mask_0 = const()[name = tensor("cache_33_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_33_squeeze_mask_0 = const()[name = tensor("cache_33_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_33_cast_fp16 = slice_by_index(begin = cache_33_begin_0, end = cache_33_end_0, end_mask = cache_33_end_mask_0, squeeze_mask = cache_33_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_33_cast_fp16")]; tensor cache_35_begin_0 = const()[name = tensor("cache_35_begin_0"), val = tensor([8, 0, 0, 0])]; tensor cache_35_end_0 = const()[name = tensor("cache_35_end_0"), val = tensor([9, 1, 512, 8])]; tensor cache_35_end_mask_0 = const()[name = tensor("cache_35_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_35_squeeze_mask_0 = const()[name = tensor("cache_35_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_35_cast_fp16 = slice_by_index(begin = cache_35_begin_0, end = cache_35_end_0, end_mask = cache_35_end_mask_0, squeeze_mask = cache_35_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_35_cast_fp16")]; tensor input_443_axes_0 = const()[name = tensor("input_443_axes_0"), val = tensor([-1])]; tensor encoder_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102682816)))]; tensor encoder_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102683904)))]; tensor input_443_cast_fp16 = layer_norm(axes = input_443_axes_0, beta = encoder_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward1_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("input_443_cast_fp16")]; tensor encoder_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102684992)))]; tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("linear_73_cast_fp16")]; tensor input_447_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_447_cast_fp16")]; tensor encoder_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104782208)))]; tensor linear_74_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("linear_74_cast_fp16")]; tensor var_2092_to_fp16 = const()[name = tensor("op_2092_to_fp16"), val = tensor(0x1p-1)]; tensor var_2093_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_2092_to_fp16)[name = tensor("op_2093_cast_fp16")]; tensor input_453_cast_fp16 = add(x = input_441_cast_fp16, y = var_2093_cast_fp16)[name = tensor("input_453_cast_fp16")]; tensor key_17_axes_0 = const()[name = tensor("key_17_axes_0"), val = tensor([-1])]; tensor encoder_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106879424)))]; tensor encoder_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106880512)))]; tensor key_17_cast_fp16 = layer_norm(axes = key_17_axes_0, beta = encoder_layers_8_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_self_att_weight_to_fp16, x = input_453_cast_fp16)[name = tensor("key_17_cast_fp16")]; tensor input_455_interleave_0 = const()[name = tensor("input_455_interleave_0"), val = tensor(false)]; tensor input_455_cast_fp16 = concat(axis = var_65, interleave = input_455_interleave_0, values = (cache_33_cast_fp16, key_17_cast_fp16))[name = tensor("input_455_cast_fp16")]; tensor var_2115_begin_0 = const()[name = tensor("op_2115_begin_0"), val = tensor([0, 4, 0])]; tensor var_2115_end_0 = const()[name = tensor("op_2115_end_0"), val = tensor([1, 70, 512])]; tensor var_2115_end_mask_0 = const()[name = tensor("op_2115_end_mask_0"), val = tensor([true, true, true])]; tensor var_2115_cast_fp16 = slice_by_index(begin = var_2115_begin_0, end = var_2115_end_0, end_mask = var_2115_end_mask_0, x = cache_33_cast_fp16)[name = tensor("op_2115_cast_fp16")]; tensor var_2118_begin_0 = const()[name = tensor("op_2118_begin_0"), val = tensor([0, 0, 0])]; tensor var_2118_end_0 = const()[name = tensor("op_2118_end_0"), val = tensor([1, 4, 512])]; tensor var_2118_end_mask_0 = const()[name = tensor("op_2118_end_mask_0"), val = tensor([true, false, true])]; tensor var_2118_cast_fp16 = slice_by_index(begin = var_2118_begin_0, end = var_2118_end_0, end_mask = var_2118_end_mask_0, x = key_17_cast_fp16)[name = tensor("op_2118_cast_fp16")]; tensor var_2121_interleave_0 = const()[name = tensor("op_2121_interleave_0"), val = tensor(false)]; tensor var_2121_cast_fp16 = concat(axis = var_65, interleave = var_2121_interleave_0, values = (var_2115_cast_fp16, var_2118_cast_fp16))[name = tensor("op_2121_cast_fp16")]; tensor encoder_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106881600)))]; tensor linear_75_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_q_weight_to_fp16, x = key_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; tensor var_2125 = const()[name = tensor("op_2125"), val = tensor([1, -1, 8, 64])]; tensor q_49_cast_fp16 = reshape(shape = var_2125, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; tensor encoder_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107405952)))]; tensor linear_76_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_k_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_76_cast_fp16")]; tensor var_2129 = const()[name = tensor("op_2129"), val = tensor([1, -1, 8, 64])]; tensor k_33_cast_fp16 = reshape(shape = var_2129, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; tensor encoder_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107930304)))]; tensor linear_77_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_v_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_77_cast_fp16")]; tensor var_2133 = const()[name = tensor("op_2133"), val = tensor([1, -1, 8, 64])]; tensor v_17_cast_fp16 = reshape(shape = var_2133, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108454656)))]; tensor var_2145_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2145_cast_fp16")]; tensor encoder_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108455744)))]; tensor var_2147_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2147_cast_fp16")]; tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_215_transpose_x_0 = const()[name = tensor("x_215_transpose_x_0"), val = tensor(false)]; tensor x_215_transpose_y_0 = const()[name = tensor("x_215_transpose_y_0"), val = tensor(false)]; tensor var_2149_to_fp16 = const()[name = tensor("op_2149_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108456832)))]; tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_2147_cast_fp16)[name = tensor("transpose_165")]; tensor x_215_cast_fp16 = matmul(transpose_x = x_215_transpose_x_0, transpose_y = x_215_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_2149_to_fp16)[name = tensor("x_215_cast_fp16")]; tensor x_217_pad_0 = const()[name = tensor("x_217_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_217_mode_0 = const()[name = tensor("x_217_mode_0"), val = tensor("constant")]; tensor const_183_to_fp16 = const()[name = tensor("const_183_to_fp16"), val = tensor(0x0p+0)]; tensor x_217_cast_fp16 = pad(constant_val = const_183_to_fp16, mode = x_217_mode_0, pad = x_217_pad_0, x = x_215_cast_fp16)[name = tensor("x_217_cast_fp16")]; tensor var_2157 = const()[name = tensor("op_2157"), val = tensor([1, 8, -1, 8])]; tensor x_219_cast_fp16 = reshape(shape = var_2157, x = x_217_cast_fp16)[name = tensor("x_219_cast_fp16")]; tensor var_2161_begin_0 = const()[name = tensor("op_2161_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2161_end_0 = const()[name = tensor("op_2161_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_2161_end_mask_0 = const()[name = tensor("op_2161_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2161_cast_fp16 = slice_by_index(begin = var_2161_begin_0, end = var_2161_end_0, end_mask = var_2161_end_mask_0, x = x_219_cast_fp16)[name = tensor("op_2161_cast_fp16")]; tensor var_2162 = const()[name = tensor("op_2162"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_33_cast_fp16 = reshape(shape = var_2162, x = var_2161_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_163")]; tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_2145_cast_fp16)[name = tensor("transpose_164")]; tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; tensor var_2171_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_2171_cast_fp16")]; tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_33_cast_fp16 = mul(x = var_2171_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; tensor scores_35_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_11)[name = tensor("scores_35_cast_fp16")]; tensor var_2177_cast_fp16 = softmax(axis = var_56, x = scores_35_cast_fp16)[name = tensor("op_2177_cast_fp16")]; tensor input_457_cast_fp16 = select(a = var_40_to_fp16, b = var_2177_cast_fp16, cond = mask_11)[name = tensor("input_457_cast_fp16")]; tensor x_221_transpose_x_0 = const()[name = tensor("x_221_transpose_x_0"), val = tensor(false)]; tensor x_221_transpose_y_0 = const()[name = tensor("x_221_transpose_y_0"), val = tensor(false)]; tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_166")]; tensor x_221_cast_fp16 = matmul(transpose_x = x_221_transpose_x_0, transpose_y = x_221_transpose_y_0, x = input_457_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_221_cast_fp16")]; tensor var_2181_perm_0 = const()[name = tensor("op_2181_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2182 = const()[name = tensor("op_2182"), val = tensor([1, -1, 512])]; tensor var_2181_cast_fp16 = transpose(perm = var_2181_perm_0, x = x_221_cast_fp16)[name = tensor("transpose_162")]; tensor input_459_cast_fp16 = reshape(shape = var_2182, x = var_2181_cast_fp16)[name = tensor("input_459_cast_fp16")]; tensor encoder_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108615616)))]; tensor linear_79_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_out_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("linear_79_cast_fp16")]; tensor input_463_cast_fp16 = add(x = input_453_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_463_cast_fp16")]; tensor x_225_axes_0 = const()[name = tensor("x_225_axes_0"), val = tensor([-1])]; tensor encoder_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109139968)))]; tensor encoder_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109141056)))]; tensor x_225_cast_fp16 = layer_norm(axes = x_225_axes_0, beta = encoder_layers_8_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_conv_weight_to_fp16, x = input_463_cast_fp16)[name = tensor("x_225_cast_fp16")]; tensor input_465_perm_0 = const()[name = tensor("input_465_perm_0"), val = tensor([0, 2, 1])]; tensor input_467_pad_type_0 = const()[name = tensor("input_467_pad_type_0"), val = tensor("valid")]; tensor input_467_strides_0 = const()[name = tensor("input_467_strides_0"), val = tensor([1])]; tensor input_467_pad_0 = const()[name = tensor("input_467_pad_0"), val = tensor([0, 0])]; tensor input_467_dilations_0 = const()[name = tensor("input_467_dilations_0"), val = tensor([1])]; tensor input_467_groups_0 = const()[name = tensor("input_467_groups_0"), val = tensor(1)]; tensor encoder_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109142144)))]; tensor input_465_cast_fp16 = transpose(perm = input_465_perm_0, x = x_225_cast_fp16)[name = tensor("transpose_161")]; tensor input_467_cast_fp16 = conv(dilations = input_467_dilations_0, groups = input_467_groups_0, pad = input_467_pad_0, pad_type = input_467_pad_type_0, strides = input_467_strides_0, weight = encoder_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_465_cast_fp16)[name = tensor("input_467_cast_fp16")]; tensor x_227_split_num_splits_0 = const()[name = tensor("x_227_split_num_splits_0"), val = tensor(2)]; tensor x_227_split_axis_0 = const()[name = tensor("x_227_split_axis_0"), val = tensor(1)]; tensor x_227_split_cast_fp16_0, tensor x_227_split_cast_fp16_1 = split(axis = x_227_split_axis_0, num_splits = x_227_split_num_splits_0, x = input_467_cast_fp16)[name = tensor("x_227_split_cast_fp16")]; tensor x_227_split_1_sigmoid_cast_fp16 = sigmoid(x = x_227_split_cast_fp16_1)[name = tensor("x_227_split_1_sigmoid_cast_fp16")]; tensor x_227_cast_fp16 = mul(x = x_227_split_cast_fp16_0, y = x_227_split_1_sigmoid_cast_fp16)[name = tensor("x_227_cast_fp16")]; tensor input_469_cast_fp16 = select(a = var_40_to_fp16, b = x_227_cast_fp16, cond = var_551)[name = tensor("input_469_cast_fp16")]; tensor new_x_35_interleave_0 = const()[name = tensor("new_x_35_interleave_0"), val = tensor(false)]; tensor new_x_35_cast_fp16 = concat(axis = var_56, interleave = new_x_35_interleave_0, values = (cache_35_cast_fp16, input_469_cast_fp16))[name = tensor("new_x_35_cast_fp16")]; tensor next_cache_17_begin_0 = const()[name = tensor("next_cache_17_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_17_end_0 = const()[name = tensor("next_cache_17_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_17_end_mask_0 = const()[name = tensor("next_cache_17_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_17_cast_fp16 = slice_by_index(begin = next_cache_17_begin_0, end = next_cache_17_end_0, end_mask = next_cache_17_end_mask_0, x = new_x_35_cast_fp16)[name = tensor("next_cache_17_cast_fp16")]; tensor var_2223_begin_0 = const()[name = tensor("op_2223_begin_0"), val = tensor([0, 0, 4])]; tensor var_2223_end_0 = const()[name = tensor("op_2223_end_0"), val = tensor([1, 512, 12])]; tensor var_2223_end_mask_0 = const()[name = tensor("op_2223_end_mask_0"), val = tensor([true, true, true])]; tensor var_2223_cast_fp16 = slice_by_index(begin = var_2223_begin_0, end = var_2223_end_0, end_mask = var_2223_end_mask_0, x = next_cache_17_cast_fp16)[name = tensor("op_2223_cast_fp16")]; tensor x_229_pad_type_0 = const()[name = tensor("x_229_pad_type_0"), val = tensor("valid")]; tensor x_229_groups_0 = const()[name = tensor("x_229_groups_0"), val = tensor(512)]; tensor x_229_strides_0 = const()[name = tensor("x_229_strides_0"), val = tensor([1])]; tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0])]; tensor x_229_dilations_0 = const()[name = tensor("x_229_dilations_0"), val = tensor([1])]; tensor encoder_layers_8_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110190784)))]; tensor x_229_cast_fp16 = conv(dilations = x_229_dilations_0, groups = x_229_groups_0, pad = x_229_pad_0, pad_type = x_229_pad_type_0, strides = x_229_strides_0, weight = encoder_layers_8_conv_depthwise_conv_weight_to_fp16, x = new_x_35_cast_fp16)[name = tensor("x_229_cast_fp16")]; tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; tensor x_231_axes_0 = const()[name = tensor("x_231_axes_0"), val = tensor([-1])]; tensor encoder_layers_8_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110200064)))]; tensor encoder_layers_8_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110201152)))]; tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_229_cast_fp16)[name = tensor("transpose_160")]; tensor x_231_cast_fp16 = layer_norm(axes = x_231_axes_0, beta = encoder_layers_8_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_conv_batch_norm_weight_to_fp16, x = input_471_cast_fp16)[name = tensor("x_231_cast_fp16")]; tensor input_473_perm_0 = const()[name = tensor("input_473_perm_0"), val = tensor([0, 2, 1])]; tensor input_473_cast_fp16 = transpose(perm = input_473_perm_0, x = x_231_cast_fp16)[name = tensor("transpose_159")]; tensor input_475_cast_fp16 = silu(x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; tensor x_233_pad_type_0 = const()[name = tensor("x_233_pad_type_0"), val = tensor("valid")]; tensor x_233_strides_0 = const()[name = tensor("x_233_strides_0"), val = tensor([1])]; tensor x_233_pad_0 = const()[name = tensor("x_233_pad_0"), val = tensor([0, 0])]; tensor x_233_dilations_0 = const()[name = tensor("x_233_dilations_0"), val = tensor([1])]; tensor x_233_groups_0 = const()[name = tensor("x_233_groups_0"), val = tensor(1)]; tensor encoder_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110202240)))]; tensor x_233_cast_fp16 = conv(dilations = x_233_dilations_0, groups = x_233_groups_0, pad = x_233_pad_0, pad_type = x_233_pad_type_0, strides = x_233_strides_0, weight = encoder_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("x_233_cast_fp16")]; tensor input_477_perm_0 = const()[name = tensor("input_477_perm_0"), val = tensor([0, 2, 1])]; tensor input_477_cast_fp16 = transpose(perm = input_477_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_158")]; tensor input_479_cast_fp16 = add(x = input_463_cast_fp16, y = input_477_cast_fp16)[name = tensor("input_479_cast_fp16")]; tensor input_481_axes_0 = const()[name = tensor("input_481_axes_0"), val = tensor([-1])]; tensor encoder_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110726592)))]; tensor encoder_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110727680)))]; tensor input_481_cast_fp16 = layer_norm(axes = input_481_axes_0, beta = encoder_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("input_481_cast_fp16")]; tensor encoder_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110728768)))]; tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_481_cast_fp16)[name = tensor("linear_80_cast_fp16")]; tensor input_485_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_485_cast_fp16")]; tensor encoder_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112825984)))]; tensor linear_81_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("linear_81_cast_fp16")]; tensor var_2264_to_fp16 = const()[name = tensor("op_2264_to_fp16"), val = tensor(0x1p-1)]; tensor var_2265_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_2264_to_fp16)[name = tensor("op_2265_cast_fp16")]; tensor input_491_cast_fp16 = add(x = input_479_cast_fp16, y = var_2265_cast_fp16)[name = tensor("input_491_cast_fp16")]; tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; tensor encoder_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114923200)))]; tensor encoder_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114924288)))]; tensor input_493_cast_fp16 = layer_norm(axes = input_493_axes_0, beta = encoder_layers_8_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_out_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("input_493_cast_fp16")]; tensor cache_37_begin_0 = const()[name = tensor("cache_37_begin_0"), val = tensor([9, 0, 0, 0])]; tensor cache_37_end_0 = const()[name = tensor("cache_37_end_0"), val = tensor([10, 1, 70, 512])]; tensor cache_37_end_mask_0 = const()[name = tensor("cache_37_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_37_squeeze_mask_0 = const()[name = tensor("cache_37_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_37_cast_fp16 = slice_by_index(begin = cache_37_begin_0, end = cache_37_end_0, end_mask = cache_37_end_mask_0, squeeze_mask = cache_37_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_37_cast_fp16")]; tensor cache_39_begin_0 = const()[name = tensor("cache_39_begin_0"), val = tensor([9, 0, 0, 0])]; tensor cache_39_end_0 = const()[name = tensor("cache_39_end_0"), val = tensor([10, 1, 512, 8])]; tensor cache_39_end_mask_0 = const()[name = tensor("cache_39_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_39_squeeze_mask_0 = const()[name = tensor("cache_39_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_39_cast_fp16 = slice_by_index(begin = cache_39_begin_0, end = cache_39_end_0, end_mask = cache_39_end_mask_0, squeeze_mask = cache_39_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_39_cast_fp16")]; tensor input_495_axes_0 = const()[name = tensor("input_495_axes_0"), val = tensor([-1])]; tensor encoder_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114925376)))]; tensor encoder_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114926464)))]; tensor input_495_cast_fp16 = layer_norm(axes = input_495_axes_0, beta = encoder_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward1_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("input_495_cast_fp16")]; tensor encoder_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114927552)))]; tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_495_cast_fp16)[name = tensor("linear_82_cast_fp16")]; tensor input_499_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_499_cast_fp16")]; tensor encoder_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117024768)))]; tensor linear_83_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("linear_83_cast_fp16")]; tensor var_2299_to_fp16 = const()[name = tensor("op_2299_to_fp16"), val = tensor(0x1p-1)]; tensor var_2300_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_2299_to_fp16)[name = tensor("op_2300_cast_fp16")]; tensor input_505_cast_fp16 = add(x = input_493_cast_fp16, y = var_2300_cast_fp16)[name = tensor("input_505_cast_fp16")]; tensor key_19_axes_0 = const()[name = tensor("key_19_axes_0"), val = tensor([-1])]; tensor encoder_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119121984)))]; tensor encoder_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119123072)))]; tensor key_19_cast_fp16 = layer_norm(axes = key_19_axes_0, beta = encoder_layers_9_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_self_att_weight_to_fp16, x = input_505_cast_fp16)[name = tensor("key_19_cast_fp16")]; tensor input_507_interleave_0 = const()[name = tensor("input_507_interleave_0"), val = tensor(false)]; tensor input_507_cast_fp16 = concat(axis = var_65, interleave = input_507_interleave_0, values = (cache_37_cast_fp16, key_19_cast_fp16))[name = tensor("input_507_cast_fp16")]; tensor var_2322_begin_0 = const()[name = tensor("op_2322_begin_0"), val = tensor([0, 4, 0])]; tensor var_2322_end_0 = const()[name = tensor("op_2322_end_0"), val = tensor([1, 70, 512])]; tensor var_2322_end_mask_0 = const()[name = tensor("op_2322_end_mask_0"), val = tensor([true, true, true])]; tensor var_2322_cast_fp16 = slice_by_index(begin = var_2322_begin_0, end = var_2322_end_0, end_mask = var_2322_end_mask_0, x = cache_37_cast_fp16)[name = tensor("op_2322_cast_fp16")]; tensor var_2325_begin_0 = const()[name = tensor("op_2325_begin_0"), val = tensor([0, 0, 0])]; tensor var_2325_end_0 = const()[name = tensor("op_2325_end_0"), val = tensor([1, 4, 512])]; tensor var_2325_end_mask_0 = const()[name = tensor("op_2325_end_mask_0"), val = tensor([true, false, true])]; tensor var_2325_cast_fp16 = slice_by_index(begin = var_2325_begin_0, end = var_2325_end_0, end_mask = var_2325_end_mask_0, x = key_19_cast_fp16)[name = tensor("op_2325_cast_fp16")]; tensor var_2328_interleave_0 = const()[name = tensor("op_2328_interleave_0"), val = tensor(false)]; tensor var_2328_cast_fp16 = concat(axis = var_65, interleave = var_2328_interleave_0, values = (var_2322_cast_fp16, var_2325_cast_fp16))[name = tensor("op_2328_cast_fp16")]; tensor encoder_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119124160)))]; tensor linear_84_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_q_weight_to_fp16, x = key_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; tensor var_2332 = const()[name = tensor("op_2332"), val = tensor([1, -1, 8, 64])]; tensor q_55_cast_fp16 = reshape(shape = var_2332, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; tensor encoder_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119648512)))]; tensor linear_85_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_k_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_85_cast_fp16")]; tensor var_2336 = const()[name = tensor("op_2336"), val = tensor([1, -1, 8, 64])]; tensor k_37_cast_fp16 = reshape(shape = var_2336, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; tensor encoder_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120172864)))]; tensor linear_86_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_v_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_86_cast_fp16")]; tensor var_2340 = const()[name = tensor("op_2340"), val = tensor([1, -1, 8, 64])]; tensor v_19_cast_fp16 = reshape(shape = var_2340, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120697216)))]; tensor var_2352_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2352_cast_fp16")]; tensor encoder_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120698304)))]; tensor var_2354_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2354_cast_fp16")]; tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_241_transpose_x_0 = const()[name = tensor("x_241_transpose_x_0"), val = tensor(false)]; tensor x_241_transpose_y_0 = const()[name = tensor("x_241_transpose_y_0"), val = tensor(false)]; tensor var_2356_to_fp16 = const()[name = tensor("op_2356_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120699392)))]; tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_2354_cast_fp16)[name = tensor("transpose_156")]; tensor x_241_cast_fp16 = matmul(transpose_x = x_241_transpose_x_0, transpose_y = x_241_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_2356_to_fp16)[name = tensor("x_241_cast_fp16")]; tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_243_mode_0 = const()[name = tensor("x_243_mode_0"), val = tensor("constant")]; tensor const_196_to_fp16 = const()[name = tensor("const_196_to_fp16"), val = tensor(0x0p+0)]; tensor x_243_cast_fp16 = pad(constant_val = const_196_to_fp16, mode = x_243_mode_0, pad = x_243_pad_0, x = x_241_cast_fp16)[name = tensor("x_243_cast_fp16")]; tensor var_2364 = const()[name = tensor("op_2364"), val = tensor([1, 8, -1, 8])]; tensor x_245_cast_fp16 = reshape(shape = var_2364, x = x_243_cast_fp16)[name = tensor("x_245_cast_fp16")]; tensor var_2368_begin_0 = const()[name = tensor("op_2368_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2368_end_0 = const()[name = tensor("op_2368_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_2368_end_mask_0 = const()[name = tensor("op_2368_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2368_cast_fp16 = slice_by_index(begin = var_2368_begin_0, end = var_2368_end_0, end_mask = var_2368_end_mask_0, x = x_245_cast_fp16)[name = tensor("op_2368_cast_fp16")]; tensor var_2369 = const()[name = tensor("op_2369"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_37_cast_fp16 = reshape(shape = var_2369, x = var_2368_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_154")]; tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_2352_cast_fp16)[name = tensor("transpose_155")]; tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; tensor var_2378_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_2378_cast_fp16")]; tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_37_cast_fp16 = mul(x = var_2378_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; tensor scores_39_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_11)[name = tensor("scores_39_cast_fp16")]; tensor var_2384_cast_fp16 = softmax(axis = var_56, x = scores_39_cast_fp16)[name = tensor("op_2384_cast_fp16")]; tensor input_509_cast_fp16 = select(a = var_40_to_fp16, b = var_2384_cast_fp16, cond = mask_11)[name = tensor("input_509_cast_fp16")]; tensor x_247_transpose_x_0 = const()[name = tensor("x_247_transpose_x_0"), val = tensor(false)]; tensor x_247_transpose_y_0 = const()[name = tensor("x_247_transpose_y_0"), val = tensor(false)]; tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_157")]; tensor x_247_cast_fp16 = matmul(transpose_x = x_247_transpose_x_0, transpose_y = x_247_transpose_y_0, x = input_509_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_247_cast_fp16")]; tensor var_2388_perm_0 = const()[name = tensor("op_2388_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2389 = const()[name = tensor("op_2389"), val = tensor([1, -1, 512])]; tensor var_2388_cast_fp16 = transpose(perm = var_2388_perm_0, x = x_247_cast_fp16)[name = tensor("transpose_153")]; tensor input_511_cast_fp16 = reshape(shape = var_2389, x = var_2388_cast_fp16)[name = tensor("input_511_cast_fp16")]; tensor encoder_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120858176)))]; tensor linear_88_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_out_weight_to_fp16, x = input_511_cast_fp16)[name = tensor("linear_88_cast_fp16")]; tensor input_515_cast_fp16 = add(x = input_505_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_515_cast_fp16")]; tensor x_251_axes_0 = const()[name = tensor("x_251_axes_0"), val = tensor([-1])]; tensor encoder_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121382528)))]; tensor encoder_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121383616)))]; tensor x_251_cast_fp16 = layer_norm(axes = x_251_axes_0, beta = encoder_layers_9_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_conv_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("x_251_cast_fp16")]; tensor input_517_perm_0 = const()[name = tensor("input_517_perm_0"), val = tensor([0, 2, 1])]; tensor input_519_pad_type_0 = const()[name = tensor("input_519_pad_type_0"), val = tensor("valid")]; tensor input_519_strides_0 = const()[name = tensor("input_519_strides_0"), val = tensor([1])]; tensor input_519_pad_0 = const()[name = tensor("input_519_pad_0"), val = tensor([0, 0])]; tensor input_519_dilations_0 = const()[name = tensor("input_519_dilations_0"), val = tensor([1])]; tensor input_519_groups_0 = const()[name = tensor("input_519_groups_0"), val = tensor(1)]; tensor encoder_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121384704)))]; tensor input_517_cast_fp16 = transpose(perm = input_517_perm_0, x = x_251_cast_fp16)[name = tensor("transpose_152")]; tensor input_519_cast_fp16 = conv(dilations = input_519_dilations_0, groups = input_519_groups_0, pad = input_519_pad_0, pad_type = input_519_pad_type_0, strides = input_519_strides_0, weight = encoder_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_517_cast_fp16)[name = tensor("input_519_cast_fp16")]; tensor x_253_split_num_splits_0 = const()[name = tensor("x_253_split_num_splits_0"), val = tensor(2)]; tensor x_253_split_axis_0 = const()[name = tensor("x_253_split_axis_0"), val = tensor(1)]; tensor x_253_split_cast_fp16_0, tensor x_253_split_cast_fp16_1 = split(axis = x_253_split_axis_0, num_splits = x_253_split_num_splits_0, x = input_519_cast_fp16)[name = tensor("x_253_split_cast_fp16")]; tensor x_253_split_1_sigmoid_cast_fp16 = sigmoid(x = x_253_split_cast_fp16_1)[name = tensor("x_253_split_1_sigmoid_cast_fp16")]; tensor x_253_cast_fp16 = mul(x = x_253_split_cast_fp16_0, y = x_253_split_1_sigmoid_cast_fp16)[name = tensor("x_253_cast_fp16")]; tensor input_521_cast_fp16 = select(a = var_40_to_fp16, b = x_253_cast_fp16, cond = var_551)[name = tensor("input_521_cast_fp16")]; tensor new_x_39_interleave_0 = const()[name = tensor("new_x_39_interleave_0"), val = tensor(false)]; tensor new_x_39_cast_fp16 = concat(axis = var_56, interleave = new_x_39_interleave_0, values = (cache_39_cast_fp16, input_521_cast_fp16))[name = tensor("new_x_39_cast_fp16")]; tensor next_cache_19_begin_0 = const()[name = tensor("next_cache_19_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_19_end_0 = const()[name = tensor("next_cache_19_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_19_end_mask_0 = const()[name = tensor("next_cache_19_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_19_cast_fp16 = slice_by_index(begin = next_cache_19_begin_0, end = next_cache_19_end_0, end_mask = next_cache_19_end_mask_0, x = new_x_39_cast_fp16)[name = tensor("next_cache_19_cast_fp16")]; tensor var_2430_begin_0 = const()[name = tensor("op_2430_begin_0"), val = tensor([0, 0, 4])]; tensor var_2430_end_0 = const()[name = tensor("op_2430_end_0"), val = tensor([1, 512, 12])]; tensor var_2430_end_mask_0 = const()[name = tensor("op_2430_end_mask_0"), val = tensor([true, true, true])]; tensor var_2430_cast_fp16 = slice_by_index(begin = var_2430_begin_0, end = var_2430_end_0, end_mask = var_2430_end_mask_0, x = next_cache_19_cast_fp16)[name = tensor("op_2430_cast_fp16")]; tensor x_255_pad_type_0 = const()[name = tensor("x_255_pad_type_0"), val = tensor("valid")]; tensor x_255_groups_0 = const()[name = tensor("x_255_groups_0"), val = tensor(512)]; tensor x_255_strides_0 = const()[name = tensor("x_255_strides_0"), val = tensor([1])]; tensor x_255_pad_0 = const()[name = tensor("x_255_pad_0"), val = tensor([0, 0])]; tensor x_255_dilations_0 = const()[name = tensor("x_255_dilations_0"), val = tensor([1])]; tensor encoder_layers_9_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122433344)))]; tensor x_255_cast_fp16 = conv(dilations = x_255_dilations_0, groups = x_255_groups_0, pad = x_255_pad_0, pad_type = x_255_pad_type_0, strides = x_255_strides_0, weight = encoder_layers_9_conv_depthwise_conv_weight_to_fp16, x = new_x_39_cast_fp16)[name = tensor("x_255_cast_fp16")]; tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; tensor x_257_axes_0 = const()[name = tensor("x_257_axes_0"), val = tensor([-1])]; tensor encoder_layers_9_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122442624)))]; tensor encoder_layers_9_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122443712)))]; tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_151")]; tensor x_257_cast_fp16 = layer_norm(axes = x_257_axes_0, beta = encoder_layers_9_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_conv_batch_norm_weight_to_fp16, x = input_523_cast_fp16)[name = tensor("x_257_cast_fp16")]; tensor input_525_perm_0 = const()[name = tensor("input_525_perm_0"), val = tensor([0, 2, 1])]; tensor input_525_cast_fp16 = transpose(perm = input_525_perm_0, x = x_257_cast_fp16)[name = tensor("transpose_150")]; tensor input_527_cast_fp16 = silu(x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; tensor x_259_pad_type_0 = const()[name = tensor("x_259_pad_type_0"), val = tensor("valid")]; tensor x_259_strides_0 = const()[name = tensor("x_259_strides_0"), val = tensor([1])]; tensor x_259_pad_0 = const()[name = tensor("x_259_pad_0"), val = tensor([0, 0])]; tensor x_259_dilations_0 = const()[name = tensor("x_259_dilations_0"), val = tensor([1])]; tensor x_259_groups_0 = const()[name = tensor("x_259_groups_0"), val = tensor(1)]; tensor encoder_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122444800)))]; tensor x_259_cast_fp16 = conv(dilations = x_259_dilations_0, groups = x_259_groups_0, pad = x_259_pad_0, pad_type = x_259_pad_type_0, strides = x_259_strides_0, weight = encoder_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("x_259_cast_fp16")]; tensor input_529_perm_0 = const()[name = tensor("input_529_perm_0"), val = tensor([0, 2, 1])]; tensor input_529_cast_fp16 = transpose(perm = input_529_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_149")]; tensor input_531_cast_fp16 = add(x = input_515_cast_fp16, y = input_529_cast_fp16)[name = tensor("input_531_cast_fp16")]; tensor input_533_axes_0 = const()[name = tensor("input_533_axes_0"), val = tensor([-1])]; tensor encoder_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122969152)))]; tensor encoder_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122970240)))]; tensor input_533_cast_fp16 = layer_norm(axes = input_533_axes_0, beta = encoder_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("input_533_cast_fp16")]; tensor encoder_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122971328)))]; tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_533_cast_fp16)[name = tensor("linear_89_cast_fp16")]; tensor input_537_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_537_cast_fp16")]; tensor encoder_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125068544)))]; tensor linear_90_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("linear_90_cast_fp16")]; tensor var_2471_to_fp16 = const()[name = tensor("op_2471_to_fp16"), val = tensor(0x1p-1)]; tensor var_2472_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2471_to_fp16)[name = tensor("op_2472_cast_fp16")]; tensor input_543_cast_fp16 = add(x = input_531_cast_fp16, y = var_2472_cast_fp16)[name = tensor("input_543_cast_fp16")]; tensor input_545_axes_0 = const()[name = tensor("input_545_axes_0"), val = tensor([-1])]; tensor encoder_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127165760)))]; tensor encoder_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127166848)))]; tensor input_545_cast_fp16 = layer_norm(axes = input_545_axes_0, beta = encoder_layers_9_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_out_weight_to_fp16, x = input_543_cast_fp16)[name = tensor("input_545_cast_fp16")]; tensor cache_41_begin_0 = const()[name = tensor("cache_41_begin_0"), val = tensor([10, 0, 0, 0])]; tensor cache_41_end_0 = const()[name = tensor("cache_41_end_0"), val = tensor([11, 1, 70, 512])]; tensor cache_41_end_mask_0 = const()[name = tensor("cache_41_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_41_squeeze_mask_0 = const()[name = tensor("cache_41_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_41_cast_fp16 = slice_by_index(begin = cache_41_begin_0, end = cache_41_end_0, end_mask = cache_41_end_mask_0, squeeze_mask = cache_41_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_41_cast_fp16")]; tensor cache_43_begin_0 = const()[name = tensor("cache_43_begin_0"), val = tensor([10, 0, 0, 0])]; tensor cache_43_end_0 = const()[name = tensor("cache_43_end_0"), val = tensor([11, 1, 512, 8])]; tensor cache_43_end_mask_0 = const()[name = tensor("cache_43_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_43_squeeze_mask_0 = const()[name = tensor("cache_43_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_43_cast_fp16 = slice_by_index(begin = cache_43_begin_0, end = cache_43_end_0, end_mask = cache_43_end_mask_0, squeeze_mask = cache_43_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_43_cast_fp16")]; tensor input_547_axes_0 = const()[name = tensor("input_547_axes_0"), val = tensor([-1])]; tensor encoder_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127167936)))]; tensor encoder_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127169024)))]; tensor input_547_cast_fp16 = layer_norm(axes = input_547_axes_0, beta = encoder_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward1_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("input_547_cast_fp16")]; tensor encoder_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127170112)))]; tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_547_cast_fp16)[name = tensor("linear_91_cast_fp16")]; tensor input_551_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_551_cast_fp16")]; tensor encoder_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129267328)))]; tensor linear_92_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("linear_92_cast_fp16")]; tensor var_2506_to_fp16 = const()[name = tensor("op_2506_to_fp16"), val = tensor(0x1p-1)]; tensor var_2507_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2506_to_fp16)[name = tensor("op_2507_cast_fp16")]; tensor input_557_cast_fp16 = add(x = input_545_cast_fp16, y = var_2507_cast_fp16)[name = tensor("input_557_cast_fp16")]; tensor key_21_axes_0 = const()[name = tensor("key_21_axes_0"), val = tensor([-1])]; tensor encoder_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131364544)))]; tensor encoder_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131365632)))]; tensor key_21_cast_fp16 = layer_norm(axes = key_21_axes_0, beta = encoder_layers_10_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_self_att_weight_to_fp16, x = input_557_cast_fp16)[name = tensor("key_21_cast_fp16")]; tensor input_559_interleave_0 = const()[name = tensor("input_559_interleave_0"), val = tensor(false)]; tensor input_559_cast_fp16 = concat(axis = var_65, interleave = input_559_interleave_0, values = (cache_41_cast_fp16, key_21_cast_fp16))[name = tensor("input_559_cast_fp16")]; tensor var_2529_begin_0 = const()[name = tensor("op_2529_begin_0"), val = tensor([0, 4, 0])]; tensor var_2529_end_0 = const()[name = tensor("op_2529_end_0"), val = tensor([1, 70, 512])]; tensor var_2529_end_mask_0 = const()[name = tensor("op_2529_end_mask_0"), val = tensor([true, true, true])]; tensor var_2529_cast_fp16 = slice_by_index(begin = var_2529_begin_0, end = var_2529_end_0, end_mask = var_2529_end_mask_0, x = cache_41_cast_fp16)[name = tensor("op_2529_cast_fp16")]; tensor var_2532_begin_0 = const()[name = tensor("op_2532_begin_0"), val = tensor([0, 0, 0])]; tensor var_2532_end_0 = const()[name = tensor("op_2532_end_0"), val = tensor([1, 4, 512])]; tensor var_2532_end_mask_0 = const()[name = tensor("op_2532_end_mask_0"), val = tensor([true, false, true])]; tensor var_2532_cast_fp16 = slice_by_index(begin = var_2532_begin_0, end = var_2532_end_0, end_mask = var_2532_end_mask_0, x = key_21_cast_fp16)[name = tensor("op_2532_cast_fp16")]; tensor var_2535_interleave_0 = const()[name = tensor("op_2535_interleave_0"), val = tensor(false)]; tensor var_2535_cast_fp16 = concat(axis = var_65, interleave = var_2535_interleave_0, values = (var_2529_cast_fp16, var_2532_cast_fp16))[name = tensor("op_2535_cast_fp16")]; tensor encoder_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131366720)))]; tensor linear_93_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_q_weight_to_fp16, x = key_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; tensor var_2539 = const()[name = tensor("op_2539"), val = tensor([1, -1, 8, 64])]; tensor q_61_cast_fp16 = reshape(shape = var_2539, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; tensor encoder_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131891072)))]; tensor linear_94_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_k_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_94_cast_fp16")]; tensor var_2543 = const()[name = tensor("op_2543"), val = tensor([1, -1, 8, 64])]; tensor k_41_cast_fp16 = reshape(shape = var_2543, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; tensor encoder_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132415424)))]; tensor linear_95_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_v_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_95_cast_fp16")]; tensor var_2547 = const()[name = tensor("op_2547"), val = tensor([1, -1, 8, 64])]; tensor v_21_cast_fp16 = reshape(shape = var_2547, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132939776)))]; tensor var_2559_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2559_cast_fp16")]; tensor encoder_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132940864)))]; tensor var_2561_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2561_cast_fp16")]; tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_267_transpose_x_0 = const()[name = tensor("x_267_transpose_x_0"), val = tensor(false)]; tensor x_267_transpose_y_0 = const()[name = tensor("x_267_transpose_y_0"), val = tensor(false)]; tensor var_2563_to_fp16 = const()[name = tensor("op_2563_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132941952)))]; tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2561_cast_fp16)[name = tensor("transpose_147")]; tensor x_267_cast_fp16 = matmul(transpose_x = x_267_transpose_x_0, transpose_y = x_267_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2563_to_fp16)[name = tensor("x_267_cast_fp16")]; tensor x_269_pad_0 = const()[name = tensor("x_269_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_269_mode_0 = const()[name = tensor("x_269_mode_0"), val = tensor("constant")]; tensor const_209_to_fp16 = const()[name = tensor("const_209_to_fp16"), val = tensor(0x0p+0)]; tensor x_269_cast_fp16 = pad(constant_val = const_209_to_fp16, mode = x_269_mode_0, pad = x_269_pad_0, x = x_267_cast_fp16)[name = tensor("x_269_cast_fp16")]; tensor var_2571 = const()[name = tensor("op_2571"), val = tensor([1, 8, -1, 8])]; tensor x_271_cast_fp16 = reshape(shape = var_2571, x = x_269_cast_fp16)[name = tensor("x_271_cast_fp16")]; tensor var_2575_begin_0 = const()[name = tensor("op_2575_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2575_end_0 = const()[name = tensor("op_2575_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_2575_end_mask_0 = const()[name = tensor("op_2575_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2575_cast_fp16 = slice_by_index(begin = var_2575_begin_0, end = var_2575_end_0, end_mask = var_2575_end_mask_0, x = x_271_cast_fp16)[name = tensor("op_2575_cast_fp16")]; tensor var_2576 = const()[name = tensor("op_2576"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2576, x = var_2575_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_145")]; tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2559_cast_fp16)[name = tensor("transpose_146")]; tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; tensor var_2585_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2585_cast_fp16")]; tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_41_cast_fp16 = mul(x = var_2585_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; tensor scores_43_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_11)[name = tensor("scores_43_cast_fp16")]; tensor var_2591_cast_fp16 = softmax(axis = var_56, x = scores_43_cast_fp16)[name = tensor("op_2591_cast_fp16")]; tensor input_561_cast_fp16 = select(a = var_40_to_fp16, b = var_2591_cast_fp16, cond = mask_11)[name = tensor("input_561_cast_fp16")]; tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_148")]; tensor x_273_cast_fp16 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = input_561_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_273_cast_fp16")]; tensor var_2595_perm_0 = const()[name = tensor("op_2595_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2596 = const()[name = tensor("op_2596"), val = tensor([1, -1, 512])]; tensor var_2595_cast_fp16 = transpose(perm = var_2595_perm_0, x = x_273_cast_fp16)[name = tensor("transpose_144")]; tensor input_563_cast_fp16 = reshape(shape = var_2596, x = var_2595_cast_fp16)[name = tensor("input_563_cast_fp16")]; tensor encoder_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133100736)))]; tensor linear_97_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_out_weight_to_fp16, x = input_563_cast_fp16)[name = tensor("linear_97_cast_fp16")]; tensor input_567_cast_fp16 = add(x = input_557_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_567_cast_fp16")]; tensor x_277_axes_0 = const()[name = tensor("x_277_axes_0"), val = tensor([-1])]; tensor encoder_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133625088)))]; tensor encoder_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133626176)))]; tensor x_277_cast_fp16 = layer_norm(axes = x_277_axes_0, beta = encoder_layers_10_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_conv_weight_to_fp16, x = input_567_cast_fp16)[name = tensor("x_277_cast_fp16")]; tensor input_569_perm_0 = const()[name = tensor("input_569_perm_0"), val = tensor([0, 2, 1])]; tensor input_571_pad_type_0 = const()[name = tensor("input_571_pad_type_0"), val = tensor("valid")]; tensor input_571_strides_0 = const()[name = tensor("input_571_strides_0"), val = tensor([1])]; tensor input_571_pad_0 = const()[name = tensor("input_571_pad_0"), val = tensor([0, 0])]; tensor input_571_dilations_0 = const()[name = tensor("input_571_dilations_0"), val = tensor([1])]; tensor input_571_groups_0 = const()[name = tensor("input_571_groups_0"), val = tensor(1)]; tensor encoder_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133627264)))]; tensor input_569_cast_fp16 = transpose(perm = input_569_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_143")]; tensor input_571_cast_fp16 = conv(dilations = input_571_dilations_0, groups = input_571_groups_0, pad = input_571_pad_0, pad_type = input_571_pad_type_0, strides = input_571_strides_0, weight = encoder_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_569_cast_fp16)[name = tensor("input_571_cast_fp16")]; tensor x_279_split_num_splits_0 = const()[name = tensor("x_279_split_num_splits_0"), val = tensor(2)]; tensor x_279_split_axis_0 = const()[name = tensor("x_279_split_axis_0"), val = tensor(1)]; tensor x_279_split_cast_fp16_0, tensor x_279_split_cast_fp16_1 = split(axis = x_279_split_axis_0, num_splits = x_279_split_num_splits_0, x = input_571_cast_fp16)[name = tensor("x_279_split_cast_fp16")]; tensor x_279_split_1_sigmoid_cast_fp16 = sigmoid(x = x_279_split_cast_fp16_1)[name = tensor("x_279_split_1_sigmoid_cast_fp16")]; tensor x_279_cast_fp16 = mul(x = x_279_split_cast_fp16_0, y = x_279_split_1_sigmoid_cast_fp16)[name = tensor("x_279_cast_fp16")]; tensor input_573_cast_fp16 = select(a = var_40_to_fp16, b = x_279_cast_fp16, cond = var_551)[name = tensor("input_573_cast_fp16")]; tensor new_x_43_interleave_0 = const()[name = tensor("new_x_43_interleave_0"), val = tensor(false)]; tensor new_x_43_cast_fp16 = concat(axis = var_56, interleave = new_x_43_interleave_0, values = (cache_43_cast_fp16, input_573_cast_fp16))[name = tensor("new_x_43_cast_fp16")]; tensor next_cache_21_begin_0 = const()[name = tensor("next_cache_21_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_21_end_0 = const()[name = tensor("next_cache_21_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_21_end_mask_0 = const()[name = tensor("next_cache_21_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_21_cast_fp16 = slice_by_index(begin = next_cache_21_begin_0, end = next_cache_21_end_0, end_mask = next_cache_21_end_mask_0, x = new_x_43_cast_fp16)[name = tensor("next_cache_21_cast_fp16")]; tensor var_2637_begin_0 = const()[name = tensor("op_2637_begin_0"), val = tensor([0, 0, 4])]; tensor var_2637_end_0 = const()[name = tensor("op_2637_end_0"), val = tensor([1, 512, 12])]; tensor var_2637_end_mask_0 = const()[name = tensor("op_2637_end_mask_0"), val = tensor([true, true, true])]; tensor var_2637_cast_fp16 = slice_by_index(begin = var_2637_begin_0, end = var_2637_end_0, end_mask = var_2637_end_mask_0, x = next_cache_21_cast_fp16)[name = tensor("op_2637_cast_fp16")]; tensor x_281_pad_type_0 = const()[name = tensor("x_281_pad_type_0"), val = tensor("valid")]; tensor x_281_groups_0 = const()[name = tensor("x_281_groups_0"), val = tensor(512)]; tensor x_281_strides_0 = const()[name = tensor("x_281_strides_0"), val = tensor([1])]; tensor x_281_pad_0 = const()[name = tensor("x_281_pad_0"), val = tensor([0, 0])]; tensor x_281_dilations_0 = const()[name = tensor("x_281_dilations_0"), val = tensor([1])]; tensor encoder_layers_10_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134675904)))]; tensor x_281_cast_fp16 = conv(dilations = x_281_dilations_0, groups = x_281_groups_0, pad = x_281_pad_0, pad_type = x_281_pad_type_0, strides = x_281_strides_0, weight = encoder_layers_10_conv_depthwise_conv_weight_to_fp16, x = new_x_43_cast_fp16)[name = tensor("x_281_cast_fp16")]; tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; tensor encoder_layers_10_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134685184)))]; tensor encoder_layers_10_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134686272)))]; tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_142")]; tensor x_283_cast_fp16 = layer_norm(axes = x_283_axes_0, beta = encoder_layers_10_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_conv_batch_norm_weight_to_fp16, x = input_575_cast_fp16)[name = tensor("x_283_cast_fp16")]; tensor input_577_perm_0 = const()[name = tensor("input_577_perm_0"), val = tensor([0, 2, 1])]; tensor input_577_cast_fp16 = transpose(perm = input_577_perm_0, x = x_283_cast_fp16)[name = tensor("transpose_141")]; tensor input_579_cast_fp16 = silu(x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; tensor encoder_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134687360)))]; tensor x_285_cast_fp16 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = encoder_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("x_285_cast_fp16")]; tensor input_581_perm_0 = const()[name = tensor("input_581_perm_0"), val = tensor([0, 2, 1])]; tensor input_581_cast_fp16 = transpose(perm = input_581_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_140")]; tensor input_583_cast_fp16 = add(x = input_567_cast_fp16, y = input_581_cast_fp16)[name = tensor("input_583_cast_fp16")]; tensor input_585_axes_0 = const()[name = tensor("input_585_axes_0"), val = tensor([-1])]; tensor encoder_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135211712)))]; tensor encoder_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135212800)))]; tensor input_585_cast_fp16 = layer_norm(axes = input_585_axes_0, beta = encoder_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("input_585_cast_fp16")]; tensor encoder_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135213888)))]; tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_585_cast_fp16)[name = tensor("linear_98_cast_fp16")]; tensor input_589_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_589_cast_fp16")]; tensor encoder_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137311104)))]; tensor linear_99_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("linear_99_cast_fp16")]; tensor var_2678_to_fp16 = const()[name = tensor("op_2678_to_fp16"), val = tensor(0x1p-1)]; tensor var_2679_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2678_to_fp16)[name = tensor("op_2679_cast_fp16")]; tensor input_595_cast_fp16 = add(x = input_583_cast_fp16, y = var_2679_cast_fp16)[name = tensor("input_595_cast_fp16")]; tensor input_597_axes_0 = const()[name = tensor("input_597_axes_0"), val = tensor([-1])]; tensor encoder_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139408320)))]; tensor encoder_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139409408)))]; tensor input_597_cast_fp16 = layer_norm(axes = input_597_axes_0, beta = encoder_layers_10_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_out_weight_to_fp16, x = input_595_cast_fp16)[name = tensor("input_597_cast_fp16")]; tensor cache_45_begin_0 = const()[name = tensor("cache_45_begin_0"), val = tensor([11, 0, 0, 0])]; tensor cache_45_end_0 = const()[name = tensor("cache_45_end_0"), val = tensor([12, 1, 70, 512])]; tensor cache_45_end_mask_0 = const()[name = tensor("cache_45_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_45_squeeze_mask_0 = const()[name = tensor("cache_45_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_45_cast_fp16 = slice_by_index(begin = cache_45_begin_0, end = cache_45_end_0, end_mask = cache_45_end_mask_0, squeeze_mask = cache_45_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_45_cast_fp16")]; tensor cache_47_begin_0 = const()[name = tensor("cache_47_begin_0"), val = tensor([11, 0, 0, 0])]; tensor cache_47_end_0 = const()[name = tensor("cache_47_end_0"), val = tensor([12, 1, 512, 8])]; tensor cache_47_end_mask_0 = const()[name = tensor("cache_47_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_47_squeeze_mask_0 = const()[name = tensor("cache_47_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_47_cast_fp16 = slice_by_index(begin = cache_47_begin_0, end = cache_47_end_0, end_mask = cache_47_end_mask_0, squeeze_mask = cache_47_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_47_cast_fp16")]; tensor input_599_axes_0 = const()[name = tensor("input_599_axes_0"), val = tensor([-1])]; tensor encoder_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139410496)))]; tensor encoder_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139411584)))]; tensor input_599_cast_fp16 = layer_norm(axes = input_599_axes_0, beta = encoder_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward1_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("input_599_cast_fp16")]; tensor encoder_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139412672)))]; tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_599_cast_fp16)[name = tensor("linear_100_cast_fp16")]; tensor input_603_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_603_cast_fp16")]; tensor encoder_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141509888)))]; tensor linear_101_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("linear_101_cast_fp16")]; tensor var_2713_to_fp16 = const()[name = tensor("op_2713_to_fp16"), val = tensor(0x1p-1)]; tensor var_2714_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2713_to_fp16)[name = tensor("op_2714_cast_fp16")]; tensor input_609_cast_fp16 = add(x = input_597_cast_fp16, y = var_2714_cast_fp16)[name = tensor("input_609_cast_fp16")]; tensor key_23_axes_0 = const()[name = tensor("key_23_axes_0"), val = tensor([-1])]; tensor encoder_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143607104)))]; tensor encoder_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143608192)))]; tensor key_23_cast_fp16 = layer_norm(axes = key_23_axes_0, beta = encoder_layers_11_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_self_att_weight_to_fp16, x = input_609_cast_fp16)[name = tensor("key_23_cast_fp16")]; tensor input_611_interleave_0 = const()[name = tensor("input_611_interleave_0"), val = tensor(false)]; tensor input_611_cast_fp16 = concat(axis = var_65, interleave = input_611_interleave_0, values = (cache_45_cast_fp16, key_23_cast_fp16))[name = tensor("input_611_cast_fp16")]; tensor var_2736_begin_0 = const()[name = tensor("op_2736_begin_0"), val = tensor([0, 4, 0])]; tensor var_2736_end_0 = const()[name = tensor("op_2736_end_0"), val = tensor([1, 70, 512])]; tensor var_2736_end_mask_0 = const()[name = tensor("op_2736_end_mask_0"), val = tensor([true, true, true])]; tensor var_2736_cast_fp16 = slice_by_index(begin = var_2736_begin_0, end = var_2736_end_0, end_mask = var_2736_end_mask_0, x = cache_45_cast_fp16)[name = tensor("op_2736_cast_fp16")]; tensor var_2739_begin_0 = const()[name = tensor("op_2739_begin_0"), val = tensor([0, 0, 0])]; tensor var_2739_end_0 = const()[name = tensor("op_2739_end_0"), val = tensor([1, 4, 512])]; tensor var_2739_end_mask_0 = const()[name = tensor("op_2739_end_mask_0"), val = tensor([true, false, true])]; tensor var_2739_cast_fp16 = slice_by_index(begin = var_2739_begin_0, end = var_2739_end_0, end_mask = var_2739_end_mask_0, x = key_23_cast_fp16)[name = tensor("op_2739_cast_fp16")]; tensor var_2742_interleave_0 = const()[name = tensor("op_2742_interleave_0"), val = tensor(false)]; tensor var_2742_cast_fp16 = concat(axis = var_65, interleave = var_2742_interleave_0, values = (var_2736_cast_fp16, var_2739_cast_fp16))[name = tensor("op_2742_cast_fp16")]; tensor encoder_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143609280)))]; tensor linear_102_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_q_weight_to_fp16, x = key_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; tensor var_2746 = const()[name = tensor("op_2746"), val = tensor([1, -1, 8, 64])]; tensor q_67_cast_fp16 = reshape(shape = var_2746, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; tensor encoder_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144133632)))]; tensor linear_103_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_k_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_103_cast_fp16")]; tensor var_2750 = const()[name = tensor("op_2750"), val = tensor([1, -1, 8, 64])]; tensor k_45_cast_fp16 = reshape(shape = var_2750, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; tensor encoder_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144657984)))]; tensor linear_104_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_v_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_104_cast_fp16")]; tensor var_2754 = const()[name = tensor("op_2754"), val = tensor([1, -1, 8, 64])]; tensor v_23_cast_fp16 = reshape(shape = var_2754, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145182336)))]; tensor var_2766_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2766_cast_fp16")]; tensor encoder_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145183424)))]; tensor var_2768_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2768_cast_fp16")]; tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; tensor var_2770_to_fp16 = const()[name = tensor("op_2770_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145184512)))]; tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2768_cast_fp16)[name = tensor("transpose_138")]; tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2770_to_fp16)[name = tensor("x_293_cast_fp16")]; tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; tensor const_222_to_fp16 = const()[name = tensor("const_222_to_fp16"), val = tensor(0x0p+0)]; tensor x_295_cast_fp16 = pad(constant_val = const_222_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; tensor var_2778 = const()[name = tensor("op_2778"), val = tensor([1, 8, -1, 8])]; tensor x_297_cast_fp16 = reshape(shape = var_2778, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; tensor var_2782_begin_0 = const()[name = tensor("op_2782_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2782_end_0 = const()[name = tensor("op_2782_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_2782_end_mask_0 = const()[name = tensor("op_2782_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2782_cast_fp16 = slice_by_index(begin = var_2782_begin_0, end = var_2782_end_0, end_mask = var_2782_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2782_cast_fp16")]; tensor var_2783 = const()[name = tensor("op_2783"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2783, x = var_2782_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_136")]; tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2766_cast_fp16)[name = tensor("transpose_137")]; tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; tensor var_2792_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2792_cast_fp16")]; tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_45_cast_fp16 = mul(x = var_2792_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; tensor scores_47_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_11)[name = tensor("scores_47_cast_fp16")]; tensor var_2798_cast_fp16 = softmax(axis = var_56, x = scores_47_cast_fp16)[name = tensor("op_2798_cast_fp16")]; tensor input_613_cast_fp16 = select(a = var_40_to_fp16, b = var_2798_cast_fp16, cond = mask_11)[name = tensor("input_613_cast_fp16")]; tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_139")]; tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_613_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_299_cast_fp16")]; tensor var_2802_perm_0 = const()[name = tensor("op_2802_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2803 = const()[name = tensor("op_2803"), val = tensor([1, -1, 512])]; tensor var_2802_cast_fp16 = transpose(perm = var_2802_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_135")]; tensor input_615_cast_fp16 = reshape(shape = var_2803, x = var_2802_cast_fp16)[name = tensor("input_615_cast_fp16")]; tensor encoder_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145343296)))]; tensor linear_106_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_out_weight_to_fp16, x = input_615_cast_fp16)[name = tensor("linear_106_cast_fp16")]; tensor input_619_cast_fp16 = add(x = input_609_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_619_cast_fp16")]; tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; tensor encoder_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145867648)))]; tensor encoder_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145868736)))]; tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = encoder_layers_11_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_conv_weight_to_fp16, x = input_619_cast_fp16)[name = tensor("x_303_cast_fp16")]; tensor input_621_perm_0 = const()[name = tensor("input_621_perm_0"), val = tensor([0, 2, 1])]; tensor input_623_pad_type_0 = const()[name = tensor("input_623_pad_type_0"), val = tensor("valid")]; tensor input_623_strides_0 = const()[name = tensor("input_623_strides_0"), val = tensor([1])]; tensor input_623_pad_0 = const()[name = tensor("input_623_pad_0"), val = tensor([0, 0])]; tensor input_623_dilations_0 = const()[name = tensor("input_623_dilations_0"), val = tensor([1])]; tensor input_623_groups_0 = const()[name = tensor("input_623_groups_0"), val = tensor(1)]; tensor encoder_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145869824)))]; tensor input_621_cast_fp16 = transpose(perm = input_621_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_134")]; tensor input_623_cast_fp16 = conv(dilations = input_623_dilations_0, groups = input_623_groups_0, pad = input_623_pad_0, pad_type = input_623_pad_type_0, strides = input_623_strides_0, weight = encoder_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_621_cast_fp16)[name = tensor("input_623_cast_fp16")]; tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_623_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; tensor input_625_cast_fp16 = select(a = var_40_to_fp16, b = x_305_cast_fp16, cond = var_551)[name = tensor("input_625_cast_fp16")]; tensor new_x_47_interleave_0 = const()[name = tensor("new_x_47_interleave_0"), val = tensor(false)]; tensor new_x_47_cast_fp16 = concat(axis = var_56, interleave = new_x_47_interleave_0, values = (cache_47_cast_fp16, input_625_cast_fp16))[name = tensor("new_x_47_cast_fp16")]; tensor next_cache_23_begin_0 = const()[name = tensor("next_cache_23_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_23_end_0 = const()[name = tensor("next_cache_23_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_23_end_mask_0 = const()[name = tensor("next_cache_23_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_23_cast_fp16 = slice_by_index(begin = next_cache_23_begin_0, end = next_cache_23_end_0, end_mask = next_cache_23_end_mask_0, x = new_x_47_cast_fp16)[name = tensor("next_cache_23_cast_fp16")]; tensor var_2844_begin_0 = const()[name = tensor("op_2844_begin_0"), val = tensor([0, 0, 4])]; tensor var_2844_end_0 = const()[name = tensor("op_2844_end_0"), val = tensor([1, 512, 12])]; tensor var_2844_end_mask_0 = const()[name = tensor("op_2844_end_mask_0"), val = tensor([true, true, true])]; tensor var_2844_cast_fp16 = slice_by_index(begin = var_2844_begin_0, end = var_2844_end_0, end_mask = var_2844_end_mask_0, x = next_cache_23_cast_fp16)[name = tensor("op_2844_cast_fp16")]; tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(512)]; tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; tensor encoder_layers_11_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146918464)))]; tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = encoder_layers_11_conv_depthwise_conv_weight_to_fp16, x = new_x_47_cast_fp16)[name = tensor("x_307_cast_fp16")]; tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; tensor x_309_axes_0 = const()[name = tensor("x_309_axes_0"), val = tensor([-1])]; tensor encoder_layers_11_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146927744)))]; tensor encoder_layers_11_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146928832)))]; tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_133")]; tensor x_309_cast_fp16 = layer_norm(axes = x_309_axes_0, beta = encoder_layers_11_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_conv_batch_norm_weight_to_fp16, x = input_627_cast_fp16)[name = tensor("x_309_cast_fp16")]; tensor input_629_perm_0 = const()[name = tensor("input_629_perm_0"), val = tensor([0, 2, 1])]; tensor input_629_cast_fp16 = transpose(perm = input_629_perm_0, x = x_309_cast_fp16)[name = tensor("transpose_132")]; tensor input_631_cast_fp16 = silu(x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; tensor x_311_pad_type_0 = const()[name = tensor("x_311_pad_type_0"), val = tensor("valid")]; tensor x_311_strides_0 = const()[name = tensor("x_311_strides_0"), val = tensor([1])]; tensor x_311_pad_0 = const()[name = tensor("x_311_pad_0"), val = tensor([0, 0])]; tensor x_311_dilations_0 = const()[name = tensor("x_311_dilations_0"), val = tensor([1])]; tensor x_311_groups_0 = const()[name = tensor("x_311_groups_0"), val = tensor(1)]; tensor encoder_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146929920)))]; tensor x_311_cast_fp16 = conv(dilations = x_311_dilations_0, groups = x_311_groups_0, pad = x_311_pad_0, pad_type = x_311_pad_type_0, strides = x_311_strides_0, weight = encoder_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("x_311_cast_fp16")]; tensor input_633_perm_0 = const()[name = tensor("input_633_perm_0"), val = tensor([0, 2, 1])]; tensor input_633_cast_fp16 = transpose(perm = input_633_perm_0, x = x_311_cast_fp16)[name = tensor("transpose_131")]; tensor input_635_cast_fp16 = add(x = input_619_cast_fp16, y = input_633_cast_fp16)[name = tensor("input_635_cast_fp16")]; tensor input_637_axes_0 = const()[name = tensor("input_637_axes_0"), val = tensor([-1])]; tensor encoder_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147454272)))]; tensor encoder_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147455360)))]; tensor input_637_cast_fp16 = layer_norm(axes = input_637_axes_0, beta = encoder_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("input_637_cast_fp16")]; tensor encoder_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147456448)))]; tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_637_cast_fp16)[name = tensor("linear_107_cast_fp16")]; tensor input_641_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_641_cast_fp16")]; tensor encoder_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149553664)))]; tensor linear_108_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("linear_108_cast_fp16")]; tensor var_2885_to_fp16 = const()[name = tensor("op_2885_to_fp16"), val = tensor(0x1p-1)]; tensor var_2886_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2885_to_fp16)[name = tensor("op_2886_cast_fp16")]; tensor input_647_cast_fp16 = add(x = input_635_cast_fp16, y = var_2886_cast_fp16)[name = tensor("input_647_cast_fp16")]; tensor input_649_axes_0 = const()[name = tensor("input_649_axes_0"), val = tensor([-1])]; tensor encoder_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151650880)))]; tensor encoder_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151651968)))]; tensor input_649_cast_fp16 = layer_norm(axes = input_649_axes_0, beta = encoder_layers_11_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_out_weight_to_fp16, x = input_647_cast_fp16)[name = tensor("input_649_cast_fp16")]; tensor cache_49_begin_0 = const()[name = tensor("cache_49_begin_0"), val = tensor([12, 0, 0, 0])]; tensor cache_49_end_0 = const()[name = tensor("cache_49_end_0"), val = tensor([13, 1, 70, 512])]; tensor cache_49_end_mask_0 = const()[name = tensor("cache_49_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_49_squeeze_mask_0 = const()[name = tensor("cache_49_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_49_cast_fp16 = slice_by_index(begin = cache_49_begin_0, end = cache_49_end_0, end_mask = cache_49_end_mask_0, squeeze_mask = cache_49_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_49_cast_fp16")]; tensor cache_51_begin_0 = const()[name = tensor("cache_51_begin_0"), val = tensor([12, 0, 0, 0])]; tensor cache_51_end_0 = const()[name = tensor("cache_51_end_0"), val = tensor([13, 1, 512, 8])]; tensor cache_51_end_mask_0 = const()[name = tensor("cache_51_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_51_squeeze_mask_0 = const()[name = tensor("cache_51_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_51_cast_fp16 = slice_by_index(begin = cache_51_begin_0, end = cache_51_end_0, end_mask = cache_51_end_mask_0, squeeze_mask = cache_51_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_51_cast_fp16")]; tensor input_651_axes_0 = const()[name = tensor("input_651_axes_0"), val = tensor([-1])]; tensor encoder_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151653056)))]; tensor encoder_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151654144)))]; tensor input_651_cast_fp16 = layer_norm(axes = input_651_axes_0, beta = encoder_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward1_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("input_651_cast_fp16")]; tensor encoder_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151655232)))]; tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_651_cast_fp16)[name = tensor("linear_109_cast_fp16")]; tensor input_655_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_655_cast_fp16")]; tensor encoder_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153752448)))]; tensor linear_110_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("linear_110_cast_fp16")]; tensor var_2920_to_fp16 = const()[name = tensor("op_2920_to_fp16"), val = tensor(0x1p-1)]; tensor var_2921_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2920_to_fp16)[name = tensor("op_2921_cast_fp16")]; tensor input_661_cast_fp16 = add(x = input_649_cast_fp16, y = var_2921_cast_fp16)[name = tensor("input_661_cast_fp16")]; tensor key_25_axes_0 = const()[name = tensor("key_25_axes_0"), val = tensor([-1])]; tensor encoder_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155849664)))]; tensor encoder_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155850752)))]; tensor key_25_cast_fp16 = layer_norm(axes = key_25_axes_0, beta = encoder_layers_12_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_self_att_weight_to_fp16, x = input_661_cast_fp16)[name = tensor("key_25_cast_fp16")]; tensor input_663_interleave_0 = const()[name = tensor("input_663_interleave_0"), val = tensor(false)]; tensor input_663_cast_fp16 = concat(axis = var_65, interleave = input_663_interleave_0, values = (cache_49_cast_fp16, key_25_cast_fp16))[name = tensor("input_663_cast_fp16")]; tensor var_2943_begin_0 = const()[name = tensor("op_2943_begin_0"), val = tensor([0, 4, 0])]; tensor var_2943_end_0 = const()[name = tensor("op_2943_end_0"), val = tensor([1, 70, 512])]; tensor var_2943_end_mask_0 = const()[name = tensor("op_2943_end_mask_0"), val = tensor([true, true, true])]; tensor var_2943_cast_fp16 = slice_by_index(begin = var_2943_begin_0, end = var_2943_end_0, end_mask = var_2943_end_mask_0, x = cache_49_cast_fp16)[name = tensor("op_2943_cast_fp16")]; tensor var_2946_begin_0 = const()[name = tensor("op_2946_begin_0"), val = tensor([0, 0, 0])]; tensor var_2946_end_0 = const()[name = tensor("op_2946_end_0"), val = tensor([1, 4, 512])]; tensor var_2946_end_mask_0 = const()[name = tensor("op_2946_end_mask_0"), val = tensor([true, false, true])]; tensor var_2946_cast_fp16 = slice_by_index(begin = var_2946_begin_0, end = var_2946_end_0, end_mask = var_2946_end_mask_0, x = key_25_cast_fp16)[name = tensor("op_2946_cast_fp16")]; tensor var_2949_interleave_0 = const()[name = tensor("op_2949_interleave_0"), val = tensor(false)]; tensor var_2949_cast_fp16 = concat(axis = var_65, interleave = var_2949_interleave_0, values = (var_2943_cast_fp16, var_2946_cast_fp16))[name = tensor("op_2949_cast_fp16")]; tensor encoder_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155851840)))]; tensor linear_111_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_q_weight_to_fp16, x = key_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; tensor var_2953 = const()[name = tensor("op_2953"), val = tensor([1, -1, 8, 64])]; tensor q_73_cast_fp16 = reshape(shape = var_2953, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; tensor encoder_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156376192)))]; tensor linear_112_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_k_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_112_cast_fp16")]; tensor var_2957 = const()[name = tensor("op_2957"), val = tensor([1, -1, 8, 64])]; tensor k_49_cast_fp16 = reshape(shape = var_2957, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; tensor encoder_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156900544)))]; tensor linear_113_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_v_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_113_cast_fp16")]; tensor var_2961 = const()[name = tensor("op_2961"), val = tensor([1, -1, 8, 64])]; tensor v_25_cast_fp16 = reshape(shape = var_2961, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157424896)))]; tensor var_2973_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2973_cast_fp16")]; tensor encoder_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157425984)))]; tensor var_2975_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2975_cast_fp16")]; tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_319_transpose_x_0 = const()[name = tensor("x_319_transpose_x_0"), val = tensor(false)]; tensor x_319_transpose_y_0 = const()[name = tensor("x_319_transpose_y_0"), val = tensor(false)]; tensor var_2977_to_fp16 = const()[name = tensor("op_2977_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157427072)))]; tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2975_cast_fp16)[name = tensor("transpose_129")]; tensor x_319_cast_fp16 = matmul(transpose_x = x_319_transpose_x_0, transpose_y = x_319_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2977_to_fp16)[name = tensor("x_319_cast_fp16")]; tensor x_321_pad_0 = const()[name = tensor("x_321_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_321_mode_0 = const()[name = tensor("x_321_mode_0"), val = tensor("constant")]; tensor const_235_to_fp16 = const()[name = tensor("const_235_to_fp16"), val = tensor(0x0p+0)]; tensor x_321_cast_fp16 = pad(constant_val = const_235_to_fp16, mode = x_321_mode_0, pad = x_321_pad_0, x = x_319_cast_fp16)[name = tensor("x_321_cast_fp16")]; tensor var_2985 = const()[name = tensor("op_2985"), val = tensor([1, 8, -1, 8])]; tensor x_323_cast_fp16 = reshape(shape = var_2985, x = x_321_cast_fp16)[name = tensor("x_323_cast_fp16")]; tensor var_2989_begin_0 = const()[name = tensor("op_2989_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2989_end_0 = const()[name = tensor("op_2989_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_2989_end_mask_0 = const()[name = tensor("op_2989_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2989_cast_fp16 = slice_by_index(begin = var_2989_begin_0, end = var_2989_end_0, end_mask = var_2989_end_mask_0, x = x_323_cast_fp16)[name = tensor("op_2989_cast_fp16")]; tensor var_2990 = const()[name = tensor("op_2990"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2990, x = var_2989_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_127")]; tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2973_cast_fp16)[name = tensor("transpose_128")]; tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; tensor var_2999_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2999_cast_fp16")]; tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_49_cast_fp16 = mul(x = var_2999_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; tensor scores_51_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_11)[name = tensor("scores_51_cast_fp16")]; tensor var_3005_cast_fp16 = softmax(axis = var_56, x = scores_51_cast_fp16)[name = tensor("op_3005_cast_fp16")]; tensor input_665_cast_fp16 = select(a = var_40_to_fp16, b = var_3005_cast_fp16, cond = mask_11)[name = tensor("input_665_cast_fp16")]; tensor x_325_transpose_x_0 = const()[name = tensor("x_325_transpose_x_0"), val = tensor(false)]; tensor x_325_transpose_y_0 = const()[name = tensor("x_325_transpose_y_0"), val = tensor(false)]; tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_130")]; tensor x_325_cast_fp16 = matmul(transpose_x = x_325_transpose_x_0, transpose_y = x_325_transpose_y_0, x = input_665_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_325_cast_fp16")]; tensor var_3009_perm_0 = const()[name = tensor("op_3009_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3010 = const()[name = tensor("op_3010"), val = tensor([1, -1, 512])]; tensor var_3009_cast_fp16 = transpose(perm = var_3009_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_126")]; tensor input_667_cast_fp16 = reshape(shape = var_3010, x = var_3009_cast_fp16)[name = tensor("input_667_cast_fp16")]; tensor encoder_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157585856)))]; tensor linear_115_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_out_weight_to_fp16, x = input_667_cast_fp16)[name = tensor("linear_115_cast_fp16")]; tensor input_671_cast_fp16 = add(x = input_661_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_671_cast_fp16")]; tensor x_329_axes_0 = const()[name = tensor("x_329_axes_0"), val = tensor([-1])]; tensor encoder_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158110208)))]; tensor encoder_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158111296)))]; tensor x_329_cast_fp16 = layer_norm(axes = x_329_axes_0, beta = encoder_layers_12_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_conv_weight_to_fp16, x = input_671_cast_fp16)[name = tensor("x_329_cast_fp16")]; tensor input_673_perm_0 = const()[name = tensor("input_673_perm_0"), val = tensor([0, 2, 1])]; tensor input_675_pad_type_0 = const()[name = tensor("input_675_pad_type_0"), val = tensor("valid")]; tensor input_675_strides_0 = const()[name = tensor("input_675_strides_0"), val = tensor([1])]; tensor input_675_pad_0 = const()[name = tensor("input_675_pad_0"), val = tensor([0, 0])]; tensor input_675_dilations_0 = const()[name = tensor("input_675_dilations_0"), val = tensor([1])]; tensor input_675_groups_0 = const()[name = tensor("input_675_groups_0"), val = tensor(1)]; tensor encoder_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158112384)))]; tensor input_673_cast_fp16 = transpose(perm = input_673_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_125")]; tensor input_675_cast_fp16 = conv(dilations = input_675_dilations_0, groups = input_675_groups_0, pad = input_675_pad_0, pad_type = input_675_pad_type_0, strides = input_675_strides_0, weight = encoder_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_673_cast_fp16)[name = tensor("input_675_cast_fp16")]; tensor x_331_split_num_splits_0 = const()[name = tensor("x_331_split_num_splits_0"), val = tensor(2)]; tensor x_331_split_axis_0 = const()[name = tensor("x_331_split_axis_0"), val = tensor(1)]; tensor x_331_split_cast_fp16_0, tensor x_331_split_cast_fp16_1 = split(axis = x_331_split_axis_0, num_splits = x_331_split_num_splits_0, x = input_675_cast_fp16)[name = tensor("x_331_split_cast_fp16")]; tensor x_331_split_1_sigmoid_cast_fp16 = sigmoid(x = x_331_split_cast_fp16_1)[name = tensor("x_331_split_1_sigmoid_cast_fp16")]; tensor x_331_cast_fp16 = mul(x = x_331_split_cast_fp16_0, y = x_331_split_1_sigmoid_cast_fp16)[name = tensor("x_331_cast_fp16")]; tensor input_677_cast_fp16 = select(a = var_40_to_fp16, b = x_331_cast_fp16, cond = var_551)[name = tensor("input_677_cast_fp16")]; tensor new_x_51_interleave_0 = const()[name = tensor("new_x_51_interleave_0"), val = tensor(false)]; tensor new_x_51_cast_fp16 = concat(axis = var_56, interleave = new_x_51_interleave_0, values = (cache_51_cast_fp16, input_677_cast_fp16))[name = tensor("new_x_51_cast_fp16")]; tensor next_cache_25_begin_0 = const()[name = tensor("next_cache_25_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_25_end_0 = const()[name = tensor("next_cache_25_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_25_end_mask_0 = const()[name = tensor("next_cache_25_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_25_cast_fp16 = slice_by_index(begin = next_cache_25_begin_0, end = next_cache_25_end_0, end_mask = next_cache_25_end_mask_0, x = new_x_51_cast_fp16)[name = tensor("next_cache_25_cast_fp16")]; tensor var_3051_begin_0 = const()[name = tensor("op_3051_begin_0"), val = tensor([0, 0, 4])]; tensor var_3051_end_0 = const()[name = tensor("op_3051_end_0"), val = tensor([1, 512, 12])]; tensor var_3051_end_mask_0 = const()[name = tensor("op_3051_end_mask_0"), val = tensor([true, true, true])]; tensor var_3051_cast_fp16 = slice_by_index(begin = var_3051_begin_0, end = var_3051_end_0, end_mask = var_3051_end_mask_0, x = next_cache_25_cast_fp16)[name = tensor("op_3051_cast_fp16")]; tensor x_333_pad_type_0 = const()[name = tensor("x_333_pad_type_0"), val = tensor("valid")]; tensor x_333_groups_0 = const()[name = tensor("x_333_groups_0"), val = tensor(512)]; tensor x_333_strides_0 = const()[name = tensor("x_333_strides_0"), val = tensor([1])]; tensor x_333_pad_0 = const()[name = tensor("x_333_pad_0"), val = tensor([0, 0])]; tensor x_333_dilations_0 = const()[name = tensor("x_333_dilations_0"), val = tensor([1])]; tensor encoder_layers_12_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159161024)))]; tensor x_333_cast_fp16 = conv(dilations = x_333_dilations_0, groups = x_333_groups_0, pad = x_333_pad_0, pad_type = x_333_pad_type_0, strides = x_333_strides_0, weight = encoder_layers_12_conv_depthwise_conv_weight_to_fp16, x = new_x_51_cast_fp16)[name = tensor("x_333_cast_fp16")]; tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; tensor x_335_axes_0 = const()[name = tensor("x_335_axes_0"), val = tensor([-1])]; tensor encoder_layers_12_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159170304)))]; tensor encoder_layers_12_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159171392)))]; tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_333_cast_fp16)[name = tensor("transpose_124")]; tensor x_335_cast_fp16 = layer_norm(axes = x_335_axes_0, beta = encoder_layers_12_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_conv_batch_norm_weight_to_fp16, x = input_679_cast_fp16)[name = tensor("x_335_cast_fp16")]; tensor input_681_perm_0 = const()[name = tensor("input_681_perm_0"), val = tensor([0, 2, 1])]; tensor input_681_cast_fp16 = transpose(perm = input_681_perm_0, x = x_335_cast_fp16)[name = tensor("transpose_123")]; tensor input_683_cast_fp16 = silu(x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; tensor x_337_pad_type_0 = const()[name = tensor("x_337_pad_type_0"), val = tensor("valid")]; tensor x_337_strides_0 = const()[name = tensor("x_337_strides_0"), val = tensor([1])]; tensor x_337_pad_0 = const()[name = tensor("x_337_pad_0"), val = tensor([0, 0])]; tensor x_337_dilations_0 = const()[name = tensor("x_337_dilations_0"), val = tensor([1])]; tensor x_337_groups_0 = const()[name = tensor("x_337_groups_0"), val = tensor(1)]; tensor encoder_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159172480)))]; tensor x_337_cast_fp16 = conv(dilations = x_337_dilations_0, groups = x_337_groups_0, pad = x_337_pad_0, pad_type = x_337_pad_type_0, strides = x_337_strides_0, weight = encoder_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("x_337_cast_fp16")]; tensor input_685_perm_0 = const()[name = tensor("input_685_perm_0"), val = tensor([0, 2, 1])]; tensor input_685_cast_fp16 = transpose(perm = input_685_perm_0, x = x_337_cast_fp16)[name = tensor("transpose_122")]; tensor input_687_cast_fp16 = add(x = input_671_cast_fp16, y = input_685_cast_fp16)[name = tensor("input_687_cast_fp16")]; tensor input_689_axes_0 = const()[name = tensor("input_689_axes_0"), val = tensor([-1])]; tensor encoder_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159696832)))]; tensor encoder_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159697920)))]; tensor input_689_cast_fp16 = layer_norm(axes = input_689_axes_0, beta = encoder_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("input_689_cast_fp16")]; tensor encoder_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159699008)))]; tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_689_cast_fp16)[name = tensor("linear_116_cast_fp16")]; tensor input_693_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_693_cast_fp16")]; tensor encoder_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161796224)))]; tensor linear_117_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("linear_117_cast_fp16")]; tensor var_3092_to_fp16 = const()[name = tensor("op_3092_to_fp16"), val = tensor(0x1p-1)]; tensor var_3093_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_3092_to_fp16)[name = tensor("op_3093_cast_fp16")]; tensor input_699_cast_fp16 = add(x = input_687_cast_fp16, y = var_3093_cast_fp16)[name = tensor("input_699_cast_fp16")]; tensor input_701_axes_0 = const()[name = tensor("input_701_axes_0"), val = tensor([-1])]; tensor encoder_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163893440)))]; tensor encoder_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163894528)))]; tensor input_701_cast_fp16 = layer_norm(axes = input_701_axes_0, beta = encoder_layers_12_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_out_weight_to_fp16, x = input_699_cast_fp16)[name = tensor("input_701_cast_fp16")]; tensor cache_53_begin_0 = const()[name = tensor("cache_53_begin_0"), val = tensor([13, 0, 0, 0])]; tensor cache_53_end_0 = const()[name = tensor("cache_53_end_0"), val = tensor([14, 1, 70, 512])]; tensor cache_53_end_mask_0 = const()[name = tensor("cache_53_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_53_squeeze_mask_0 = const()[name = tensor("cache_53_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_53_cast_fp16 = slice_by_index(begin = cache_53_begin_0, end = cache_53_end_0, end_mask = cache_53_end_mask_0, squeeze_mask = cache_53_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_53_cast_fp16")]; tensor cache_55_begin_0 = const()[name = tensor("cache_55_begin_0"), val = tensor([13, 0, 0, 0])]; tensor cache_55_end_0 = const()[name = tensor("cache_55_end_0"), val = tensor([14, 1, 512, 8])]; tensor cache_55_end_mask_0 = const()[name = tensor("cache_55_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_55_squeeze_mask_0 = const()[name = tensor("cache_55_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_55_cast_fp16 = slice_by_index(begin = cache_55_begin_0, end = cache_55_end_0, end_mask = cache_55_end_mask_0, squeeze_mask = cache_55_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_55_cast_fp16")]; tensor input_703_axes_0 = const()[name = tensor("input_703_axes_0"), val = tensor([-1])]; tensor encoder_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163895616)))]; tensor encoder_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163896704)))]; tensor input_703_cast_fp16 = layer_norm(axes = input_703_axes_0, beta = encoder_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward1_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("input_703_cast_fp16")]; tensor encoder_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163897792)))]; tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_703_cast_fp16)[name = tensor("linear_118_cast_fp16")]; tensor input_707_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_707_cast_fp16")]; tensor encoder_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165995008)))]; tensor linear_119_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("linear_119_cast_fp16")]; tensor var_3127_to_fp16 = const()[name = tensor("op_3127_to_fp16"), val = tensor(0x1p-1)]; tensor var_3128_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_3127_to_fp16)[name = tensor("op_3128_cast_fp16")]; tensor input_713_cast_fp16 = add(x = input_701_cast_fp16, y = var_3128_cast_fp16)[name = tensor("input_713_cast_fp16")]; tensor key_27_axes_0 = const()[name = tensor("key_27_axes_0"), val = tensor([-1])]; tensor encoder_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168092224)))]; tensor encoder_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168093312)))]; tensor key_27_cast_fp16 = layer_norm(axes = key_27_axes_0, beta = encoder_layers_13_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_self_att_weight_to_fp16, x = input_713_cast_fp16)[name = tensor("key_27_cast_fp16")]; tensor input_715_interleave_0 = const()[name = tensor("input_715_interleave_0"), val = tensor(false)]; tensor input_715_cast_fp16 = concat(axis = var_65, interleave = input_715_interleave_0, values = (cache_53_cast_fp16, key_27_cast_fp16))[name = tensor("input_715_cast_fp16")]; tensor var_3150_begin_0 = const()[name = tensor("op_3150_begin_0"), val = tensor([0, 4, 0])]; tensor var_3150_end_0 = const()[name = tensor("op_3150_end_0"), val = tensor([1, 70, 512])]; tensor var_3150_end_mask_0 = const()[name = tensor("op_3150_end_mask_0"), val = tensor([true, true, true])]; tensor var_3150_cast_fp16 = slice_by_index(begin = var_3150_begin_0, end = var_3150_end_0, end_mask = var_3150_end_mask_0, x = cache_53_cast_fp16)[name = tensor("op_3150_cast_fp16")]; tensor var_3153_begin_0 = const()[name = tensor("op_3153_begin_0"), val = tensor([0, 0, 0])]; tensor var_3153_end_0 = const()[name = tensor("op_3153_end_0"), val = tensor([1, 4, 512])]; tensor var_3153_end_mask_0 = const()[name = tensor("op_3153_end_mask_0"), val = tensor([true, false, true])]; tensor var_3153_cast_fp16 = slice_by_index(begin = var_3153_begin_0, end = var_3153_end_0, end_mask = var_3153_end_mask_0, x = key_27_cast_fp16)[name = tensor("op_3153_cast_fp16")]; tensor var_3156_interleave_0 = const()[name = tensor("op_3156_interleave_0"), val = tensor(false)]; tensor var_3156_cast_fp16 = concat(axis = var_65, interleave = var_3156_interleave_0, values = (var_3150_cast_fp16, var_3153_cast_fp16))[name = tensor("op_3156_cast_fp16")]; tensor encoder_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168094400)))]; tensor linear_120_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_q_weight_to_fp16, x = key_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; tensor var_3160 = const()[name = tensor("op_3160"), val = tensor([1, -1, 8, 64])]; tensor q_79_cast_fp16 = reshape(shape = var_3160, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; tensor encoder_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168618752)))]; tensor linear_121_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_k_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_121_cast_fp16")]; tensor var_3164 = const()[name = tensor("op_3164"), val = tensor([1, -1, 8, 64])]; tensor k_53_cast_fp16 = reshape(shape = var_3164, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; tensor encoder_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169143104)))]; tensor linear_122_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_v_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_122_cast_fp16")]; tensor var_3168 = const()[name = tensor("op_3168"), val = tensor([1, -1, 8, 64])]; tensor v_27_cast_fp16 = reshape(shape = var_3168, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169667456)))]; tensor var_3180_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3180_cast_fp16")]; tensor encoder_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169668544)))]; tensor var_3182_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3182_cast_fp16")]; tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; tensor var_3184_to_fp16 = const()[name = tensor("op_3184_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169669632)))]; tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_3182_cast_fp16)[name = tensor("transpose_120")]; tensor x_345_cast_fp16 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_3184_to_fp16)[name = tensor("x_345_cast_fp16")]; tensor x_347_pad_0 = const()[name = tensor("x_347_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_347_mode_0 = const()[name = tensor("x_347_mode_0"), val = tensor("constant")]; tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(0x0p+0)]; tensor x_347_cast_fp16 = pad(constant_val = const_248_to_fp16, mode = x_347_mode_0, pad = x_347_pad_0, x = x_345_cast_fp16)[name = tensor("x_347_cast_fp16")]; tensor var_3192 = const()[name = tensor("op_3192"), val = tensor([1, 8, -1, 8])]; tensor x_349_cast_fp16 = reshape(shape = var_3192, x = x_347_cast_fp16)[name = tensor("x_349_cast_fp16")]; tensor var_3196_begin_0 = const()[name = tensor("op_3196_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3196_end_0 = const()[name = tensor("op_3196_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_3196_end_mask_0 = const()[name = tensor("op_3196_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3196_cast_fp16 = slice_by_index(begin = var_3196_begin_0, end = var_3196_end_0, end_mask = var_3196_end_mask_0, x = x_349_cast_fp16)[name = tensor("op_3196_cast_fp16")]; tensor var_3197 = const()[name = tensor("op_3197"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_53_cast_fp16 = reshape(shape = var_3197, x = var_3196_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_118")]; tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_3180_cast_fp16)[name = tensor("transpose_119")]; tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; tensor var_3206_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_3206_cast_fp16")]; tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_53_cast_fp16 = mul(x = var_3206_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; tensor scores_55_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_11)[name = tensor("scores_55_cast_fp16")]; tensor var_3212_cast_fp16 = softmax(axis = var_56, x = scores_55_cast_fp16)[name = tensor("op_3212_cast_fp16")]; tensor input_717_cast_fp16 = select(a = var_40_to_fp16, b = var_3212_cast_fp16, cond = mask_11)[name = tensor("input_717_cast_fp16")]; tensor x_351_transpose_x_0 = const()[name = tensor("x_351_transpose_x_0"), val = tensor(false)]; tensor x_351_transpose_y_0 = const()[name = tensor("x_351_transpose_y_0"), val = tensor(false)]; tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_121")]; tensor x_351_cast_fp16 = matmul(transpose_x = x_351_transpose_x_0, transpose_y = x_351_transpose_y_0, x = input_717_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_351_cast_fp16")]; tensor var_3216_perm_0 = const()[name = tensor("op_3216_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3217 = const()[name = tensor("op_3217"), val = tensor([1, -1, 512])]; tensor var_3216_cast_fp16 = transpose(perm = var_3216_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_117")]; tensor input_719_cast_fp16 = reshape(shape = var_3217, x = var_3216_cast_fp16)[name = tensor("input_719_cast_fp16")]; tensor encoder_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169828416)))]; tensor linear_124_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_out_weight_to_fp16, x = input_719_cast_fp16)[name = tensor("linear_124_cast_fp16")]; tensor input_723_cast_fp16 = add(x = input_713_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_723_cast_fp16")]; tensor x_355_axes_0 = const()[name = tensor("x_355_axes_0"), val = tensor([-1])]; tensor encoder_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170352768)))]; tensor encoder_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170353856)))]; tensor x_355_cast_fp16 = layer_norm(axes = x_355_axes_0, beta = encoder_layers_13_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_conv_weight_to_fp16, x = input_723_cast_fp16)[name = tensor("x_355_cast_fp16")]; tensor input_725_perm_0 = const()[name = tensor("input_725_perm_0"), val = tensor([0, 2, 1])]; tensor input_727_pad_type_0 = const()[name = tensor("input_727_pad_type_0"), val = tensor("valid")]; tensor input_727_strides_0 = const()[name = tensor("input_727_strides_0"), val = tensor([1])]; tensor input_727_pad_0 = const()[name = tensor("input_727_pad_0"), val = tensor([0, 0])]; tensor input_727_dilations_0 = const()[name = tensor("input_727_dilations_0"), val = tensor([1])]; tensor input_727_groups_0 = const()[name = tensor("input_727_groups_0"), val = tensor(1)]; tensor encoder_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170354944)))]; tensor input_725_cast_fp16 = transpose(perm = input_725_perm_0, x = x_355_cast_fp16)[name = tensor("transpose_116")]; tensor input_727_cast_fp16 = conv(dilations = input_727_dilations_0, groups = input_727_groups_0, pad = input_727_pad_0, pad_type = input_727_pad_type_0, strides = input_727_strides_0, weight = encoder_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_725_cast_fp16)[name = tensor("input_727_cast_fp16")]; tensor x_357_split_num_splits_0 = const()[name = tensor("x_357_split_num_splits_0"), val = tensor(2)]; tensor x_357_split_axis_0 = const()[name = tensor("x_357_split_axis_0"), val = tensor(1)]; tensor x_357_split_cast_fp16_0, tensor x_357_split_cast_fp16_1 = split(axis = x_357_split_axis_0, num_splits = x_357_split_num_splits_0, x = input_727_cast_fp16)[name = tensor("x_357_split_cast_fp16")]; tensor x_357_split_1_sigmoid_cast_fp16 = sigmoid(x = x_357_split_cast_fp16_1)[name = tensor("x_357_split_1_sigmoid_cast_fp16")]; tensor x_357_cast_fp16 = mul(x = x_357_split_cast_fp16_0, y = x_357_split_1_sigmoid_cast_fp16)[name = tensor("x_357_cast_fp16")]; tensor input_729_cast_fp16 = select(a = var_40_to_fp16, b = x_357_cast_fp16, cond = var_551)[name = tensor("input_729_cast_fp16")]; tensor new_x_55_interleave_0 = const()[name = tensor("new_x_55_interleave_0"), val = tensor(false)]; tensor new_x_55_cast_fp16 = concat(axis = var_56, interleave = new_x_55_interleave_0, values = (cache_55_cast_fp16, input_729_cast_fp16))[name = tensor("new_x_55_cast_fp16")]; tensor next_cache_27_begin_0 = const()[name = tensor("next_cache_27_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_27_end_0 = const()[name = tensor("next_cache_27_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_27_end_mask_0 = const()[name = tensor("next_cache_27_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_27_cast_fp16 = slice_by_index(begin = next_cache_27_begin_0, end = next_cache_27_end_0, end_mask = next_cache_27_end_mask_0, x = new_x_55_cast_fp16)[name = tensor("next_cache_27_cast_fp16")]; tensor var_3258_begin_0 = const()[name = tensor("op_3258_begin_0"), val = tensor([0, 0, 4])]; tensor var_3258_end_0 = const()[name = tensor("op_3258_end_0"), val = tensor([1, 512, 12])]; tensor var_3258_end_mask_0 = const()[name = tensor("op_3258_end_mask_0"), val = tensor([true, true, true])]; tensor var_3258_cast_fp16 = slice_by_index(begin = var_3258_begin_0, end = var_3258_end_0, end_mask = var_3258_end_mask_0, x = next_cache_27_cast_fp16)[name = tensor("op_3258_cast_fp16")]; tensor x_359_pad_type_0 = const()[name = tensor("x_359_pad_type_0"), val = tensor("valid")]; tensor x_359_groups_0 = const()[name = tensor("x_359_groups_0"), val = tensor(512)]; tensor x_359_strides_0 = const()[name = tensor("x_359_strides_0"), val = tensor([1])]; tensor x_359_pad_0 = const()[name = tensor("x_359_pad_0"), val = tensor([0, 0])]; tensor x_359_dilations_0 = const()[name = tensor("x_359_dilations_0"), val = tensor([1])]; tensor encoder_layers_13_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171403584)))]; tensor x_359_cast_fp16 = conv(dilations = x_359_dilations_0, groups = x_359_groups_0, pad = x_359_pad_0, pad_type = x_359_pad_type_0, strides = x_359_strides_0, weight = encoder_layers_13_conv_depthwise_conv_weight_to_fp16, x = new_x_55_cast_fp16)[name = tensor("x_359_cast_fp16")]; tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; tensor x_361_axes_0 = const()[name = tensor("x_361_axes_0"), val = tensor([-1])]; tensor encoder_layers_13_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171412864)))]; tensor encoder_layers_13_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171413952)))]; tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_359_cast_fp16)[name = tensor("transpose_115")]; tensor x_361_cast_fp16 = layer_norm(axes = x_361_axes_0, beta = encoder_layers_13_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_conv_batch_norm_weight_to_fp16, x = input_731_cast_fp16)[name = tensor("x_361_cast_fp16")]; tensor input_733_perm_0 = const()[name = tensor("input_733_perm_0"), val = tensor([0, 2, 1])]; tensor input_733_cast_fp16 = transpose(perm = input_733_perm_0, x = x_361_cast_fp16)[name = tensor("transpose_114")]; tensor input_735_cast_fp16 = silu(x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; tensor x_363_pad_type_0 = const()[name = tensor("x_363_pad_type_0"), val = tensor("valid")]; tensor x_363_strides_0 = const()[name = tensor("x_363_strides_0"), val = tensor([1])]; tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0])]; tensor x_363_dilations_0 = const()[name = tensor("x_363_dilations_0"), val = tensor([1])]; tensor x_363_groups_0 = const()[name = tensor("x_363_groups_0"), val = tensor(1)]; tensor encoder_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171415040)))]; tensor x_363_cast_fp16 = conv(dilations = x_363_dilations_0, groups = x_363_groups_0, pad = x_363_pad_0, pad_type = x_363_pad_type_0, strides = x_363_strides_0, weight = encoder_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("x_363_cast_fp16")]; tensor input_737_perm_0 = const()[name = tensor("input_737_perm_0"), val = tensor([0, 2, 1])]; tensor input_737_cast_fp16 = transpose(perm = input_737_perm_0, x = x_363_cast_fp16)[name = tensor("transpose_113")]; tensor input_739_cast_fp16 = add(x = input_723_cast_fp16, y = input_737_cast_fp16)[name = tensor("input_739_cast_fp16")]; tensor input_741_axes_0 = const()[name = tensor("input_741_axes_0"), val = tensor([-1])]; tensor encoder_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171939392)))]; tensor encoder_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171940480)))]; tensor input_741_cast_fp16 = layer_norm(axes = input_741_axes_0, beta = encoder_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("input_741_cast_fp16")]; tensor encoder_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171941568)))]; tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_741_cast_fp16)[name = tensor("linear_125_cast_fp16")]; tensor input_745_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_745_cast_fp16")]; tensor encoder_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174038784)))]; tensor linear_126_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("linear_126_cast_fp16")]; tensor var_3299_to_fp16 = const()[name = tensor("op_3299_to_fp16"), val = tensor(0x1p-1)]; tensor var_3300_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_3299_to_fp16)[name = tensor("op_3300_cast_fp16")]; tensor input_751_cast_fp16 = add(x = input_739_cast_fp16, y = var_3300_cast_fp16)[name = tensor("input_751_cast_fp16")]; tensor input_753_axes_0 = const()[name = tensor("input_753_axes_0"), val = tensor([-1])]; tensor encoder_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176136000)))]; tensor encoder_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176137088)))]; tensor input_753_cast_fp16 = layer_norm(axes = input_753_axes_0, beta = encoder_layers_13_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_out_weight_to_fp16, x = input_751_cast_fp16)[name = tensor("input_753_cast_fp16")]; tensor cache_57_begin_0 = const()[name = tensor("cache_57_begin_0"), val = tensor([14, 0, 0, 0])]; tensor cache_57_end_0 = const()[name = tensor("cache_57_end_0"), val = tensor([15, 1, 70, 512])]; tensor cache_57_end_mask_0 = const()[name = tensor("cache_57_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_57_squeeze_mask_0 = const()[name = tensor("cache_57_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_57_cast_fp16 = slice_by_index(begin = cache_57_begin_0, end = cache_57_end_0, end_mask = cache_57_end_mask_0, squeeze_mask = cache_57_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_57_cast_fp16")]; tensor cache_59_begin_0 = const()[name = tensor("cache_59_begin_0"), val = tensor([14, 0, 0, 0])]; tensor cache_59_end_0 = const()[name = tensor("cache_59_end_0"), val = tensor([15, 1, 512, 8])]; tensor cache_59_end_mask_0 = const()[name = tensor("cache_59_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_59_squeeze_mask_0 = const()[name = tensor("cache_59_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_59_cast_fp16 = slice_by_index(begin = cache_59_begin_0, end = cache_59_end_0, end_mask = cache_59_end_mask_0, squeeze_mask = cache_59_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_59_cast_fp16")]; tensor input_755_axes_0 = const()[name = tensor("input_755_axes_0"), val = tensor([-1])]; tensor encoder_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176138176)))]; tensor encoder_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176139264)))]; tensor input_755_cast_fp16 = layer_norm(axes = input_755_axes_0, beta = encoder_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward1_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("input_755_cast_fp16")]; tensor encoder_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176140352)))]; tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_755_cast_fp16)[name = tensor("linear_127_cast_fp16")]; tensor input_759_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_759_cast_fp16")]; tensor encoder_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178237568)))]; tensor linear_128_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("linear_128_cast_fp16")]; tensor var_3334_to_fp16 = const()[name = tensor("op_3334_to_fp16"), val = tensor(0x1p-1)]; tensor var_3335_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_3334_to_fp16)[name = tensor("op_3335_cast_fp16")]; tensor input_765_cast_fp16 = add(x = input_753_cast_fp16, y = var_3335_cast_fp16)[name = tensor("input_765_cast_fp16")]; tensor key_29_axes_0 = const()[name = tensor("key_29_axes_0"), val = tensor([-1])]; tensor encoder_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180334784)))]; tensor encoder_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180335872)))]; tensor key_29_cast_fp16 = layer_norm(axes = key_29_axes_0, beta = encoder_layers_14_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_self_att_weight_to_fp16, x = input_765_cast_fp16)[name = tensor("key_29_cast_fp16")]; tensor input_767_interleave_0 = const()[name = tensor("input_767_interleave_0"), val = tensor(false)]; tensor input_767_cast_fp16 = concat(axis = var_65, interleave = input_767_interleave_0, values = (cache_57_cast_fp16, key_29_cast_fp16))[name = tensor("input_767_cast_fp16")]; tensor var_3357_begin_0 = const()[name = tensor("op_3357_begin_0"), val = tensor([0, 4, 0])]; tensor var_3357_end_0 = const()[name = tensor("op_3357_end_0"), val = tensor([1, 70, 512])]; tensor var_3357_end_mask_0 = const()[name = tensor("op_3357_end_mask_0"), val = tensor([true, true, true])]; tensor var_3357_cast_fp16 = slice_by_index(begin = var_3357_begin_0, end = var_3357_end_0, end_mask = var_3357_end_mask_0, x = cache_57_cast_fp16)[name = tensor("op_3357_cast_fp16")]; tensor var_3360_begin_0 = const()[name = tensor("op_3360_begin_0"), val = tensor([0, 0, 0])]; tensor var_3360_end_0 = const()[name = tensor("op_3360_end_0"), val = tensor([1, 4, 512])]; tensor var_3360_end_mask_0 = const()[name = tensor("op_3360_end_mask_0"), val = tensor([true, false, true])]; tensor var_3360_cast_fp16 = slice_by_index(begin = var_3360_begin_0, end = var_3360_end_0, end_mask = var_3360_end_mask_0, x = key_29_cast_fp16)[name = tensor("op_3360_cast_fp16")]; tensor var_3363_interleave_0 = const()[name = tensor("op_3363_interleave_0"), val = tensor(false)]; tensor var_3363_cast_fp16 = concat(axis = var_65, interleave = var_3363_interleave_0, values = (var_3357_cast_fp16, var_3360_cast_fp16))[name = tensor("op_3363_cast_fp16")]; tensor encoder_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180336960)))]; tensor linear_129_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_q_weight_to_fp16, x = key_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; tensor var_3367 = const()[name = tensor("op_3367"), val = tensor([1, -1, 8, 64])]; tensor q_85_cast_fp16 = reshape(shape = var_3367, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; tensor encoder_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180861312)))]; tensor linear_130_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_k_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_130_cast_fp16")]; tensor var_3371 = const()[name = tensor("op_3371"), val = tensor([1, -1, 8, 64])]; tensor k_57_cast_fp16 = reshape(shape = var_3371, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; tensor encoder_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181385664)))]; tensor linear_131_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_v_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_131_cast_fp16")]; tensor var_3375 = const()[name = tensor("op_3375"), val = tensor([1, -1, 8, 64])]; tensor v_29_cast_fp16 = reshape(shape = var_3375, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181910016)))]; tensor var_3387_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3387_cast_fp16")]; tensor encoder_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181911104)))]; tensor var_3389_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3389_cast_fp16")]; tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_371_transpose_x_0 = const()[name = tensor("x_371_transpose_x_0"), val = tensor(false)]; tensor x_371_transpose_y_0 = const()[name = tensor("x_371_transpose_y_0"), val = tensor(false)]; tensor var_3391_to_fp16 = const()[name = tensor("op_3391_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181912192)))]; tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_3389_cast_fp16)[name = tensor("transpose_111")]; tensor x_371_cast_fp16 = matmul(transpose_x = x_371_transpose_x_0, transpose_y = x_371_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_3391_to_fp16)[name = tensor("x_371_cast_fp16")]; tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_373_mode_0 = const()[name = tensor("x_373_mode_0"), val = tensor("constant")]; tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(0x0p+0)]; tensor x_373_cast_fp16 = pad(constant_val = const_261_to_fp16, mode = x_373_mode_0, pad = x_373_pad_0, x = x_371_cast_fp16)[name = tensor("x_373_cast_fp16")]; tensor var_3399 = const()[name = tensor("op_3399"), val = tensor([1, 8, -1, 8])]; tensor x_375_cast_fp16 = reshape(shape = var_3399, x = x_373_cast_fp16)[name = tensor("x_375_cast_fp16")]; tensor var_3403_begin_0 = const()[name = tensor("op_3403_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3403_end_0 = const()[name = tensor("op_3403_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_3403_end_mask_0 = const()[name = tensor("op_3403_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3403_cast_fp16 = slice_by_index(begin = var_3403_begin_0, end = var_3403_end_0, end_mask = var_3403_end_mask_0, x = x_375_cast_fp16)[name = tensor("op_3403_cast_fp16")]; tensor var_3404 = const()[name = tensor("op_3404"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_57_cast_fp16 = reshape(shape = var_3404, x = var_3403_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_109")]; tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_3387_cast_fp16)[name = tensor("transpose_110")]; tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; tensor var_3413_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_3413_cast_fp16")]; tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_57_cast_fp16 = mul(x = var_3413_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; tensor scores_59_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_11)[name = tensor("scores_59_cast_fp16")]; tensor var_3419_cast_fp16 = softmax(axis = var_56, x = scores_59_cast_fp16)[name = tensor("op_3419_cast_fp16")]; tensor input_769_cast_fp16 = select(a = var_40_to_fp16, b = var_3419_cast_fp16, cond = mask_11)[name = tensor("input_769_cast_fp16")]; tensor x_377_transpose_x_0 = const()[name = tensor("x_377_transpose_x_0"), val = tensor(false)]; tensor x_377_transpose_y_0 = const()[name = tensor("x_377_transpose_y_0"), val = tensor(false)]; tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_112")]; tensor x_377_cast_fp16 = matmul(transpose_x = x_377_transpose_x_0, transpose_y = x_377_transpose_y_0, x = input_769_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_377_cast_fp16")]; tensor var_3423_perm_0 = const()[name = tensor("op_3423_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3424 = const()[name = tensor("op_3424"), val = tensor([1, -1, 512])]; tensor var_3423_cast_fp16 = transpose(perm = var_3423_perm_0, x = x_377_cast_fp16)[name = tensor("transpose_108")]; tensor input_771_cast_fp16 = reshape(shape = var_3424, x = var_3423_cast_fp16)[name = tensor("input_771_cast_fp16")]; tensor encoder_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182070976)))]; tensor linear_133_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_out_weight_to_fp16, x = input_771_cast_fp16)[name = tensor("linear_133_cast_fp16")]; tensor input_775_cast_fp16 = add(x = input_765_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_775_cast_fp16")]; tensor x_381_axes_0 = const()[name = tensor("x_381_axes_0"), val = tensor([-1])]; tensor encoder_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182595328)))]; tensor encoder_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182596416)))]; tensor x_381_cast_fp16 = layer_norm(axes = x_381_axes_0, beta = encoder_layers_14_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_conv_weight_to_fp16, x = input_775_cast_fp16)[name = tensor("x_381_cast_fp16")]; tensor input_777_perm_0 = const()[name = tensor("input_777_perm_0"), val = tensor([0, 2, 1])]; tensor input_779_pad_type_0 = const()[name = tensor("input_779_pad_type_0"), val = tensor("valid")]; tensor input_779_strides_0 = const()[name = tensor("input_779_strides_0"), val = tensor([1])]; tensor input_779_pad_0 = const()[name = tensor("input_779_pad_0"), val = tensor([0, 0])]; tensor input_779_dilations_0 = const()[name = tensor("input_779_dilations_0"), val = tensor([1])]; tensor input_779_groups_0 = const()[name = tensor("input_779_groups_0"), val = tensor(1)]; tensor encoder_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182597504)))]; tensor input_777_cast_fp16 = transpose(perm = input_777_perm_0, x = x_381_cast_fp16)[name = tensor("transpose_107")]; tensor input_779_cast_fp16 = conv(dilations = input_779_dilations_0, groups = input_779_groups_0, pad = input_779_pad_0, pad_type = input_779_pad_type_0, strides = input_779_strides_0, weight = encoder_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_777_cast_fp16)[name = tensor("input_779_cast_fp16")]; tensor x_383_split_num_splits_0 = const()[name = tensor("x_383_split_num_splits_0"), val = tensor(2)]; tensor x_383_split_axis_0 = const()[name = tensor("x_383_split_axis_0"), val = tensor(1)]; tensor x_383_split_cast_fp16_0, tensor x_383_split_cast_fp16_1 = split(axis = x_383_split_axis_0, num_splits = x_383_split_num_splits_0, x = input_779_cast_fp16)[name = tensor("x_383_split_cast_fp16")]; tensor x_383_split_1_sigmoid_cast_fp16 = sigmoid(x = x_383_split_cast_fp16_1)[name = tensor("x_383_split_1_sigmoid_cast_fp16")]; tensor x_383_cast_fp16 = mul(x = x_383_split_cast_fp16_0, y = x_383_split_1_sigmoid_cast_fp16)[name = tensor("x_383_cast_fp16")]; tensor input_781_cast_fp16 = select(a = var_40_to_fp16, b = x_383_cast_fp16, cond = var_551)[name = tensor("input_781_cast_fp16")]; tensor new_x_59_interleave_0 = const()[name = tensor("new_x_59_interleave_0"), val = tensor(false)]; tensor new_x_59_cast_fp16 = concat(axis = var_56, interleave = new_x_59_interleave_0, values = (cache_59_cast_fp16, input_781_cast_fp16))[name = tensor("new_x_59_cast_fp16")]; tensor next_cache_29_begin_0 = const()[name = tensor("next_cache_29_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_29_end_0 = const()[name = tensor("next_cache_29_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_29_end_mask_0 = const()[name = tensor("next_cache_29_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_29_cast_fp16 = slice_by_index(begin = next_cache_29_begin_0, end = next_cache_29_end_0, end_mask = next_cache_29_end_mask_0, x = new_x_59_cast_fp16)[name = tensor("next_cache_29_cast_fp16")]; tensor var_3465_begin_0 = const()[name = tensor("op_3465_begin_0"), val = tensor([0, 0, 4])]; tensor var_3465_end_0 = const()[name = tensor("op_3465_end_0"), val = tensor([1, 512, 12])]; tensor var_3465_end_mask_0 = const()[name = tensor("op_3465_end_mask_0"), val = tensor([true, true, true])]; tensor var_3465_cast_fp16 = slice_by_index(begin = var_3465_begin_0, end = var_3465_end_0, end_mask = var_3465_end_mask_0, x = next_cache_29_cast_fp16)[name = tensor("op_3465_cast_fp16")]; tensor x_385_pad_type_0 = const()[name = tensor("x_385_pad_type_0"), val = tensor("valid")]; tensor x_385_groups_0 = const()[name = tensor("x_385_groups_0"), val = tensor(512)]; tensor x_385_strides_0 = const()[name = tensor("x_385_strides_0"), val = tensor([1])]; tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0])]; tensor x_385_dilations_0 = const()[name = tensor("x_385_dilations_0"), val = tensor([1])]; tensor encoder_layers_14_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183646144)))]; tensor x_385_cast_fp16 = conv(dilations = x_385_dilations_0, groups = x_385_groups_0, pad = x_385_pad_0, pad_type = x_385_pad_type_0, strides = x_385_strides_0, weight = encoder_layers_14_conv_depthwise_conv_weight_to_fp16, x = new_x_59_cast_fp16)[name = tensor("x_385_cast_fp16")]; tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; tensor x_387_axes_0 = const()[name = tensor("x_387_axes_0"), val = tensor([-1])]; tensor encoder_layers_14_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183655424)))]; tensor encoder_layers_14_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183656512)))]; tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_385_cast_fp16)[name = tensor("transpose_106")]; tensor x_387_cast_fp16 = layer_norm(axes = x_387_axes_0, beta = encoder_layers_14_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_conv_batch_norm_weight_to_fp16, x = input_783_cast_fp16)[name = tensor("x_387_cast_fp16")]; tensor input_785_perm_0 = const()[name = tensor("input_785_perm_0"), val = tensor([0, 2, 1])]; tensor input_785_cast_fp16 = transpose(perm = input_785_perm_0, x = x_387_cast_fp16)[name = tensor("transpose_105")]; tensor input_787_cast_fp16 = silu(x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; tensor x_389_pad_type_0 = const()[name = tensor("x_389_pad_type_0"), val = tensor("valid")]; tensor x_389_strides_0 = const()[name = tensor("x_389_strides_0"), val = tensor([1])]; tensor x_389_pad_0 = const()[name = tensor("x_389_pad_0"), val = tensor([0, 0])]; tensor x_389_dilations_0 = const()[name = tensor("x_389_dilations_0"), val = tensor([1])]; tensor x_389_groups_0 = const()[name = tensor("x_389_groups_0"), val = tensor(1)]; tensor encoder_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183657600)))]; tensor x_389_cast_fp16 = conv(dilations = x_389_dilations_0, groups = x_389_groups_0, pad = x_389_pad_0, pad_type = x_389_pad_type_0, strides = x_389_strides_0, weight = encoder_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("x_389_cast_fp16")]; tensor input_789_perm_0 = const()[name = tensor("input_789_perm_0"), val = tensor([0, 2, 1])]; tensor input_789_cast_fp16 = transpose(perm = input_789_perm_0, x = x_389_cast_fp16)[name = tensor("transpose_104")]; tensor input_791_cast_fp16 = add(x = input_775_cast_fp16, y = input_789_cast_fp16)[name = tensor("input_791_cast_fp16")]; tensor input_793_axes_0 = const()[name = tensor("input_793_axes_0"), val = tensor([-1])]; tensor encoder_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184181952)))]; tensor encoder_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184183040)))]; tensor input_793_cast_fp16 = layer_norm(axes = input_793_axes_0, beta = encoder_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("input_793_cast_fp16")]; tensor encoder_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184184128)))]; tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_793_cast_fp16)[name = tensor("linear_134_cast_fp16")]; tensor input_797_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_797_cast_fp16")]; tensor encoder_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186281344)))]; tensor linear_135_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("linear_135_cast_fp16")]; tensor var_3506_to_fp16 = const()[name = tensor("op_3506_to_fp16"), val = tensor(0x1p-1)]; tensor var_3507_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_3506_to_fp16)[name = tensor("op_3507_cast_fp16")]; tensor input_803_cast_fp16 = add(x = input_791_cast_fp16, y = var_3507_cast_fp16)[name = tensor("input_803_cast_fp16")]; tensor input_805_axes_0 = const()[name = tensor("input_805_axes_0"), val = tensor([-1])]; tensor encoder_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188378560)))]; tensor encoder_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188379648)))]; tensor input_805_cast_fp16 = layer_norm(axes = input_805_axes_0, beta = encoder_layers_14_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_out_weight_to_fp16, x = input_803_cast_fp16)[name = tensor("input_805_cast_fp16")]; tensor cache_61_begin_0 = const()[name = tensor("cache_61_begin_0"), val = tensor([15, 0, 0, 0])]; tensor cache_61_end_0 = const()[name = tensor("cache_61_end_0"), val = tensor([16, 1, 70, 512])]; tensor cache_61_end_mask_0 = const()[name = tensor("cache_61_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_61_squeeze_mask_0 = const()[name = tensor("cache_61_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_61_cast_fp16 = slice_by_index(begin = cache_61_begin_0, end = cache_61_end_0, end_mask = cache_61_end_mask_0, squeeze_mask = cache_61_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_61_cast_fp16")]; tensor cache_63_begin_0 = const()[name = tensor("cache_63_begin_0"), val = tensor([15, 0, 0, 0])]; tensor cache_63_end_0 = const()[name = tensor("cache_63_end_0"), val = tensor([16, 1, 512, 8])]; tensor cache_63_end_mask_0 = const()[name = tensor("cache_63_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_63_squeeze_mask_0 = const()[name = tensor("cache_63_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_63_cast_fp16 = slice_by_index(begin = cache_63_begin_0, end = cache_63_end_0, end_mask = cache_63_end_mask_0, squeeze_mask = cache_63_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_63_cast_fp16")]; tensor input_807_axes_0 = const()[name = tensor("input_807_axes_0"), val = tensor([-1])]; tensor encoder_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188380736)))]; tensor encoder_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188381824)))]; tensor input_807_cast_fp16 = layer_norm(axes = input_807_axes_0, beta = encoder_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward1_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("input_807_cast_fp16")]; tensor encoder_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188382912)))]; tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_807_cast_fp16)[name = tensor("linear_136_cast_fp16")]; tensor input_811_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_811_cast_fp16")]; tensor encoder_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190480128)))]; tensor linear_137_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("linear_137_cast_fp16")]; tensor var_3541_to_fp16 = const()[name = tensor("op_3541_to_fp16"), val = tensor(0x1p-1)]; tensor var_3542_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_3541_to_fp16)[name = tensor("op_3542_cast_fp16")]; tensor input_817_cast_fp16 = add(x = input_805_cast_fp16, y = var_3542_cast_fp16)[name = tensor("input_817_cast_fp16")]; tensor key_31_axes_0 = const()[name = tensor("key_31_axes_0"), val = tensor([-1])]; tensor encoder_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192577344)))]; tensor encoder_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192578432)))]; tensor key_31_cast_fp16 = layer_norm(axes = key_31_axes_0, beta = encoder_layers_15_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_self_att_weight_to_fp16, x = input_817_cast_fp16)[name = tensor("key_31_cast_fp16")]; tensor input_819_interleave_0 = const()[name = tensor("input_819_interleave_0"), val = tensor(false)]; tensor input_819_cast_fp16 = concat(axis = var_65, interleave = input_819_interleave_0, values = (cache_61_cast_fp16, key_31_cast_fp16))[name = tensor("input_819_cast_fp16")]; tensor var_3564_begin_0 = const()[name = tensor("op_3564_begin_0"), val = tensor([0, 4, 0])]; tensor var_3564_end_0 = const()[name = tensor("op_3564_end_0"), val = tensor([1, 70, 512])]; tensor var_3564_end_mask_0 = const()[name = tensor("op_3564_end_mask_0"), val = tensor([true, true, true])]; tensor var_3564_cast_fp16 = slice_by_index(begin = var_3564_begin_0, end = var_3564_end_0, end_mask = var_3564_end_mask_0, x = cache_61_cast_fp16)[name = tensor("op_3564_cast_fp16")]; tensor var_3567_begin_0 = const()[name = tensor("op_3567_begin_0"), val = tensor([0, 0, 0])]; tensor var_3567_end_0 = const()[name = tensor("op_3567_end_0"), val = tensor([1, 4, 512])]; tensor var_3567_end_mask_0 = const()[name = tensor("op_3567_end_mask_0"), val = tensor([true, false, true])]; tensor var_3567_cast_fp16 = slice_by_index(begin = var_3567_begin_0, end = var_3567_end_0, end_mask = var_3567_end_mask_0, x = key_31_cast_fp16)[name = tensor("op_3567_cast_fp16")]; tensor var_3570_interleave_0 = const()[name = tensor("op_3570_interleave_0"), val = tensor(false)]; tensor var_3570_cast_fp16 = concat(axis = var_65, interleave = var_3570_interleave_0, values = (var_3564_cast_fp16, var_3567_cast_fp16))[name = tensor("op_3570_cast_fp16")]; tensor encoder_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192579520)))]; tensor linear_138_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_q_weight_to_fp16, x = key_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; tensor var_3574 = const()[name = tensor("op_3574"), val = tensor([1, -1, 8, 64])]; tensor q_91_cast_fp16 = reshape(shape = var_3574, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; tensor encoder_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193103872)))]; tensor linear_139_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_k_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_139_cast_fp16")]; tensor var_3578 = const()[name = tensor("op_3578"), val = tensor([1, -1, 8, 64])]; tensor k_61_cast_fp16 = reshape(shape = var_3578, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; tensor encoder_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193628224)))]; tensor linear_140_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_v_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_140_cast_fp16")]; tensor var_3582 = const()[name = tensor("op_3582"), val = tensor([1, -1, 8, 64])]; tensor v_31_cast_fp16 = reshape(shape = var_3582, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194152576)))]; tensor var_3594_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3594_cast_fp16")]; tensor encoder_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194153664)))]; tensor var_3596_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3596_cast_fp16")]; tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_397_transpose_x_0 = const()[name = tensor("x_397_transpose_x_0"), val = tensor(false)]; tensor x_397_transpose_y_0 = const()[name = tensor("x_397_transpose_y_0"), val = tensor(false)]; tensor var_3598_to_fp16 = const()[name = tensor("op_3598_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194154752)))]; tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3596_cast_fp16)[name = tensor("transpose_102")]; tensor x_397_cast_fp16 = matmul(transpose_x = x_397_transpose_x_0, transpose_y = x_397_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3598_to_fp16)[name = tensor("x_397_cast_fp16")]; tensor x_399_pad_0 = const()[name = tensor("x_399_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_399_mode_0 = const()[name = tensor("x_399_mode_0"), val = tensor("constant")]; tensor const_274_to_fp16 = const()[name = tensor("const_274_to_fp16"), val = tensor(0x0p+0)]; tensor x_399_cast_fp16 = pad(constant_val = const_274_to_fp16, mode = x_399_mode_0, pad = x_399_pad_0, x = x_397_cast_fp16)[name = tensor("x_399_cast_fp16")]; tensor var_3606 = const()[name = tensor("op_3606"), val = tensor([1, 8, -1, 8])]; tensor x_401_cast_fp16 = reshape(shape = var_3606, x = x_399_cast_fp16)[name = tensor("x_401_cast_fp16")]; tensor var_3610_begin_0 = const()[name = tensor("op_3610_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3610_end_0 = const()[name = tensor("op_3610_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_3610_end_mask_0 = const()[name = tensor("op_3610_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3610_cast_fp16 = slice_by_index(begin = var_3610_begin_0, end = var_3610_end_0, end_mask = var_3610_end_mask_0, x = x_401_cast_fp16)[name = tensor("op_3610_cast_fp16")]; tensor var_3611 = const()[name = tensor("op_3611"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3611, x = var_3610_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_100")]; tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3594_cast_fp16)[name = tensor("transpose_101")]; tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; tensor var_3620_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3620_cast_fp16")]; tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_61_cast_fp16 = mul(x = var_3620_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; tensor scores_63_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_11)[name = tensor("scores_63_cast_fp16")]; tensor var_3626_cast_fp16 = softmax(axis = var_56, x = scores_63_cast_fp16)[name = tensor("op_3626_cast_fp16")]; tensor input_821_cast_fp16 = select(a = var_40_to_fp16, b = var_3626_cast_fp16, cond = mask_11)[name = tensor("input_821_cast_fp16")]; tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_103")]; tensor x_403_cast_fp16 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = input_821_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_403_cast_fp16")]; tensor var_3630_perm_0 = const()[name = tensor("op_3630_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3631 = const()[name = tensor("op_3631"), val = tensor([1, -1, 512])]; tensor var_3630_cast_fp16 = transpose(perm = var_3630_perm_0, x = x_403_cast_fp16)[name = tensor("transpose_99")]; tensor input_823_cast_fp16 = reshape(shape = var_3631, x = var_3630_cast_fp16)[name = tensor("input_823_cast_fp16")]; tensor encoder_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194313536)))]; tensor linear_142_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_out_weight_to_fp16, x = input_823_cast_fp16)[name = tensor("linear_142_cast_fp16")]; tensor input_827_cast_fp16 = add(x = input_817_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_827_cast_fp16")]; tensor x_407_axes_0 = const()[name = tensor("x_407_axes_0"), val = tensor([-1])]; tensor encoder_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194837888)))]; tensor encoder_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194838976)))]; tensor x_407_cast_fp16 = layer_norm(axes = x_407_axes_0, beta = encoder_layers_15_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_conv_weight_to_fp16, x = input_827_cast_fp16)[name = tensor("x_407_cast_fp16")]; tensor input_829_perm_0 = const()[name = tensor("input_829_perm_0"), val = tensor([0, 2, 1])]; tensor input_831_pad_type_0 = const()[name = tensor("input_831_pad_type_0"), val = tensor("valid")]; tensor input_831_strides_0 = const()[name = tensor("input_831_strides_0"), val = tensor([1])]; tensor input_831_pad_0 = const()[name = tensor("input_831_pad_0"), val = tensor([0, 0])]; tensor input_831_dilations_0 = const()[name = tensor("input_831_dilations_0"), val = tensor([1])]; tensor input_831_groups_0 = const()[name = tensor("input_831_groups_0"), val = tensor(1)]; tensor encoder_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194840064)))]; tensor input_829_cast_fp16 = transpose(perm = input_829_perm_0, x = x_407_cast_fp16)[name = tensor("transpose_98")]; tensor input_831_cast_fp16 = conv(dilations = input_831_dilations_0, groups = input_831_groups_0, pad = input_831_pad_0, pad_type = input_831_pad_type_0, strides = input_831_strides_0, weight = encoder_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_829_cast_fp16)[name = tensor("input_831_cast_fp16")]; tensor x_409_split_num_splits_0 = const()[name = tensor("x_409_split_num_splits_0"), val = tensor(2)]; tensor x_409_split_axis_0 = const()[name = tensor("x_409_split_axis_0"), val = tensor(1)]; tensor x_409_split_cast_fp16_0, tensor x_409_split_cast_fp16_1 = split(axis = x_409_split_axis_0, num_splits = x_409_split_num_splits_0, x = input_831_cast_fp16)[name = tensor("x_409_split_cast_fp16")]; tensor x_409_split_1_sigmoid_cast_fp16 = sigmoid(x = x_409_split_cast_fp16_1)[name = tensor("x_409_split_1_sigmoid_cast_fp16")]; tensor x_409_cast_fp16 = mul(x = x_409_split_cast_fp16_0, y = x_409_split_1_sigmoid_cast_fp16)[name = tensor("x_409_cast_fp16")]; tensor input_833_cast_fp16 = select(a = var_40_to_fp16, b = x_409_cast_fp16, cond = var_551)[name = tensor("input_833_cast_fp16")]; tensor new_x_63_interleave_0 = const()[name = tensor("new_x_63_interleave_0"), val = tensor(false)]; tensor new_x_63_cast_fp16 = concat(axis = var_56, interleave = new_x_63_interleave_0, values = (cache_63_cast_fp16, input_833_cast_fp16))[name = tensor("new_x_63_cast_fp16")]; tensor next_cache_31_begin_0 = const()[name = tensor("next_cache_31_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_31_end_0 = const()[name = tensor("next_cache_31_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_31_end_mask_0 = const()[name = tensor("next_cache_31_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_31_cast_fp16 = slice_by_index(begin = next_cache_31_begin_0, end = next_cache_31_end_0, end_mask = next_cache_31_end_mask_0, x = new_x_63_cast_fp16)[name = tensor("next_cache_31_cast_fp16")]; tensor var_3672_begin_0 = const()[name = tensor("op_3672_begin_0"), val = tensor([0, 0, 4])]; tensor var_3672_end_0 = const()[name = tensor("op_3672_end_0"), val = tensor([1, 512, 12])]; tensor var_3672_end_mask_0 = const()[name = tensor("op_3672_end_mask_0"), val = tensor([true, true, true])]; tensor var_3672_cast_fp16 = slice_by_index(begin = var_3672_begin_0, end = var_3672_end_0, end_mask = var_3672_end_mask_0, x = next_cache_31_cast_fp16)[name = tensor("op_3672_cast_fp16")]; tensor x_411_pad_type_0 = const()[name = tensor("x_411_pad_type_0"), val = tensor("valid")]; tensor x_411_groups_0 = const()[name = tensor("x_411_groups_0"), val = tensor(512)]; tensor x_411_strides_0 = const()[name = tensor("x_411_strides_0"), val = tensor([1])]; tensor x_411_pad_0 = const()[name = tensor("x_411_pad_0"), val = tensor([0, 0])]; tensor x_411_dilations_0 = const()[name = tensor("x_411_dilations_0"), val = tensor([1])]; tensor encoder_layers_15_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195888704)))]; tensor x_411_cast_fp16 = conv(dilations = x_411_dilations_0, groups = x_411_groups_0, pad = x_411_pad_0, pad_type = x_411_pad_type_0, strides = x_411_strides_0, weight = encoder_layers_15_conv_depthwise_conv_weight_to_fp16, x = new_x_63_cast_fp16)[name = tensor("x_411_cast_fp16")]; tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; tensor encoder_layers_15_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195897984)))]; tensor encoder_layers_15_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195899072)))]; tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_411_cast_fp16)[name = tensor("transpose_97")]; tensor x_413_cast_fp16 = layer_norm(axes = x_413_axes_0, beta = encoder_layers_15_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_conv_batch_norm_weight_to_fp16, x = input_835_cast_fp16)[name = tensor("x_413_cast_fp16")]; tensor input_837_perm_0 = const()[name = tensor("input_837_perm_0"), val = tensor([0, 2, 1])]; tensor input_837_cast_fp16 = transpose(perm = input_837_perm_0, x = x_413_cast_fp16)[name = tensor("transpose_96")]; tensor input_839_cast_fp16 = silu(x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; tensor x_415_pad_type_0 = const()[name = tensor("x_415_pad_type_0"), val = tensor("valid")]; tensor x_415_strides_0 = const()[name = tensor("x_415_strides_0"), val = tensor([1])]; tensor x_415_pad_0 = const()[name = tensor("x_415_pad_0"), val = tensor([0, 0])]; tensor x_415_dilations_0 = const()[name = tensor("x_415_dilations_0"), val = tensor([1])]; tensor x_415_groups_0 = const()[name = tensor("x_415_groups_0"), val = tensor(1)]; tensor encoder_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195900160)))]; tensor x_415_cast_fp16 = conv(dilations = x_415_dilations_0, groups = x_415_groups_0, pad = x_415_pad_0, pad_type = x_415_pad_type_0, strides = x_415_strides_0, weight = encoder_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("x_415_cast_fp16")]; tensor input_841_perm_0 = const()[name = tensor("input_841_perm_0"), val = tensor([0, 2, 1])]; tensor input_841_cast_fp16 = transpose(perm = input_841_perm_0, x = x_415_cast_fp16)[name = tensor("transpose_95")]; tensor input_843_cast_fp16 = add(x = input_827_cast_fp16, y = input_841_cast_fp16)[name = tensor("input_843_cast_fp16")]; tensor input_845_axes_0 = const()[name = tensor("input_845_axes_0"), val = tensor([-1])]; tensor encoder_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196424512)))]; tensor encoder_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196425600)))]; tensor input_845_cast_fp16 = layer_norm(axes = input_845_axes_0, beta = encoder_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("input_845_cast_fp16")]; tensor encoder_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196426688)))]; tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_845_cast_fp16)[name = tensor("linear_143_cast_fp16")]; tensor input_849_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_849_cast_fp16")]; tensor encoder_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198523904)))]; tensor linear_144_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("linear_144_cast_fp16")]; tensor var_3713_to_fp16 = const()[name = tensor("op_3713_to_fp16"), val = tensor(0x1p-1)]; tensor var_3714_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3713_to_fp16)[name = tensor("op_3714_cast_fp16")]; tensor input_855_cast_fp16 = add(x = input_843_cast_fp16, y = var_3714_cast_fp16)[name = tensor("input_855_cast_fp16")]; tensor input_857_axes_0 = const()[name = tensor("input_857_axes_0"), val = tensor([-1])]; tensor encoder_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200621120)))]; tensor encoder_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200622208)))]; tensor input_857_cast_fp16 = layer_norm(axes = input_857_axes_0, beta = encoder_layers_15_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_out_weight_to_fp16, x = input_855_cast_fp16)[name = tensor("input_857_cast_fp16")]; tensor cache_65_begin_0 = const()[name = tensor("cache_65_begin_0"), val = tensor([16, 0, 0, 0])]; tensor cache_65_end_0 = const()[name = tensor("cache_65_end_0"), val = tensor([17, 1, 70, 512])]; tensor cache_65_end_mask_0 = const()[name = tensor("cache_65_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_65_squeeze_mask_0 = const()[name = tensor("cache_65_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_65_cast_fp16 = slice_by_index(begin = cache_65_begin_0, end = cache_65_end_0, end_mask = cache_65_end_mask_0, squeeze_mask = cache_65_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_65_cast_fp16")]; tensor cache_begin_0 = const()[name = tensor("cache_begin_0"), val = tensor([16, 0, 0, 0])]; tensor cache_end_0 = const()[name = tensor("cache_end_0"), val = tensor([17, 1, 512, 8])]; tensor cache_end_mask_0 = const()[name = tensor("cache_end_mask_0"), val = tensor([false, true, true, true])]; tensor cache_squeeze_mask_0 = const()[name = tensor("cache_squeeze_mask_0"), val = tensor([true, false, false, false])]; tensor cache_cast_fp16 = slice_by_index(begin = cache_begin_0, end = cache_end_0, end_mask = cache_end_mask_0, squeeze_mask = cache_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_cast_fp16")]; tensor input_859_axes_0 = const()[name = tensor("input_859_axes_0"), val = tensor([-1])]; tensor encoder_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200623296)))]; tensor encoder_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200624384)))]; tensor input_859_cast_fp16 = layer_norm(axes = input_859_axes_0, beta = encoder_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward1_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("input_859_cast_fp16")]; tensor encoder_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200625472)))]; tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_859_cast_fp16)[name = tensor("linear_145_cast_fp16")]; tensor input_863_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_863_cast_fp16")]; tensor encoder_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202722688)))]; tensor linear_146_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("linear_146_cast_fp16")]; tensor var_3748_to_fp16 = const()[name = tensor("op_3748_to_fp16"), val = tensor(0x1p-1)]; tensor var_3749_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3748_to_fp16)[name = tensor("op_3749_cast_fp16")]; tensor input_869_cast_fp16 = add(x = input_857_cast_fp16, y = var_3749_cast_fp16)[name = tensor("input_869_cast_fp16")]; tensor key_axes_0 = const()[name = tensor("key_axes_0"), val = tensor([-1])]; tensor encoder_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204819904)))]; tensor encoder_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204820992)))]; tensor key_cast_fp16 = layer_norm(axes = key_axes_0, beta = encoder_layers_16_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_self_att_weight_to_fp16, x = input_869_cast_fp16)[name = tensor("key_cast_fp16")]; tensor input_871_interleave_0 = const()[name = tensor("input_871_interleave_0"), val = tensor(false)]; tensor input_871_cast_fp16 = concat(axis = var_65, interleave = input_871_interleave_0, values = (cache_65_cast_fp16, key_cast_fp16))[name = tensor("input_871_cast_fp16")]; tensor var_3771_begin_0 = const()[name = tensor("op_3771_begin_0"), val = tensor([0, 4, 0])]; tensor var_3771_end_0 = const()[name = tensor("op_3771_end_0"), val = tensor([1, 70, 512])]; tensor var_3771_end_mask_0 = const()[name = tensor("op_3771_end_mask_0"), val = tensor([true, true, true])]; tensor var_3771_cast_fp16 = slice_by_index(begin = var_3771_begin_0, end = var_3771_end_0, end_mask = var_3771_end_mask_0, x = cache_65_cast_fp16)[name = tensor("op_3771_cast_fp16")]; tensor var_3774_begin_0 = const()[name = tensor("op_3774_begin_0"), val = tensor([0, 0, 0])]; tensor var_3774_end_0 = const()[name = tensor("op_3774_end_0"), val = tensor([1, 4, 512])]; tensor var_3774_end_mask_0 = const()[name = tensor("op_3774_end_mask_0"), val = tensor([true, false, true])]; tensor var_3774_cast_fp16 = slice_by_index(begin = var_3774_begin_0, end = var_3774_end_0, end_mask = var_3774_end_mask_0, x = key_cast_fp16)[name = tensor("op_3774_cast_fp16")]; tensor cache_last_channel_cur_interleave_0 = const()[name = tensor("cache_last_channel_cur_interleave_0"), val = tensor(false)]; tensor cache_last_channel_cur_cast_fp16 = concat(axis = var_65, interleave = cache_last_channel_cur_interleave_0, values = (var_3771_cast_fp16, var_3774_cast_fp16))[name = tensor("cache_last_channel_cur_cast_fp16")]; tensor encoder_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204822080)))]; tensor linear_147_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_q_weight_to_fp16, x = key_cast_fp16)[name = tensor("linear_147_cast_fp16")]; tensor var_3781 = const()[name = tensor("op_3781"), val = tensor([1, -1, 8, 64])]; tensor q_97_cast_fp16 = reshape(shape = var_3781, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; tensor encoder_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205346432)))]; tensor linear_148_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_k_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_148_cast_fp16")]; tensor var_3785 = const()[name = tensor("op_3785"), val = tensor([1, -1, 8, 64])]; tensor k_65_cast_fp16 = reshape(shape = var_3785, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; tensor encoder_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205870784)))]; tensor linear_149_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_v_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_149_cast_fp16")]; tensor var_3789 = const()[name = tensor("op_3789"), val = tensor([1, -1, 8, 64])]; tensor v_cast_fp16 = reshape(shape = var_3789, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206395136)))]; tensor var_3801_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3801_cast_fp16")]; tensor encoder_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206396224)))]; tensor var_3803_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3803_cast_fp16")]; tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_423_transpose_x_0 = const()[name = tensor("x_423_transpose_x_0"), val = tensor(false)]; tensor x_423_transpose_y_0 = const()[name = tensor("x_423_transpose_y_0"), val = tensor(false)]; tensor var_3805_to_fp16 = const()[name = tensor("op_3805_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206397312)))]; tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3803_cast_fp16)[name = tensor("transpose_93")]; tensor x_423_cast_fp16 = matmul(transpose_x = x_423_transpose_x_0, transpose_y = x_423_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3805_to_fp16)[name = tensor("x_423_cast_fp16")]; tensor x_425_pad_0 = const()[name = tensor("x_425_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_425_mode_0 = const()[name = tensor("x_425_mode_0"), val = tensor("constant")]; tensor const_287_to_fp16 = const()[name = tensor("const_287_to_fp16"), val = tensor(0x0p+0)]; tensor x_425_cast_fp16 = pad(constant_val = const_287_to_fp16, mode = x_425_mode_0, pad = x_425_pad_0, x = x_423_cast_fp16)[name = tensor("x_425_cast_fp16")]; tensor var_3813 = const()[name = tensor("op_3813"), val = tensor([1, 8, -1, 8])]; tensor x_427_cast_fp16 = reshape(shape = var_3813, x = x_425_cast_fp16)[name = tensor("x_427_cast_fp16")]; tensor var_3817_begin_0 = const()[name = tensor("op_3817_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3817_end_0 = const()[name = tensor("op_3817_end_0"), val = tensor([1, 8, 156, 8])]; tensor var_3817_end_mask_0 = const()[name = tensor("op_3817_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3817_cast_fp16 = slice_by_index(begin = var_3817_begin_0, end = var_3817_end_0, end_mask = var_3817_end_mask_0, x = x_427_cast_fp16)[name = tensor("op_3817_cast_fp16")]; tensor var_3818 = const()[name = tensor("op_3818"), val = tensor([1, 8, 8, 155])]; tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3818, x = var_3817_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_91")]; tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3801_cast_fp16)[name = tensor("transpose_92")]; tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 8, 78])]; tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; tensor var_3827_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3827_cast_fp16")]; tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_65_cast_fp16 = mul(x = var_3827_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; tensor scores_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_11)[name = tensor("scores_cast_fp16")]; tensor var_3833_cast_fp16 = softmax(axis = var_56, x = scores_cast_fp16)[name = tensor("op_3833_cast_fp16")]; tensor input_873_cast_fp16 = select(a = var_40_to_fp16, b = var_3833_cast_fp16, cond = mask_11)[name = tensor("input_873_cast_fp16")]; tensor x_429_transpose_x_0 = const()[name = tensor("x_429_transpose_x_0"), val = tensor(false)]; tensor x_429_transpose_y_0 = const()[name = tensor("x_429_transpose_y_0"), val = tensor(false)]; tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_94")]; tensor x_429_cast_fp16 = matmul(transpose_x = x_429_transpose_x_0, transpose_y = x_429_transpose_y_0, x = input_873_cast_fp16, y = value_cast_fp16)[name = tensor("x_429_cast_fp16")]; tensor var_3837_perm_0 = const()[name = tensor("op_3837_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3838 = const()[name = tensor("op_3838"), val = tensor([1, -1, 512])]; tensor var_3837_cast_fp16 = transpose(perm = var_3837_perm_0, x = x_429_cast_fp16)[name = tensor("transpose_90")]; tensor input_875_cast_fp16 = reshape(shape = var_3838, x = var_3837_cast_fp16)[name = tensor("input_875_cast_fp16")]; tensor encoder_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206556096)))]; tensor linear_151_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_out_weight_to_fp16, x = input_875_cast_fp16)[name = tensor("linear_151_cast_fp16")]; tensor input_879_cast_fp16 = add(x = input_869_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_879_cast_fp16")]; tensor x_433_axes_0 = const()[name = tensor("x_433_axes_0"), val = tensor([-1])]; tensor encoder_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207080448)))]; tensor encoder_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207081536)))]; tensor x_433_cast_fp16 = layer_norm(axes = x_433_axes_0, beta = encoder_layers_16_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_conv_weight_to_fp16, x = input_879_cast_fp16)[name = tensor("x_433_cast_fp16")]; tensor input_881_perm_0 = const()[name = tensor("input_881_perm_0"), val = tensor([0, 2, 1])]; tensor input_883_pad_type_0 = const()[name = tensor("input_883_pad_type_0"), val = tensor("valid")]; tensor input_883_strides_0 = const()[name = tensor("input_883_strides_0"), val = tensor([1])]; tensor input_883_pad_0 = const()[name = tensor("input_883_pad_0"), val = tensor([0, 0])]; tensor input_883_dilations_0 = const()[name = tensor("input_883_dilations_0"), val = tensor([1])]; tensor input_883_groups_0 = const()[name = tensor("input_883_groups_0"), val = tensor(1)]; tensor encoder_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207082624)))]; tensor input_881_cast_fp16 = transpose(perm = input_881_perm_0, x = x_433_cast_fp16)[name = tensor("transpose_89")]; tensor input_883_cast_fp16 = conv(dilations = input_883_dilations_0, groups = input_883_groups_0, pad = input_883_pad_0, pad_type = input_883_pad_type_0, strides = input_883_strides_0, weight = encoder_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_881_cast_fp16)[name = tensor("input_883_cast_fp16")]; tensor x_435_split_num_splits_0 = const()[name = tensor("x_435_split_num_splits_0"), val = tensor(2)]; tensor x_435_split_axis_0 = const()[name = tensor("x_435_split_axis_0"), val = tensor(1)]; tensor x_435_split_cast_fp16_0, tensor x_435_split_cast_fp16_1 = split(axis = x_435_split_axis_0, num_splits = x_435_split_num_splits_0, x = input_883_cast_fp16)[name = tensor("x_435_split_cast_fp16")]; tensor x_435_split_1_sigmoid_cast_fp16 = sigmoid(x = x_435_split_cast_fp16_1)[name = tensor("x_435_split_1_sigmoid_cast_fp16")]; tensor x_435_cast_fp16 = mul(x = x_435_split_cast_fp16_0, y = x_435_split_1_sigmoid_cast_fp16)[name = tensor("x_435_cast_fp16")]; tensor input_885_cast_fp16 = select(a = var_40_to_fp16, b = x_435_cast_fp16, cond = var_551)[name = tensor("input_885_cast_fp16")]; tensor new_x_interleave_0 = const()[name = tensor("new_x_interleave_0"), val = tensor(false)]; tensor new_x_cast_fp16 = concat(axis = var_56, interleave = new_x_interleave_0, values = (cache_cast_fp16, input_885_cast_fp16))[name = tensor("new_x_cast_fp16")]; tensor next_cache_begin_0 = const()[name = tensor("next_cache_begin_0"), val = tensor([0, 0, 0])]; tensor next_cache_end_0 = const()[name = tensor("next_cache_end_0"), val = tensor([1, 512, 12])]; tensor next_cache_end_mask_0 = const()[name = tensor("next_cache_end_mask_0"), val = tensor([true, true, false])]; tensor next_cache_cast_fp16 = slice_by_index(begin = next_cache_begin_0, end = next_cache_end_0, end_mask = next_cache_end_mask_0, x = new_x_cast_fp16)[name = tensor("next_cache_cast_fp16")]; tensor cache_last_time_cur_begin_0 = const()[name = tensor("cache_last_time_cur_begin_0"), val = tensor([0, 0, 4])]; tensor cache_last_time_cur_end_0 = const()[name = tensor("cache_last_time_cur_end_0"), val = tensor([1, 512, 12])]; tensor cache_last_time_cur_end_mask_0 = const()[name = tensor("cache_last_time_cur_end_mask_0"), val = tensor([true, true, true])]; tensor cache_last_time_cur_cast_fp16 = slice_by_index(begin = cache_last_time_cur_begin_0, end = cache_last_time_cur_end_0, end_mask = cache_last_time_cur_end_mask_0, x = next_cache_cast_fp16)[name = tensor("cache_last_time_cur_cast_fp16")]; tensor x_437_pad_type_0 = const()[name = tensor("x_437_pad_type_0"), val = tensor("valid")]; tensor x_437_groups_0 = const()[name = tensor("x_437_groups_0"), val = tensor(512)]; tensor x_437_strides_0 = const()[name = tensor("x_437_strides_0"), val = tensor([1])]; tensor x_437_pad_0 = const()[name = tensor("x_437_pad_0"), val = tensor([0, 0])]; tensor x_437_dilations_0 = const()[name = tensor("x_437_dilations_0"), val = tensor([1])]; tensor encoder_layers_16_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208131264)))]; tensor x_437_cast_fp16 = conv(dilations = x_437_dilations_0, groups = x_437_groups_0, pad = x_437_pad_0, pad_type = x_437_pad_type_0, strides = x_437_strides_0, weight = encoder_layers_16_conv_depthwise_conv_weight_to_fp16, x = new_x_cast_fp16)[name = tensor("x_437_cast_fp16")]; tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; tensor x_439_axes_0 = const()[name = tensor("x_439_axes_0"), val = tensor([-1])]; tensor encoder_layers_16_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208140544)))]; tensor encoder_layers_16_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208141632)))]; tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_437_cast_fp16)[name = tensor("transpose_88")]; tensor x_439_cast_fp16 = layer_norm(axes = x_439_axes_0, beta = encoder_layers_16_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_conv_batch_norm_weight_to_fp16, x = input_887_cast_fp16)[name = tensor("x_439_cast_fp16")]; tensor input_889_perm_0 = const()[name = tensor("input_889_perm_0"), val = tensor([0, 2, 1])]; tensor input_889_cast_fp16 = transpose(perm = input_889_perm_0, x = x_439_cast_fp16)[name = tensor("transpose_87")]; tensor input_891_cast_fp16 = silu(x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; tensor encoder_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208142720)))]; tensor x_441_cast_fp16 = conv(dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = encoder_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("x_441_cast_fp16")]; tensor input_893_perm_0 = const()[name = tensor("input_893_perm_0"), val = tensor([0, 2, 1])]; tensor input_893_cast_fp16 = transpose(perm = input_893_perm_0, x = x_441_cast_fp16)[name = tensor("transpose_86")]; tensor input_895_cast_fp16 = add(x = input_879_cast_fp16, y = input_893_cast_fp16)[name = tensor("input_895_cast_fp16")]; tensor input_897_axes_0 = const()[name = tensor("input_897_axes_0"), val = tensor([-1])]; tensor encoder_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208667072)))]; tensor encoder_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208668160)))]; tensor input_897_cast_fp16 = layer_norm(axes = input_897_axes_0, beta = encoder_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("input_897_cast_fp16")]; tensor encoder_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208669248)))]; tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_897_cast_fp16)[name = tensor("linear_152_cast_fp16")]; tensor input_901_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_901_cast_fp16")]; tensor encoder_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210766464)))]; tensor linear_153_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("linear_153_cast_fp16")]; tensor var_3920_to_fp16 = const()[name = tensor("op_3920_to_fp16"), val = tensor(0x1p-1)]; tensor var_3921_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3920_to_fp16)[name = tensor("op_3921_cast_fp16")]; tensor input_cast_fp16 = add(x = input_895_cast_fp16, y = var_3921_cast_fp16)[name = tensor("input_cast_fp16")]; tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; tensor encoder_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212863680)))]; tensor encoder_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212864768)))]; tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = encoder_layers_16_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor obj_5_axis_0 = const()[name = tensor("obj_5_axis_0"), val = tensor(0)]; tensor obj_5_cast_fp16 = stack(axis = obj_5_axis_0, values = (var_465_cast_fp16, var_672_cast_fp16, var_879_cast_fp16, var_1086_cast_fp16, var_1293_cast_fp16, var_1500_cast_fp16, var_1707_cast_fp16, var_1914_cast_fp16, var_2121_cast_fp16, var_2328_cast_fp16, var_2535_cast_fp16, var_2742_cast_fp16, var_2949_cast_fp16, var_3156_cast_fp16, var_3363_cast_fp16, var_3570_cast_fp16, cache_last_channel_cur_cast_fp16))[name = tensor("obj_5_cast_fp16")]; tensor obj_7_axis_0 = const()[name = tensor("obj_7_axis_0"), val = tensor(0)]; tensor obj_7_cast_fp16 = stack(axis = obj_7_axis_0, values = (var_567_cast_fp16, var_774_cast_fp16, var_981_cast_fp16, var_1188_cast_fp16, var_1395_cast_fp16, var_1602_cast_fp16, var_1809_cast_fp16, var_2016_cast_fp16, var_2223_cast_fp16, var_2430_cast_fp16, var_2637_cast_fp16, var_2844_cast_fp16, var_3051_cast_fp16, var_3258_cast_fp16, var_3465_cast_fp16, var_3672_cast_fp16, cache_last_time_cur_cast_fp16))[name = tensor("obj_7_cast_fp16")]; tensor obj_7_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_7_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor var_3937 = add(x = cache_last_channel_len, y = cache_keep_size)[name = tensor("op_3937")]; tensor var_3937_promoted_to_fp16_dtype_0 = const()[name = tensor("op_3937_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; tensor const_293_to_fp16 = const()[name = tensor("const_293_to_fp16"), val = tensor(-inf)]; tensor var_46_promoted_to_fp16 = const()[name = tensor("op_46_promoted_to_fp16"), val = tensor(0x1.18p+6)]; tensor var_3937_to_fp16 = cast(dtype = var_3937_promoted_to_fp16_dtype_0, x = var_3937)[name = tensor("cast_184")]; tensor clip_1_cast_fp16 = clip(alpha = const_293_to_fp16, beta = var_46_promoted_to_fp16, x = var_3937_to_fp16)[name = tensor("clip_1_cast_fp16")]; tensor var_3964_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_3964_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor cast_179_dtype_0 = const()[name = tensor("cast_179_dtype_0"), val = tensor("int32")]; tensor cast_180_dtype_0 = const()[name = tensor("cast_180_dtype_0"), val = tensor("int32")]; tensor new_cache_last_channel_len = cast(dtype = cast_180_dtype_0, x = clip_1_cast_fp16)[name = tensor("cast_181")]; tensor encoded_length = cast(dtype = cast_179_dtype_0, x = clip_0_cast_fp16)[name = tensor("cast_182")]; tensor new_cache_last_channel = cast(dtype = var_3964_cast_fp16_to_fp32_dtype_0, x = obj_5_cast_fp16)[name = tensor("cast_183")]; tensor new_cache_last_time = cast(dtype = obj_7_cast_fp16_to_fp32_dtype_0, x = obj_7_cast_fp16)[name = tensor("cast_185")]; tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; tensor encoded_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_186")]; tensor new_pre_cache = cast(dtype = var_28_cast_fp16_to_fp32_dtype_0, x = var_28_cast_fp16)[name = tensor("cast_198")]; } -> (encoded_output, encoded_length, new_pre_cache, new_cache_last_channel, new_cache_last_time, new_cache_last_channel_len); }