program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3520.4.1"}, {"coremlc-version", "3520.5.1"}})] { func main(tensor mel_features, tensor mel_length) { tensor var_30 = const()[name = tensor("op_30"), val = tensor(-1)]; tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; tensor mel_features_to_fp16_dtype_0 = const()[name = tensor("mel_features_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_86_to_fp16_dtype_0 = const()[name = tensor("op_86_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_87_promoted_to_fp16 = const()[name = tensor("op_87_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor mel_length_to_fp16 = cast(dtype = var_86_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_3")]; tensor var_88_cast_fp16 = add(x = mel_length_to_fp16, y = var_87_promoted_to_fp16)[name = tensor("op_88_cast_fp16")]; tensor _inversed_90_y_0_to_fp16 = const()[name = tensor("_inversed_90_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_90_cast_fp16 = mul(x = var_88_cast_fp16, y = _inversed_90_y_0_to_fp16)[name = tensor("_inversed_90_cast_fp16")]; tensor var_91_to_fp16 = const()[name = tensor("op_91_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_1_cast_fp16 = add(x = _inversed_90_cast_fp16, y = var_91_to_fp16)[name = tensor("lengths_1_cast_fp16")]; tensor lengths_3_cast_fp16 = floor(x = lengths_1_cast_fp16)[name = tensor("lengths_3_cast_fp16")]; tensor var_95_promoted_to_fp16 = const()[name = tensor("op_95_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor var_96_cast_fp16 = add(x = lengths_3_cast_fp16, y = var_95_promoted_to_fp16)[name = tensor("op_96_cast_fp16")]; tensor _inversed_98_y_0_to_fp16 = const()[name = tensor("_inversed_98_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_98_cast_fp16 = mul(x = var_96_cast_fp16, y = _inversed_98_y_0_to_fp16)[name = tensor("_inversed_98_cast_fp16")]; tensor var_99_to_fp16 = const()[name = tensor("op_99_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_7_cast_fp16 = add(x = _inversed_98_cast_fp16, y = var_99_to_fp16)[name = tensor("lengths_7_cast_fp16")]; tensor lengths_9_cast_fp16 = floor(x = lengths_7_cast_fp16)[name = tensor("lengths_9_cast_fp16")]; tensor var_103_promoted_to_fp16 = const()[name = tensor("op_103_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor var_104_cast_fp16 = add(x = lengths_9_cast_fp16, y = var_103_promoted_to_fp16)[name = tensor("op_104_cast_fp16")]; tensor _inversed_106_y_0_to_fp16 = const()[name = tensor("_inversed_106_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_106_cast_fp16 = mul(x = var_104_cast_fp16, y = _inversed_106_y_0_to_fp16)[name = tensor("_inversed_106_cast_fp16")]; tensor var_107_to_fp16 = const()[name = tensor("op_107_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_13_cast_fp16 = add(x = _inversed_106_cast_fp16, y = var_107_to_fp16)[name = tensor("lengths_13_cast_fp16")]; tensor lengths_cast_fp16 = floor(x = lengths_13_cast_fp16)[name = tensor("lengths_cast_fp16")]; tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; tensor mel_features_to_fp16 = cast(dtype = mel_features_to_fp16_dtype_0, x = mel_features)[name = tensor("cast_2")]; tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_features_to_fp16)[name = tensor("transpose_315")]; tensor input_1_cast_fp16 = expand_dims(axes = input_1_axes_0, x = x_1_cast_fp16)[name = tensor("input_1_cast_fp16")]; tensor input_3_pad_type_0 = const()[name = tensor("input_3_pad_type_0"), val = tensor("custom")]; tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_3_strides_0 = const()[name = tensor("input_3_strides_0"), val = tensor([2, 2])]; tensor input_3_dilations_0 = const()[name = tensor("input_3_dilations_0"), val = tensor([1, 1])]; tensor input_3_groups_0 = const()[name = tensor("input_3_groups_0"), val = tensor(1)]; tensor module_pre_encode_conv_0_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_0_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2752))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2432)))]; tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3328)))]; tensor input_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = input_3_dilations_0, groups = input_3_groups_0, pad = input_3_pad_0, pad_type = input_3_pad_type_0, strides = input_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16_quantized, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; tensor input_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("custom")]; tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_7_strides_0 = const()[name = tensor("input_7_strides_0"), val = tensor([2, 2])]; tensor input_7_groups_0 = const()[name = tensor("input_7_groups_0"), val = tensor(256)]; tensor input_7_dilations_0 = const()[name = tensor("input_7_dilations_0"), val = tensor([1, 1])]; tensor module_pre_encode_conv_2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3904))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6592))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6272)))]; tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7168)))]; tensor input_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16_quantized, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; tensor input_9_pad_type_0 = const()[name = tensor("input_9_pad_type_0"), val = tensor("valid")]; tensor input_9_strides_0 = const()[name = tensor("input_9_strides_0"), val = tensor([1, 1])]; tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_9_dilations_0 = const()[name = tensor("input_9_dilations_0"), val = tensor([1, 1])]; tensor input_9_groups_0 = const()[name = tensor("input_9_groups_0"), val = tensor(1)]; tensor module_pre_encode_conv_3_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_3_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73344)))]; tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74240)))]; tensor input_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16_quantized, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; tensor input_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("custom")]; tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([2, 2])]; tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(256)]; tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; tensor module_pre_encode_conv_5_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_5_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77184)))]; tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78080)))]; tensor input_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16_quantized, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("valid")]; tensor input_15_strides_0 = const()[name = tensor("input_15_strides_0"), val = tensor([1, 1])]; tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_15_dilations_0 = const()[name = tensor("input_15_dilations_0"), val = tensor([1, 1])]; tensor input_15_groups_0 = const()[name = tensor("input_15_groups_0"), val = tensor(1)]; tensor module_pre_encode_conv_6_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_6_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78656))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144576))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144256)))]; tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145152)))]; tensor input_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = input_15_dilations_0, groups = input_15_groups_0, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = input_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16_quantized, x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; tensor x_3_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("x_3_cast_fp16")]; tensor var_157_perm_0 = const()[name = tensor("op_157_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_158 = const()[name = tensor("op_158"), val = tensor([1, 188, -1])]; tensor var_157_cast_fp16 = transpose(perm = var_157_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_314")]; tensor input_17_cast_fp16 = reshape(shape = var_158, x = var_157_cast_fp16)[name = tensor("input_17_cast_fp16")]; tensor module_pre_encode_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2768320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2767232)))]; tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2770432)))]; tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16_quantized, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; tensor padding_length_dtype_0 = const()[name = tensor("padding_length_dtype_0"), val = tensor("int32")]; tensor var_169_to_fp16 = const()[name = tensor("op_169_to_fp16"), val = tensor(0x1p+5)]; tensor x_5_cast_fp16 = mul(x = linear_0_cast_fp16, y = var_169_to_fp16)[name = tensor("x_5_cast_fp16")]; tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; tensor var_198_axes_0 = const()[name = tensor("op_198_axes_0"), val = tensor([-1])]; tensor encoder_length = cast(dtype = padding_length_dtype_0, x = lengths_cast_fp16)[name = tensor("cast_1")]; tensor var_198 = expand_dims(axes = var_198_axes_0, x = encoder_length)[name = tensor("op_198")]; tensor pad_mask_1 = less(x = expand_dims_0, y = var_198)[name = tensor("pad_mask_1")]; tensor var_200_axes_0 = const()[name = tensor("op_200_axes_0"), val = tensor([1])]; tensor var_200 = expand_dims(axes = var_200_axes_0, x = pad_mask_1)[name = tensor("op_200")]; tensor var_201 = const()[name = tensor("op_201"), val = tensor([1, 188, 1])]; tensor pad_mask_for_att_mask_1 = tile(reps = var_201, x = var_200)[name = tensor("pad_mask_for_att_mask_1")]; tensor var_203_perm_0 = const()[name = tensor("op_203_perm_0"), val = tensor([0, 2, 1])]; tensor var_203 = transpose(perm = var_203_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_313")]; tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_203)[name = tensor("pad_mask_for_att_mask")]; tensor const_7 = const()[name = tensor("const_7"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_7)[name = tensor("att_mask")]; tensor mask_1 = logical_not(x = att_mask)[name = tensor("mask_1")]; tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2772544)))]; tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2774656)))]; tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = x_5_cast_fp16)[name = tensor("input_21_cast_fp16")]; tensor module_layers_0_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2776768))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6975296))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6971136)))]; tensor module_layers_0_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6983552)))]; tensor linear_1_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear1_bias_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16_quantized, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; tensor module_layers_0_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6991808))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11187264))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11186176)))]; tensor module_layers_0_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11189376)))]; tensor linear_2_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear2_bias_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16_quantized, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; tensor var_236_to_fp16 = const()[name = tensor("op_236_to_fp16"), val = tensor(0x1p-1)]; tensor var_237_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_236_to_fp16)[name = tensor("op_237_cast_fp16")]; tensor input_31_cast_fp16 = add(x = x_5_cast_fp16, y = var_237_cast_fp16)[name = tensor("input_31_cast_fp16")]; tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11191488)))]; tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11193600)))]; tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; tensor module_layers_0_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11195712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12245440))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12244352)))]; tensor module_layers_0_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12247552)))]; tensor linear_3_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_q_bias_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16_quantized, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; tensor var_254 = const()[name = tensor("op_254"), val = tensor([1, -1, 8, 128])]; tensor q_1_cast_fp16 = reshape(shape = var_254, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; tensor module_layers_0_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12249664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13299392))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13298304)))]; tensor module_layers_0_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13301504)))]; tensor linear_4_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_k_bias_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16_quantized, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; tensor var_259 = const()[name = tensor("op_259"), val = tensor([1, -1, 8, 128])]; tensor k_1_cast_fp16 = reshape(shape = var_259, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; tensor module_layers_0_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13303616))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14353344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14352256)))]; tensor module_layers_0_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14355456)))]; tensor linear_5_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_v_bias_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16_quantized, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; tensor var_264 = const()[name = tensor("op_264"), val = tensor([1, -1, 8, 128])]; tensor v_1_cast_fp16 = reshape(shape = var_264, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14357568)))]; tensor var_276_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_276_cast_fp16")]; tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14359680)))]; tensor var_278_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_278_cast_fp16")]; tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_9_transpose_x_0 = const()[name = tensor("x_9_transpose_x_0"), val = tensor(false)]; tensor x_9_transpose_y_0 = const()[name = tensor("x_9_transpose_y_0"), val = tensor(false)]; tensor op_280_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_280_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14361792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14746304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14745856)))]; tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_278_cast_fp16)[name = tensor("transpose_312")]; tensor x_9_cast_fp16 = matmul(transpose_x = x_9_transpose_x_0, transpose_y = x_9_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = op_280_to_fp16_quantized)[name = tensor("x_9_cast_fp16")]; tensor x_11_pad_0 = const()[name = tensor("x_11_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_11_mode_0 = const()[name = tensor("x_11_mode_0"), val = tensor("constant")]; tensor const_14_to_fp16 = const()[name = tensor("const_14_to_fp16"), val = tensor(0x0p+0)]; tensor x_11_cast_fp16 = pad(constant_val = const_14_to_fp16, mode = x_11_mode_0, pad = x_11_pad_0, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; tensor var_288 = const()[name = tensor("op_288"), val = tensor([1, 8, -1, 188])]; tensor x_13_cast_fp16 = reshape(shape = var_288, x = x_11_cast_fp16)[name = tensor("x_13_cast_fp16")]; tensor var_292_begin_0 = const()[name = tensor("op_292_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_292_end_0 = const()[name = tensor("op_292_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_292_end_mask_0 = const()[name = tensor("op_292_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_292_cast_fp16 = slice_by_index(begin = var_292_begin_0, end = var_292_end_0, end_mask = var_292_end_mask_0, x = x_13_cast_fp16)[name = tensor("op_292_cast_fp16")]; tensor var_293 = const()[name = tensor("op_293"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_1_cast_fp16 = reshape(shape = var_293, x = var_292_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; tensor transpose_96_perm_0 = const()[name = tensor("transpose_96_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_97_perm_0 = const()[name = tensor("transpose_97_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_97 = transpose(perm = transpose_97_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_310")]; tensor transpose_96 = transpose(perm = transpose_96_perm_0, x = var_276_cast_fp16)[name = tensor("transpose_311")]; tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_96, y = transpose_97)[name = tensor("matrix_ac_1_cast_fp16")]; tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; tensor var_302_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_302_cast_fp16")]; tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_1_cast_fp16 = mul(x = var_302_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_3)[name = tensor("scores_3_cast_fp16")]; tensor var_308_cast_fp16 = softmax(axis = var_30, x = scores_3_cast_fp16)[name = tensor("op_308_cast_fp16")]; tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_308_cast_fp16, cond = mask_3)[name = tensor("input_33_cast_fp16")]; tensor x_15_transpose_x_0 = const()[name = tensor("x_15_transpose_x_0"), val = tensor(false)]; tensor x_15_transpose_y_0 = const()[name = tensor("x_15_transpose_y_0"), val = tensor(false)]; tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_309")]; tensor x_15_cast_fp16 = matmul(transpose_x = x_15_transpose_x_0, transpose_y = x_15_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_15_cast_fp16")]; tensor var_312_perm_0 = const()[name = tensor("op_312_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_313 = const()[name = tensor("op_313"), val = tensor([1, -1, 1024])]; tensor var_312_cast_fp16 = transpose(perm = var_312_perm_0, x = x_15_cast_fp16)[name = tensor("transpose_308")]; tensor input_35_cast_fp16 = reshape(shape = var_313, x = var_312_cast_fp16)[name = tensor("input_35_cast_fp16")]; tensor module_layers_0_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14747136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15796864))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15795776)))]; tensor module_layers_0_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15798976)))]; tensor linear_7_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_out_bias_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16_quantized, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; tensor x_19_axes_0 = const()[name = tensor("x_19_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15801088)))]; tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15803200)))]; tensor x_19_cast_fp16 = layer_norm(axes = x_19_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_19_cast_fp16")]; tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15805312))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17904640))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17902528)))]; tensor module_layers_0_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17908800)))]; tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_19_cast_fp16)[name = tensor("transpose_307")]; tensor input_43_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv1_bias_to_fp16, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; tensor x_21_split_num_splits_0 = const()[name = tensor("x_21_split_num_splits_0"), val = tensor(2)]; tensor x_21_split_axis_0 = const()[name = tensor("x_21_split_axis_0"), val = tensor(1)]; tensor x_21_split_cast_fp16_0, tensor x_21_split_cast_fp16_1 = split(axis = x_21_split_axis_0, num_splits = x_21_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_21_split_cast_fp16")]; tensor x_21_split_1_sigmoid_cast_fp16 = sigmoid(x = x_21_split_cast_fp16_1)[name = tensor("x_21_split_1_sigmoid_cast_fp16")]; tensor x_21_cast_fp16 = mul(x = x_21_split_cast_fp16_0, y = x_21_split_1_sigmoid_cast_fp16)[name = tensor("x_21_cast_fp16")]; tensor var_337_axes_0 = const()[name = tensor("op_337_axes_0"), val = tensor([1])]; tensor var_337 = expand_dims(axes = var_337_axes_0, x = pad_mask)[name = tensor("op_337")]; tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_21_cast_fp16, cond = var_337)[name = tensor("input_45_cast_fp16")]; tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; tensor const_17_to_fp16 = const()[name = tensor("const_17_to_fp16"), val = tensor(0x0p+0)]; tensor input_47_cast_fp16 = pad(constant_val = const_17_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(1024)]; tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; tensor const_248_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_248_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17912960))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17923328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17922240)))]; tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17925440)))]; tensor input_51_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_248_to_fp16_quantized, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; tensor x_23_pad_type_0 = const()[name = tensor("x_23_pad_type_0"), val = tensor("valid")]; tensor x_23_strides_0 = const()[name = tensor("x_23_strides_0"), val = tensor([1])]; tensor x_23_pad_0 = const()[name = tensor("x_23_pad_0"), val = tensor([0, 0])]; tensor x_23_dilations_0 = const()[name = tensor("x_23_dilations_0"), val = tensor([1])]; tensor x_23_groups_0 = const()[name = tensor("x_23_groups_0"), val = tensor(1)]; tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17927552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18977280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18976192)))]; tensor module_layers_0_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18979392)))]; tensor x_23_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv2_bias_to_fp16, dilations = x_23_dilations_0, groups = x_23_groups_0, pad = x_23_pad_0, pad_type = x_23_pad_type_0, strides = x_23_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_53_cast_fp16)[name = tensor("x_23_cast_fp16")]; tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_23_cast_fp16)[name = tensor("transpose_306")]; tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18981504)))]; tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18983616)))]; tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; tensor module_layers_0_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18985728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23184256))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23180096)))]; tensor module_layers_0_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23192512)))]; tensor linear_8_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear1_bias_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16_quantized, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; tensor module_layers_0_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23200768))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27396224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27395136)))]; tensor module_layers_0_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27398336)))]; tensor linear_9_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear2_bias_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16_quantized, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; tensor var_379_to_fp16 = const()[name = tensor("op_379_to_fp16"), val = tensor(0x1p-1)]; tensor var_380_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_379_to_fp16)[name = tensor("op_380_cast_fp16")]; tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_380_cast_fp16)[name = tensor("input_69_cast_fp16")]; tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27400448)))]; tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27402560)))]; tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27404672)))]; tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27406784)))]; tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; tensor module_layers_1_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27408896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31607424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31603264)))]; tensor module_layers_1_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31615680)))]; tensor linear_10_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear1_bias_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16_quantized, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; tensor module_layers_1_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31623936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35819392))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35818304)))]; tensor module_layers_1_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35821504)))]; tensor linear_11_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear2_bias_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16_quantized, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; tensor var_410_to_fp16 = const()[name = tensor("op_410_to_fp16"), val = tensor(0x1p-1)]; tensor var_411_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_410_to_fp16)[name = tensor("op_411_cast_fp16")]; tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_411_cast_fp16)[name = tensor("input_83_cast_fp16")]; tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35823616)))]; tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35825728)))]; tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; tensor module_layers_1_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35827840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36877568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36876480)))]; tensor module_layers_1_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36879680)))]; tensor linear_12_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_q_bias_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16_quantized, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; tensor var_428 = const()[name = tensor("op_428"), val = tensor([1, -1, 8, 128])]; tensor q_7_cast_fp16 = reshape(shape = var_428, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; tensor module_layers_1_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36881792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37931520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37930432)))]; tensor module_layers_1_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37933632)))]; tensor linear_13_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_k_bias_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16_quantized, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; tensor var_433 = const()[name = tensor("op_433"), val = tensor([1, -1, 8, 128])]; tensor k_5_cast_fp16 = reshape(shape = var_433, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; tensor module_layers_1_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37935744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38985472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38984384)))]; tensor module_layers_1_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38987584)))]; tensor linear_14_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_v_bias_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16_quantized, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; tensor var_438 = const()[name = tensor("op_438"), val = tensor([1, -1, 8, 128])]; tensor v_3_cast_fp16 = reshape(shape = var_438, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38989696)))]; tensor var_450_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_450_cast_fp16")]; tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38991808)))]; tensor var_452_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_452_cast_fp16")]; tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_31_transpose_x_0 = const()[name = tensor("x_31_transpose_x_0"), val = tensor(false)]; tensor x_31_transpose_y_0 = const()[name = tensor("x_31_transpose_y_0"), val = tensor(false)]; tensor op_454_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_454_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38993920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39378432))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39377984)))]; tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_452_cast_fp16)[name = tensor("transpose_305")]; tensor x_31_cast_fp16 = matmul(transpose_x = x_31_transpose_x_0, transpose_y = x_31_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = op_454_to_fp16_quantized)[name = tensor("x_31_cast_fp16")]; tensor x_33_pad_0 = const()[name = tensor("x_33_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_33_mode_0 = const()[name = tensor("x_33_mode_0"), val = tensor("constant")]; tensor const_24_to_fp16 = const()[name = tensor("const_24_to_fp16"), val = tensor(0x0p+0)]; tensor x_33_cast_fp16 = pad(constant_val = const_24_to_fp16, mode = x_33_mode_0, pad = x_33_pad_0, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; tensor var_462 = const()[name = tensor("op_462"), val = tensor([1, 8, -1, 188])]; tensor x_35_cast_fp16 = reshape(shape = var_462, x = x_33_cast_fp16)[name = tensor("x_35_cast_fp16")]; tensor var_466_begin_0 = const()[name = tensor("op_466_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_466_end_0 = const()[name = tensor("op_466_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_466_end_mask_0 = const()[name = tensor("op_466_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_466_cast_fp16 = slice_by_index(begin = var_466_begin_0, end = var_466_end_0, end_mask = var_466_end_mask_0, x = x_35_cast_fp16)[name = tensor("op_466_cast_fp16")]; tensor var_467 = const()[name = tensor("op_467"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_5_cast_fp16 = reshape(shape = var_467, x = var_466_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; tensor transpose_98_perm_0 = const()[name = tensor("transpose_98_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_99_perm_0 = const()[name = tensor("transpose_99_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_99 = transpose(perm = transpose_99_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_303")]; tensor transpose_98 = transpose(perm = transpose_98_perm_0, x = var_450_cast_fp16)[name = tensor("transpose_304")]; tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_98, y = transpose_99)[name = tensor("matrix_ac_3_cast_fp16")]; tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; tensor var_476_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_476_cast_fp16")]; tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_5_cast_fp16 = mul(x = var_476_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_3)[name = tensor("scores_7_cast_fp16")]; tensor var_482_cast_fp16 = softmax(axis = var_30, x = scores_7_cast_fp16)[name = tensor("op_482_cast_fp16")]; tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_482_cast_fp16, cond = mask_3)[name = tensor("input_85_cast_fp16")]; tensor x_37_transpose_x_0 = const()[name = tensor("x_37_transpose_x_0"), val = tensor(false)]; tensor x_37_transpose_y_0 = const()[name = tensor("x_37_transpose_y_0"), val = tensor(false)]; tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_302")]; tensor x_37_cast_fp16 = matmul(transpose_x = x_37_transpose_x_0, transpose_y = x_37_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_37_cast_fp16")]; tensor var_486_perm_0 = const()[name = tensor("op_486_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_487 = const()[name = tensor("op_487"), val = tensor([1, -1, 1024])]; tensor var_486_cast_fp16 = transpose(perm = var_486_perm_0, x = x_37_cast_fp16)[name = tensor("transpose_301")]; tensor input_87_cast_fp16 = reshape(shape = var_487, x = var_486_cast_fp16)[name = tensor("input_87_cast_fp16")]; tensor module_layers_1_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39379264))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40428992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40427904)))]; tensor module_layers_1_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40431104)))]; tensor linear_16_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_out_bias_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16_quantized, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; tensor x_41_axes_0 = const()[name = tensor("x_41_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40433216)))]; tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40435328)))]; tensor x_41_cast_fp16 = layer_norm(axes = x_41_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_41_cast_fp16")]; tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40437440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42536768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42534656)))]; tensor module_layers_1_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42540928)))]; tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_41_cast_fp16)[name = tensor("transpose_300")]; tensor input_95_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv1_bias_to_fp16, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; tensor x_43_split_num_splits_0 = const()[name = tensor("x_43_split_num_splits_0"), val = tensor(2)]; tensor x_43_split_axis_0 = const()[name = tensor("x_43_split_axis_0"), val = tensor(1)]; tensor x_43_split_cast_fp16_0, tensor x_43_split_cast_fp16_1 = split(axis = x_43_split_axis_0, num_splits = x_43_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_43_split_cast_fp16")]; tensor x_43_split_1_sigmoid_cast_fp16 = sigmoid(x = x_43_split_cast_fp16_1)[name = tensor("x_43_split_1_sigmoid_cast_fp16")]; tensor x_43_cast_fp16 = mul(x = x_43_split_cast_fp16_0, y = x_43_split_1_sigmoid_cast_fp16)[name = tensor("x_43_cast_fp16")]; tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_43_cast_fp16, cond = var_337)[name = tensor("input_97_cast_fp16")]; tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; tensor const_27_to_fp16 = const()[name = tensor("const_27_to_fp16"), val = tensor(0x0p+0)]; tensor input_99_cast_fp16 = pad(constant_val = const_27_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(1024)]; tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; tensor const_250_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_250_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42545088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42555456))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42554368)))]; tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42557568)))]; tensor input_103_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_250_to_fp16_quantized, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; tensor x_45_pad_type_0 = const()[name = tensor("x_45_pad_type_0"), val = tensor("valid")]; tensor x_45_strides_0 = const()[name = tensor("x_45_strides_0"), val = tensor([1])]; tensor x_45_pad_0 = const()[name = tensor("x_45_pad_0"), val = tensor([0, 0])]; tensor x_45_dilations_0 = const()[name = tensor("x_45_dilations_0"), val = tensor([1])]; tensor x_45_groups_0 = const()[name = tensor("x_45_groups_0"), val = tensor(1)]; tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42559680))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43609408))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43608320)))]; tensor module_layers_1_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43611520)))]; tensor x_45_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv2_bias_to_fp16, dilations = x_45_dilations_0, groups = x_45_groups_0, pad = x_45_pad_0, pad_type = x_45_pad_type_0, strides = x_45_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_105_cast_fp16)[name = tensor("x_45_cast_fp16")]; tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_45_cast_fp16)[name = tensor("transpose_299")]; tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43613632)))]; tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43615744)))]; tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; tensor module_layers_1_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43617856))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47816384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47812224)))]; tensor module_layers_1_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47824640)))]; tensor linear_17_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear1_bias_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16_quantized, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; tensor module_layers_1_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47832896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52028352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52027264)))]; tensor module_layers_1_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52030464)))]; tensor linear_18_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear2_bias_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16_quantized, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; tensor var_553_to_fp16 = const()[name = tensor("op_553_to_fp16"), val = tensor(0x1p-1)]; tensor var_554_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_553_to_fp16)[name = tensor("op_554_cast_fp16")]; tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_554_cast_fp16)[name = tensor("input_121_cast_fp16")]; tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52032576)))]; tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52034688)))]; tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52036800)))]; tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52038912)))]; tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; tensor module_layers_2_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52041024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56239552))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56235392)))]; tensor module_layers_2_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56247808)))]; tensor linear_19_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear1_bias_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16_quantized, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; tensor module_layers_2_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56256064))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60451520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60450432)))]; tensor module_layers_2_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60453632)))]; tensor linear_20_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear2_bias_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16_quantized, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; tensor var_584_to_fp16 = const()[name = tensor("op_584_to_fp16"), val = tensor(0x1p-1)]; tensor var_585_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_584_to_fp16)[name = tensor("op_585_cast_fp16")]; tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_585_cast_fp16)[name = tensor("input_135_cast_fp16")]; tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60455744)))]; tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60457856)))]; tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; tensor module_layers_2_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60459968))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61509696))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61508608)))]; tensor module_layers_2_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61511808)))]; tensor linear_21_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_q_bias_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16_quantized, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; tensor var_602 = const()[name = tensor("op_602"), val = tensor([1, -1, 8, 128])]; tensor q_13_cast_fp16 = reshape(shape = var_602, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; tensor module_layers_2_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61513920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62563648))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62562560)))]; tensor module_layers_2_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62565760)))]; tensor linear_22_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_k_bias_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16_quantized, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; tensor var_607 = const()[name = tensor("op_607"), val = tensor([1, -1, 8, 128])]; tensor k_9_cast_fp16 = reshape(shape = var_607, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; tensor module_layers_2_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62567872))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63617600))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63616512)))]; tensor module_layers_2_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63619712)))]; tensor linear_23_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_v_bias_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16_quantized, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; tensor var_612 = const()[name = tensor("op_612"), val = tensor([1, -1, 8, 128])]; tensor v_5_cast_fp16 = reshape(shape = var_612, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63621824)))]; tensor var_624_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_624_cast_fp16")]; tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63623936)))]; tensor var_626_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_626_cast_fp16")]; tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_53_transpose_x_0 = const()[name = tensor("x_53_transpose_x_0"), val = tensor(false)]; tensor x_53_transpose_y_0 = const()[name = tensor("x_53_transpose_y_0"), val = tensor(false)]; tensor op_628_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_628_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63626048))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64010560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64010112)))]; tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_626_cast_fp16)[name = tensor("transpose_298")]; tensor x_53_cast_fp16 = matmul(transpose_x = x_53_transpose_x_0, transpose_y = x_53_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = op_628_to_fp16_quantized)[name = tensor("x_53_cast_fp16")]; tensor x_55_pad_0 = const()[name = tensor("x_55_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_55_mode_0 = const()[name = tensor("x_55_mode_0"), val = tensor("constant")]; tensor const_34_to_fp16 = const()[name = tensor("const_34_to_fp16"), val = tensor(0x0p+0)]; tensor x_55_cast_fp16 = pad(constant_val = const_34_to_fp16, mode = x_55_mode_0, pad = x_55_pad_0, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; tensor var_636 = const()[name = tensor("op_636"), val = tensor([1, 8, -1, 188])]; tensor x_57_cast_fp16 = reshape(shape = var_636, x = x_55_cast_fp16)[name = tensor("x_57_cast_fp16")]; tensor var_640_begin_0 = const()[name = tensor("op_640_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_640_end_0 = const()[name = tensor("op_640_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_640_end_mask_0 = const()[name = tensor("op_640_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_640_cast_fp16 = slice_by_index(begin = var_640_begin_0, end = var_640_end_0, end_mask = var_640_end_mask_0, x = x_57_cast_fp16)[name = tensor("op_640_cast_fp16")]; tensor var_641 = const()[name = tensor("op_641"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_9_cast_fp16 = reshape(shape = var_641, x = var_640_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; tensor transpose_100_perm_0 = const()[name = tensor("transpose_100_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_101_perm_0 = const()[name = tensor("transpose_101_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_101 = transpose(perm = transpose_101_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_296")]; tensor transpose_100 = transpose(perm = transpose_100_perm_0, x = var_624_cast_fp16)[name = tensor("transpose_297")]; tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_100, y = transpose_101)[name = tensor("matrix_ac_5_cast_fp16")]; tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; tensor var_650_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_650_cast_fp16")]; tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_9_cast_fp16 = mul(x = var_650_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_3)[name = tensor("scores_11_cast_fp16")]; tensor var_656_cast_fp16 = softmax(axis = var_30, x = scores_11_cast_fp16)[name = tensor("op_656_cast_fp16")]; tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_656_cast_fp16, cond = mask_3)[name = tensor("input_137_cast_fp16")]; tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_295")]; tensor x_59_cast_fp16 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_59_cast_fp16")]; tensor var_660_perm_0 = const()[name = tensor("op_660_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_661 = const()[name = tensor("op_661"), val = tensor([1, -1, 1024])]; tensor var_660_cast_fp16 = transpose(perm = var_660_perm_0, x = x_59_cast_fp16)[name = tensor("transpose_294")]; tensor input_139_cast_fp16 = reshape(shape = var_661, x = var_660_cast_fp16)[name = tensor("input_139_cast_fp16")]; tensor module_layers_2_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64011392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65061120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65060032)))]; tensor module_layers_2_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65063232)))]; tensor linear_25_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_out_bias_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16_quantized, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; tensor x_63_axes_0 = const()[name = tensor("x_63_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65065344)))]; tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65067456)))]; tensor x_63_cast_fp16 = layer_norm(axes = x_63_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_63_cast_fp16")]; tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65069568))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67168896))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67166784)))]; tensor module_layers_2_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67173056)))]; tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_63_cast_fp16)[name = tensor("transpose_293")]; tensor input_147_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv1_bias_to_fp16, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; tensor x_65_split_num_splits_0 = const()[name = tensor("x_65_split_num_splits_0"), val = tensor(2)]; tensor x_65_split_axis_0 = const()[name = tensor("x_65_split_axis_0"), val = tensor(1)]; tensor x_65_split_cast_fp16_0, tensor x_65_split_cast_fp16_1 = split(axis = x_65_split_axis_0, num_splits = x_65_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_65_split_cast_fp16")]; tensor x_65_split_1_sigmoid_cast_fp16 = sigmoid(x = x_65_split_cast_fp16_1)[name = tensor("x_65_split_1_sigmoid_cast_fp16")]; tensor x_65_cast_fp16 = mul(x = x_65_split_cast_fp16_0, y = x_65_split_1_sigmoid_cast_fp16)[name = tensor("x_65_cast_fp16")]; tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_65_cast_fp16, cond = var_337)[name = tensor("input_149_cast_fp16")]; tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; tensor const_37_to_fp16 = const()[name = tensor("const_37_to_fp16"), val = tensor(0x0p+0)]; tensor input_151_cast_fp16 = pad(constant_val = const_37_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(1024)]; tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; tensor const_252_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_252_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67177216))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67187584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67186496)))]; tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67189696)))]; tensor input_155_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_252_to_fp16_quantized, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; tensor x_67_pad_type_0 = const()[name = tensor("x_67_pad_type_0"), val = tensor("valid")]; tensor x_67_strides_0 = const()[name = tensor("x_67_strides_0"), val = tensor([1])]; tensor x_67_pad_0 = const()[name = tensor("x_67_pad_0"), val = tensor([0, 0])]; tensor x_67_dilations_0 = const()[name = tensor("x_67_dilations_0"), val = tensor([1])]; tensor x_67_groups_0 = const()[name = tensor("x_67_groups_0"), val = tensor(1)]; tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67191808))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68241536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68240448)))]; tensor module_layers_2_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68243648)))]; tensor x_67_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv2_bias_to_fp16, dilations = x_67_dilations_0, groups = x_67_groups_0, pad = x_67_pad_0, pad_type = x_67_pad_type_0, strides = x_67_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_157_cast_fp16)[name = tensor("x_67_cast_fp16")]; tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_67_cast_fp16)[name = tensor("transpose_292")]; tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68245760)))]; tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68247872)))]; tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; tensor module_layers_2_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68249984))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72448512))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72444352)))]; tensor module_layers_2_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72456768)))]; tensor linear_26_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear1_bias_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16_quantized, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; tensor module_layers_2_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72465024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76660480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76659392)))]; tensor module_layers_2_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76662592)))]; tensor linear_27_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear2_bias_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16_quantized, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; tensor var_727_to_fp16 = const()[name = tensor("op_727_to_fp16"), val = tensor(0x1p-1)]; tensor var_728_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_727_to_fp16)[name = tensor("op_728_cast_fp16")]; tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_728_cast_fp16)[name = tensor("input_173_cast_fp16")]; tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76664704)))]; tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76666816)))]; tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76668928)))]; tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76671040)))]; tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; tensor module_layers_3_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76673152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80871680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80867520)))]; tensor module_layers_3_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80879936)))]; tensor linear_28_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear1_bias_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16_quantized, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; tensor module_layers_3_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80888192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85083648))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85082560)))]; tensor module_layers_3_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85085760)))]; tensor linear_29_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear2_bias_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16_quantized, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; tensor var_758_to_fp16 = const()[name = tensor("op_758_to_fp16"), val = tensor(0x1p-1)]; tensor var_759_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_758_to_fp16)[name = tensor("op_759_cast_fp16")]; tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_759_cast_fp16)[name = tensor("input_187_cast_fp16")]; tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85087872)))]; tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85089984)))]; tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; tensor module_layers_3_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85092096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86141824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86140736)))]; tensor module_layers_3_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86143936)))]; tensor linear_30_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_q_bias_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16_quantized, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; tensor var_776 = const()[name = tensor("op_776"), val = tensor([1, -1, 8, 128])]; tensor q_19_cast_fp16 = reshape(shape = var_776, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; tensor module_layers_3_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86146048))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87195776))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87194688)))]; tensor module_layers_3_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87197888)))]; tensor linear_31_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_k_bias_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16_quantized, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; tensor var_781 = const()[name = tensor("op_781"), val = tensor([1, -1, 8, 128])]; tensor k_13_cast_fp16 = reshape(shape = var_781, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; tensor module_layers_3_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87200000))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88249728))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88248640)))]; tensor module_layers_3_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88251840)))]; tensor linear_32_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_v_bias_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16_quantized, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; tensor var_786 = const()[name = tensor("op_786"), val = tensor([1, -1, 8, 128])]; tensor v_7_cast_fp16 = reshape(shape = var_786, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88253952)))]; tensor var_798_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_798_cast_fp16")]; tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88256064)))]; tensor var_800_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_800_cast_fp16")]; tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_75_transpose_x_0 = const()[name = tensor("x_75_transpose_x_0"), val = tensor(false)]; tensor x_75_transpose_y_0 = const()[name = tensor("x_75_transpose_y_0"), val = tensor(false)]; tensor op_802_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_802_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88258176))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88642688))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88642240)))]; tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_800_cast_fp16)[name = tensor("transpose_291")]; tensor x_75_cast_fp16 = matmul(transpose_x = x_75_transpose_x_0, transpose_y = x_75_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = op_802_to_fp16_quantized)[name = tensor("x_75_cast_fp16")]; tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_77_mode_0 = const()[name = tensor("x_77_mode_0"), val = tensor("constant")]; tensor const_44_to_fp16 = const()[name = tensor("const_44_to_fp16"), val = tensor(0x0p+0)]; tensor x_77_cast_fp16 = pad(constant_val = const_44_to_fp16, mode = x_77_mode_0, pad = x_77_pad_0, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; tensor var_810 = const()[name = tensor("op_810"), val = tensor([1, 8, -1, 188])]; tensor x_79_cast_fp16 = reshape(shape = var_810, x = x_77_cast_fp16)[name = tensor("x_79_cast_fp16")]; tensor var_814_begin_0 = const()[name = tensor("op_814_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_814_end_0 = const()[name = tensor("op_814_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_814_end_mask_0 = const()[name = tensor("op_814_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_814_cast_fp16 = slice_by_index(begin = var_814_begin_0, end = var_814_end_0, end_mask = var_814_end_mask_0, x = x_79_cast_fp16)[name = tensor("op_814_cast_fp16")]; tensor var_815 = const()[name = tensor("op_815"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_13_cast_fp16 = reshape(shape = var_815, x = var_814_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; tensor transpose_102_perm_0 = const()[name = tensor("transpose_102_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_103_perm_0 = const()[name = tensor("transpose_103_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_103 = transpose(perm = transpose_103_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_289")]; tensor transpose_102 = transpose(perm = transpose_102_perm_0, x = var_798_cast_fp16)[name = tensor("transpose_290")]; tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_102, y = transpose_103)[name = tensor("matrix_ac_7_cast_fp16")]; tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; tensor var_824_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_824_cast_fp16")]; tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_13_cast_fp16 = mul(x = var_824_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_3)[name = tensor("scores_15_cast_fp16")]; tensor var_830_cast_fp16 = softmax(axis = var_30, x = scores_15_cast_fp16)[name = tensor("op_830_cast_fp16")]; tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_830_cast_fp16, cond = mask_3)[name = tensor("input_189_cast_fp16")]; tensor x_81_transpose_x_0 = const()[name = tensor("x_81_transpose_x_0"), val = tensor(false)]; tensor x_81_transpose_y_0 = const()[name = tensor("x_81_transpose_y_0"), val = tensor(false)]; tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_288")]; tensor x_81_cast_fp16 = matmul(transpose_x = x_81_transpose_x_0, transpose_y = x_81_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_81_cast_fp16")]; tensor var_834_perm_0 = const()[name = tensor("op_834_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_835 = const()[name = tensor("op_835"), val = tensor([1, -1, 1024])]; tensor var_834_cast_fp16 = transpose(perm = var_834_perm_0, x = x_81_cast_fp16)[name = tensor("transpose_287")]; tensor input_191_cast_fp16 = reshape(shape = var_835, x = var_834_cast_fp16)[name = tensor("input_191_cast_fp16")]; tensor module_layers_3_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88643520))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89693248))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89692160)))]; tensor module_layers_3_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89695360)))]; tensor linear_34_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_out_bias_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16_quantized, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; tensor x_85_axes_0 = const()[name = tensor("x_85_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89697472)))]; tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89699584)))]; tensor x_85_cast_fp16 = layer_norm(axes = x_85_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_85_cast_fp16")]; tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89701696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91801024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91798912)))]; tensor module_layers_3_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91805184)))]; tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_85_cast_fp16)[name = tensor("transpose_286")]; tensor input_199_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv1_bias_to_fp16, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; tensor x_87_split_num_splits_0 = const()[name = tensor("x_87_split_num_splits_0"), val = tensor(2)]; tensor x_87_split_axis_0 = const()[name = tensor("x_87_split_axis_0"), val = tensor(1)]; tensor x_87_split_cast_fp16_0, tensor x_87_split_cast_fp16_1 = split(axis = x_87_split_axis_0, num_splits = x_87_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_87_split_cast_fp16")]; tensor x_87_split_1_sigmoid_cast_fp16 = sigmoid(x = x_87_split_cast_fp16_1)[name = tensor("x_87_split_1_sigmoid_cast_fp16")]; tensor x_87_cast_fp16 = mul(x = x_87_split_cast_fp16_0, y = x_87_split_1_sigmoid_cast_fp16)[name = tensor("x_87_cast_fp16")]; tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_87_cast_fp16, cond = var_337)[name = tensor("input_201_cast_fp16")]; tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; tensor const_47_to_fp16 = const()[name = tensor("const_47_to_fp16"), val = tensor(0x0p+0)]; tensor input_203_cast_fp16 = pad(constant_val = const_47_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(1024)]; tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; tensor const_254_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_254_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91809344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91819712))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91818624)))]; tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91821824)))]; tensor input_207_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_254_to_fp16_quantized, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; tensor x_89_pad_type_0 = const()[name = tensor("x_89_pad_type_0"), val = tensor("valid")]; tensor x_89_strides_0 = const()[name = tensor("x_89_strides_0"), val = tensor([1])]; tensor x_89_pad_0 = const()[name = tensor("x_89_pad_0"), val = tensor([0, 0])]; tensor x_89_dilations_0 = const()[name = tensor("x_89_dilations_0"), val = tensor([1])]; tensor x_89_groups_0 = const()[name = tensor("x_89_groups_0"), val = tensor(1)]; tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91823936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92873664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92872576)))]; tensor module_layers_3_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92875776)))]; tensor x_89_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv2_bias_to_fp16, dilations = x_89_dilations_0, groups = x_89_groups_0, pad = x_89_pad_0, pad_type = x_89_pad_type_0, strides = x_89_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_209_cast_fp16)[name = tensor("x_89_cast_fp16")]; tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_89_cast_fp16)[name = tensor("transpose_285")]; tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92877888)))]; tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92880000)))]; tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; tensor module_layers_3_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92882112))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97080640))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97076480)))]; tensor module_layers_3_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97088896)))]; tensor linear_35_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear1_bias_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16_quantized, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; tensor module_layers_3_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97097152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101292608))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101291520)))]; tensor module_layers_3_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101294720)))]; tensor linear_36_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear2_bias_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16_quantized, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; tensor var_901_to_fp16 = const()[name = tensor("op_901_to_fp16"), val = tensor(0x1p-1)]; tensor var_902_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_901_to_fp16)[name = tensor("op_902_cast_fp16")]; tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_902_cast_fp16)[name = tensor("input_225_cast_fp16")]; tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101296832)))]; tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101298944)))]; tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101301056)))]; tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101303168)))]; tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; tensor module_layers_4_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101305280))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105503808))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105499648)))]; tensor module_layers_4_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105512064)))]; tensor linear_37_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear1_bias_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16_quantized, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; tensor module_layers_4_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105520320))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109715776))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109714688)))]; tensor module_layers_4_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109717888)))]; tensor linear_38_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear2_bias_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16_quantized, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; tensor var_932_to_fp16 = const()[name = tensor("op_932_to_fp16"), val = tensor(0x1p-1)]; tensor var_933_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_932_to_fp16)[name = tensor("op_933_cast_fp16")]; tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_933_cast_fp16)[name = tensor("input_239_cast_fp16")]; tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109720000)))]; tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109722112)))]; tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; tensor module_layers_4_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109724224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110773952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110772864)))]; tensor module_layers_4_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110776064)))]; tensor linear_39_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_q_bias_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16_quantized, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; tensor var_950 = const()[name = tensor("op_950"), val = tensor([1, -1, 8, 128])]; tensor q_25_cast_fp16 = reshape(shape = var_950, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; tensor module_layers_4_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110778176))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111827904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111826816)))]; tensor module_layers_4_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111830016)))]; tensor linear_40_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_k_bias_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16_quantized, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; tensor var_955 = const()[name = tensor("op_955"), val = tensor([1, -1, 8, 128])]; tensor k_17_cast_fp16 = reshape(shape = var_955, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; tensor module_layers_4_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111832128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112881856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112880768)))]; tensor module_layers_4_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112883968)))]; tensor linear_41_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_v_bias_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16_quantized, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; tensor var_960 = const()[name = tensor("op_960"), val = tensor([1, -1, 8, 128])]; tensor v_9_cast_fp16 = reshape(shape = var_960, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112886080)))]; tensor var_972_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_972_cast_fp16")]; tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112888192)))]; tensor var_974_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_974_cast_fp16")]; tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_97_transpose_x_0 = const()[name = tensor("x_97_transpose_x_0"), val = tensor(false)]; tensor x_97_transpose_y_0 = const()[name = tensor("x_97_transpose_y_0"), val = tensor(false)]; tensor op_976_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_976_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112890304))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113274816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113274368)))]; tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_974_cast_fp16)[name = tensor("transpose_284")]; tensor x_97_cast_fp16 = matmul(transpose_x = x_97_transpose_x_0, transpose_y = x_97_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = op_976_to_fp16_quantized)[name = tensor("x_97_cast_fp16")]; tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_99_mode_0 = const()[name = tensor("x_99_mode_0"), val = tensor("constant")]; tensor const_54_to_fp16 = const()[name = tensor("const_54_to_fp16"), val = tensor(0x0p+0)]; tensor x_99_cast_fp16 = pad(constant_val = const_54_to_fp16, mode = x_99_mode_0, pad = x_99_pad_0, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; tensor var_984 = const()[name = tensor("op_984"), val = tensor([1, 8, -1, 188])]; tensor x_101_cast_fp16 = reshape(shape = var_984, x = x_99_cast_fp16)[name = tensor("x_101_cast_fp16")]; tensor var_988_begin_0 = const()[name = tensor("op_988_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_988_end_0 = const()[name = tensor("op_988_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_988_end_mask_0 = const()[name = tensor("op_988_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_988_cast_fp16 = slice_by_index(begin = var_988_begin_0, end = var_988_end_0, end_mask = var_988_end_mask_0, x = x_101_cast_fp16)[name = tensor("op_988_cast_fp16")]; tensor var_989 = const()[name = tensor("op_989"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_17_cast_fp16 = reshape(shape = var_989, x = var_988_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; tensor transpose_104_perm_0 = const()[name = tensor("transpose_104_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_105_perm_0 = const()[name = tensor("transpose_105_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_105 = transpose(perm = transpose_105_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_282")]; tensor transpose_104 = transpose(perm = transpose_104_perm_0, x = var_972_cast_fp16)[name = tensor("transpose_283")]; tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_104, y = transpose_105)[name = tensor("matrix_ac_9_cast_fp16")]; tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; tensor var_998_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_998_cast_fp16")]; tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_17_cast_fp16 = mul(x = var_998_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_3)[name = tensor("scores_19_cast_fp16")]; tensor var_1004_cast_fp16 = softmax(axis = var_30, x = scores_19_cast_fp16)[name = tensor("op_1004_cast_fp16")]; tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_1004_cast_fp16, cond = mask_3)[name = tensor("input_241_cast_fp16")]; tensor x_103_transpose_x_0 = const()[name = tensor("x_103_transpose_x_0"), val = tensor(false)]; tensor x_103_transpose_y_0 = const()[name = tensor("x_103_transpose_y_0"), val = tensor(false)]; tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_281")]; tensor x_103_cast_fp16 = matmul(transpose_x = x_103_transpose_x_0, transpose_y = x_103_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_103_cast_fp16")]; tensor var_1008_perm_0 = const()[name = tensor("op_1008_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1009 = const()[name = tensor("op_1009"), val = tensor([1, -1, 1024])]; tensor var_1008_cast_fp16 = transpose(perm = var_1008_perm_0, x = x_103_cast_fp16)[name = tensor("transpose_280")]; tensor input_243_cast_fp16 = reshape(shape = var_1009, x = var_1008_cast_fp16)[name = tensor("input_243_cast_fp16")]; tensor module_layers_4_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113275648))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114325376))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114324288)))]; tensor module_layers_4_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114327488)))]; tensor linear_43_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_out_bias_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16_quantized, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; tensor x_107_axes_0 = const()[name = tensor("x_107_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114329600)))]; tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114331712)))]; tensor x_107_cast_fp16 = layer_norm(axes = x_107_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_107_cast_fp16")]; tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114333824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116433152))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116431040)))]; tensor module_layers_4_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116437312)))]; tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_107_cast_fp16)[name = tensor("transpose_279")]; tensor input_251_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv1_bias_to_fp16, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; tensor x_109_split_num_splits_0 = const()[name = tensor("x_109_split_num_splits_0"), val = tensor(2)]; tensor x_109_split_axis_0 = const()[name = tensor("x_109_split_axis_0"), val = tensor(1)]; tensor x_109_split_cast_fp16_0, tensor x_109_split_cast_fp16_1 = split(axis = x_109_split_axis_0, num_splits = x_109_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_109_split_cast_fp16")]; tensor x_109_split_1_sigmoid_cast_fp16 = sigmoid(x = x_109_split_cast_fp16_1)[name = tensor("x_109_split_1_sigmoid_cast_fp16")]; tensor x_109_cast_fp16 = mul(x = x_109_split_cast_fp16_0, y = x_109_split_1_sigmoid_cast_fp16)[name = tensor("x_109_cast_fp16")]; tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_109_cast_fp16, cond = var_337)[name = tensor("input_253_cast_fp16")]; tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; tensor const_57_to_fp16 = const()[name = tensor("const_57_to_fp16"), val = tensor(0x0p+0)]; tensor input_255_cast_fp16 = pad(constant_val = const_57_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(1024)]; tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; tensor const_256_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_256_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116441472))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116451840))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116450752)))]; tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116453952)))]; tensor input_259_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_256_to_fp16_quantized, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; tensor x_111_pad_type_0 = const()[name = tensor("x_111_pad_type_0"), val = tensor("valid")]; tensor x_111_strides_0 = const()[name = tensor("x_111_strides_0"), val = tensor([1])]; tensor x_111_pad_0 = const()[name = tensor("x_111_pad_0"), val = tensor([0, 0])]; tensor x_111_dilations_0 = const()[name = tensor("x_111_dilations_0"), val = tensor([1])]; tensor x_111_groups_0 = const()[name = tensor("x_111_groups_0"), val = tensor(1)]; tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116456064))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117505792))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117504704)))]; tensor module_layers_4_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117507904)))]; tensor x_111_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv2_bias_to_fp16, dilations = x_111_dilations_0, groups = x_111_groups_0, pad = x_111_pad_0, pad_type = x_111_pad_type_0, strides = x_111_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_261_cast_fp16)[name = tensor("x_111_cast_fp16")]; tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_111_cast_fp16)[name = tensor("transpose_278")]; tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117510016)))]; tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117512128)))]; tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; tensor module_layers_4_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117514240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121712768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121708608)))]; tensor module_layers_4_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121721024)))]; tensor linear_44_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear1_bias_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16_quantized, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; tensor module_layers_4_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121729280))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125924736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125923648)))]; tensor module_layers_4_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125926848)))]; tensor linear_45_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear2_bias_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16_quantized, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; tensor var_1075_to_fp16 = const()[name = tensor("op_1075_to_fp16"), val = tensor(0x1p-1)]; tensor var_1076_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1075_to_fp16)[name = tensor("op_1076_cast_fp16")]; tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1076_cast_fp16)[name = tensor("input_277_cast_fp16")]; tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125928960)))]; tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125931072)))]; tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125933184)))]; tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125935296)))]; tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; tensor module_layers_5_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125937408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130135936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130131776)))]; tensor module_layers_5_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130144192)))]; tensor linear_46_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear1_bias_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16_quantized, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; tensor module_layers_5_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130152448))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134347904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134346816)))]; tensor module_layers_5_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134350016)))]; tensor linear_47_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear2_bias_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16_quantized, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; tensor var_1106_to_fp16 = const()[name = tensor("op_1106_to_fp16"), val = tensor(0x1p-1)]; tensor var_1107_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1106_to_fp16)[name = tensor("op_1107_cast_fp16")]; tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1107_cast_fp16)[name = tensor("input_291_cast_fp16")]; tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134352128)))]; tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134354240)))]; tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; tensor module_layers_5_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134356352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135406080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135404992)))]; tensor module_layers_5_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135408192)))]; tensor linear_48_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_q_bias_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16_quantized, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; tensor var_1124 = const()[name = tensor("op_1124"), val = tensor([1, -1, 8, 128])]; tensor q_31_cast_fp16 = reshape(shape = var_1124, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; tensor module_layers_5_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135410304))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136460032))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136458944)))]; tensor module_layers_5_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136462144)))]; tensor linear_49_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_k_bias_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16_quantized, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; tensor var_1129 = const()[name = tensor("op_1129"), val = tensor([1, -1, 8, 128])]; tensor k_21_cast_fp16 = reshape(shape = var_1129, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; tensor module_layers_5_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136464256))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137513984))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137512896)))]; tensor module_layers_5_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137516096)))]; tensor linear_50_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_v_bias_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16_quantized, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; tensor var_1134 = const()[name = tensor("op_1134"), val = tensor([1, -1, 8, 128])]; tensor v_11_cast_fp16 = reshape(shape = var_1134, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137518208)))]; tensor var_1146_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1146_cast_fp16")]; tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137520320)))]; tensor var_1148_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1148_cast_fp16")]; tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_119_transpose_x_0 = const()[name = tensor("x_119_transpose_x_0"), val = tensor(false)]; tensor x_119_transpose_y_0 = const()[name = tensor("x_119_transpose_y_0"), val = tensor(false)]; tensor op_1150_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1150_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137522432))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137906944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137906496)))]; tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1148_cast_fp16)[name = tensor("transpose_277")]; tensor x_119_cast_fp16 = matmul(transpose_x = x_119_transpose_x_0, transpose_y = x_119_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = op_1150_to_fp16_quantized)[name = tensor("x_119_cast_fp16")]; tensor x_121_pad_0 = const()[name = tensor("x_121_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_121_mode_0 = const()[name = tensor("x_121_mode_0"), val = tensor("constant")]; tensor const_64_to_fp16 = const()[name = tensor("const_64_to_fp16"), val = tensor(0x0p+0)]; tensor x_121_cast_fp16 = pad(constant_val = const_64_to_fp16, mode = x_121_mode_0, pad = x_121_pad_0, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; tensor var_1158 = const()[name = tensor("op_1158"), val = tensor([1, 8, -1, 188])]; tensor x_123_cast_fp16 = reshape(shape = var_1158, x = x_121_cast_fp16)[name = tensor("x_123_cast_fp16")]; tensor var_1162_begin_0 = const()[name = tensor("op_1162_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1162_end_0 = const()[name = tensor("op_1162_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1162_end_mask_0 = const()[name = tensor("op_1162_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1162_cast_fp16 = slice_by_index(begin = var_1162_begin_0, end = var_1162_end_0, end_mask = var_1162_end_mask_0, x = x_123_cast_fp16)[name = tensor("op_1162_cast_fp16")]; tensor var_1163 = const()[name = tensor("op_1163"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1163, x = var_1162_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; tensor transpose_106_perm_0 = const()[name = tensor("transpose_106_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_107_perm_0 = const()[name = tensor("transpose_107_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_107 = transpose(perm = transpose_107_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_275")]; tensor transpose_106 = transpose(perm = transpose_106_perm_0, x = var_1146_cast_fp16)[name = tensor("transpose_276")]; tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_106, y = transpose_107)[name = tensor("matrix_ac_11_cast_fp16")]; tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; tensor var_1172_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1172_cast_fp16")]; tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_21_cast_fp16 = mul(x = var_1172_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_3)[name = tensor("scores_23_cast_fp16")]; tensor var_1178_cast_fp16 = softmax(axis = var_30, x = scores_23_cast_fp16)[name = tensor("op_1178_cast_fp16")]; tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1178_cast_fp16, cond = mask_3)[name = tensor("input_293_cast_fp16")]; tensor x_125_transpose_x_0 = const()[name = tensor("x_125_transpose_x_0"), val = tensor(false)]; tensor x_125_transpose_y_0 = const()[name = tensor("x_125_transpose_y_0"), val = tensor(false)]; tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_274")]; tensor x_125_cast_fp16 = matmul(transpose_x = x_125_transpose_x_0, transpose_y = x_125_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_125_cast_fp16")]; tensor var_1182_perm_0 = const()[name = tensor("op_1182_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1183 = const()[name = tensor("op_1183"), val = tensor([1, -1, 1024])]; tensor var_1182_cast_fp16 = transpose(perm = var_1182_perm_0, x = x_125_cast_fp16)[name = tensor("transpose_273")]; tensor input_295_cast_fp16 = reshape(shape = var_1183, x = var_1182_cast_fp16)[name = tensor("input_295_cast_fp16")]; tensor module_layers_5_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137907776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138957504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138956416)))]; tensor module_layers_5_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138959616)))]; tensor linear_52_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_out_bias_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16_quantized, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; tensor x_129_axes_0 = const()[name = tensor("x_129_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138961728)))]; tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138963840)))]; tensor x_129_cast_fp16 = layer_norm(axes = x_129_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_129_cast_fp16")]; tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138965952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141065280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141063168)))]; tensor module_layers_5_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141069440)))]; tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_129_cast_fp16)[name = tensor("transpose_272")]; tensor input_303_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv1_bias_to_fp16, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; tensor x_131_split_num_splits_0 = const()[name = tensor("x_131_split_num_splits_0"), val = tensor(2)]; tensor x_131_split_axis_0 = const()[name = tensor("x_131_split_axis_0"), val = tensor(1)]; tensor x_131_split_cast_fp16_0, tensor x_131_split_cast_fp16_1 = split(axis = x_131_split_axis_0, num_splits = x_131_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_131_split_cast_fp16")]; tensor x_131_split_1_sigmoid_cast_fp16 = sigmoid(x = x_131_split_cast_fp16_1)[name = tensor("x_131_split_1_sigmoid_cast_fp16")]; tensor x_131_cast_fp16 = mul(x = x_131_split_cast_fp16_0, y = x_131_split_1_sigmoid_cast_fp16)[name = tensor("x_131_cast_fp16")]; tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_131_cast_fp16, cond = var_337)[name = tensor("input_305_cast_fp16")]; tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; tensor const_67_to_fp16 = const()[name = tensor("const_67_to_fp16"), val = tensor(0x0p+0)]; tensor input_307_cast_fp16 = pad(constant_val = const_67_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(1024)]; tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; tensor const_258_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_258_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141073600))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141083968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141082880)))]; tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141086080)))]; tensor input_311_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_258_to_fp16_quantized, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; tensor x_133_pad_type_0 = const()[name = tensor("x_133_pad_type_0"), val = tensor("valid")]; tensor x_133_strides_0 = const()[name = tensor("x_133_strides_0"), val = tensor([1])]; tensor x_133_pad_0 = const()[name = tensor("x_133_pad_0"), val = tensor([0, 0])]; tensor x_133_dilations_0 = const()[name = tensor("x_133_dilations_0"), val = tensor([1])]; tensor x_133_groups_0 = const()[name = tensor("x_133_groups_0"), val = tensor(1)]; tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141088192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142137920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142136832)))]; tensor module_layers_5_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142140032)))]; tensor x_133_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv2_bias_to_fp16, dilations = x_133_dilations_0, groups = x_133_groups_0, pad = x_133_pad_0, pad_type = x_133_pad_type_0, strides = x_133_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_313_cast_fp16)[name = tensor("x_133_cast_fp16")]; tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_133_cast_fp16)[name = tensor("transpose_271")]; tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142142144)))]; tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142144256)))]; tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; tensor module_layers_5_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142146368))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146344896))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146340736)))]; tensor module_layers_5_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146353152)))]; tensor linear_53_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear1_bias_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16_quantized, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; tensor module_layers_5_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146361408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150556864))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150555776)))]; tensor module_layers_5_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150558976)))]; tensor linear_54_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear2_bias_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16_quantized, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; tensor var_1249_to_fp16 = const()[name = tensor("op_1249_to_fp16"), val = tensor(0x1p-1)]; tensor var_1250_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1249_to_fp16)[name = tensor("op_1250_cast_fp16")]; tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1250_cast_fp16)[name = tensor("input_329_cast_fp16")]; tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150561088)))]; tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150563200)))]; tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150565312)))]; tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150567424)))]; tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; tensor module_layers_6_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150569536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154768064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154763904)))]; tensor module_layers_6_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154776320)))]; tensor linear_55_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear1_bias_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16_quantized, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; tensor module_layers_6_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154784576))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158980032))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158978944)))]; tensor module_layers_6_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158982144)))]; tensor linear_56_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear2_bias_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16_quantized, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; tensor var_1280_to_fp16 = const()[name = tensor("op_1280_to_fp16"), val = tensor(0x1p-1)]; tensor var_1281_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1280_to_fp16)[name = tensor("op_1281_cast_fp16")]; tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1281_cast_fp16)[name = tensor("input_343_cast_fp16")]; tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158984256)))]; tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158986368)))]; tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; tensor module_layers_6_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158988480))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160038208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160037120)))]; tensor module_layers_6_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160040320)))]; tensor linear_57_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_q_bias_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16_quantized, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; tensor var_1298 = const()[name = tensor("op_1298"), val = tensor([1, -1, 8, 128])]; tensor q_37_cast_fp16 = reshape(shape = var_1298, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; tensor module_layers_6_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160042432))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161092160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161091072)))]; tensor module_layers_6_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161094272)))]; tensor linear_58_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_k_bias_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16_quantized, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; tensor var_1303 = const()[name = tensor("op_1303"), val = tensor([1, -1, 8, 128])]; tensor k_25_cast_fp16 = reshape(shape = var_1303, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; tensor module_layers_6_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161096384))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162146112))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162145024)))]; tensor module_layers_6_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162148224)))]; tensor linear_59_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_v_bias_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16_quantized, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; tensor var_1308 = const()[name = tensor("op_1308"), val = tensor([1, -1, 8, 128])]; tensor v_13_cast_fp16 = reshape(shape = var_1308, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162150336)))]; tensor var_1320_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1320_cast_fp16")]; tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162152448)))]; tensor var_1322_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1322_cast_fp16")]; tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_141_transpose_x_0 = const()[name = tensor("x_141_transpose_x_0"), val = tensor(false)]; tensor x_141_transpose_y_0 = const()[name = tensor("x_141_transpose_y_0"), val = tensor(false)]; tensor op_1324_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1324_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162154560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162539072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162538624)))]; tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1322_cast_fp16)[name = tensor("transpose_270")]; tensor x_141_cast_fp16 = matmul(transpose_x = x_141_transpose_x_0, transpose_y = x_141_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = op_1324_to_fp16_quantized)[name = tensor("x_141_cast_fp16")]; tensor x_143_pad_0 = const()[name = tensor("x_143_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_143_mode_0 = const()[name = tensor("x_143_mode_0"), val = tensor("constant")]; tensor const_74_to_fp16 = const()[name = tensor("const_74_to_fp16"), val = tensor(0x0p+0)]; tensor x_143_cast_fp16 = pad(constant_val = const_74_to_fp16, mode = x_143_mode_0, pad = x_143_pad_0, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; tensor var_1332 = const()[name = tensor("op_1332"), val = tensor([1, 8, -1, 188])]; tensor x_145_cast_fp16 = reshape(shape = var_1332, x = x_143_cast_fp16)[name = tensor("x_145_cast_fp16")]; tensor var_1336_begin_0 = const()[name = tensor("op_1336_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1336_end_0 = const()[name = tensor("op_1336_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1336_end_mask_0 = const()[name = tensor("op_1336_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1336_cast_fp16 = slice_by_index(begin = var_1336_begin_0, end = var_1336_end_0, end_mask = var_1336_end_mask_0, x = x_145_cast_fp16)[name = tensor("op_1336_cast_fp16")]; tensor var_1337 = const()[name = tensor("op_1337"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1337, x = var_1336_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; tensor transpose_108_perm_0 = const()[name = tensor("transpose_108_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_109_perm_0 = const()[name = tensor("transpose_109_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_109 = transpose(perm = transpose_109_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_268")]; tensor transpose_108 = transpose(perm = transpose_108_perm_0, x = var_1320_cast_fp16)[name = tensor("transpose_269")]; tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_108, y = transpose_109)[name = tensor("matrix_ac_13_cast_fp16")]; tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; tensor var_1346_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1346_cast_fp16")]; tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_25_cast_fp16 = mul(x = var_1346_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_3)[name = tensor("scores_27_cast_fp16")]; tensor var_1352_cast_fp16 = softmax(axis = var_30, x = scores_27_cast_fp16)[name = tensor("op_1352_cast_fp16")]; tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1352_cast_fp16, cond = mask_3)[name = tensor("input_345_cast_fp16")]; tensor x_147_transpose_x_0 = const()[name = tensor("x_147_transpose_x_0"), val = tensor(false)]; tensor x_147_transpose_y_0 = const()[name = tensor("x_147_transpose_y_0"), val = tensor(false)]; tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_267")]; tensor x_147_cast_fp16 = matmul(transpose_x = x_147_transpose_x_0, transpose_y = x_147_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_147_cast_fp16")]; tensor var_1356_perm_0 = const()[name = tensor("op_1356_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1357 = const()[name = tensor("op_1357"), val = tensor([1, -1, 1024])]; tensor var_1356_cast_fp16 = transpose(perm = var_1356_perm_0, x = x_147_cast_fp16)[name = tensor("transpose_266")]; tensor input_347_cast_fp16 = reshape(shape = var_1357, x = var_1356_cast_fp16)[name = tensor("input_347_cast_fp16")]; tensor module_layers_6_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162539904))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163589632))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163588544)))]; tensor module_layers_6_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163591744)))]; tensor linear_61_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_out_bias_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16_quantized, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; tensor x_151_axes_0 = const()[name = tensor("x_151_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163593856)))]; tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163595968)))]; tensor x_151_cast_fp16 = layer_norm(axes = x_151_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_151_cast_fp16")]; tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163598080))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165697408))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165695296)))]; tensor module_layers_6_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165701568)))]; tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_151_cast_fp16)[name = tensor("transpose_265")]; tensor input_355_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv1_bias_to_fp16, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; tensor x_153_split_num_splits_0 = const()[name = tensor("x_153_split_num_splits_0"), val = tensor(2)]; tensor x_153_split_axis_0 = const()[name = tensor("x_153_split_axis_0"), val = tensor(1)]; tensor x_153_split_cast_fp16_0, tensor x_153_split_cast_fp16_1 = split(axis = x_153_split_axis_0, num_splits = x_153_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_153_split_cast_fp16")]; tensor x_153_split_1_sigmoid_cast_fp16 = sigmoid(x = x_153_split_cast_fp16_1)[name = tensor("x_153_split_1_sigmoid_cast_fp16")]; tensor x_153_cast_fp16 = mul(x = x_153_split_cast_fp16_0, y = x_153_split_1_sigmoid_cast_fp16)[name = tensor("x_153_cast_fp16")]; tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_153_cast_fp16, cond = var_337)[name = tensor("input_357_cast_fp16")]; tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; tensor const_77_to_fp16 = const()[name = tensor("const_77_to_fp16"), val = tensor(0x0p+0)]; tensor input_359_cast_fp16 = pad(constant_val = const_77_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(1024)]; tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; tensor const_260_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_260_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165705728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165716096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165715008)))]; tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165718208)))]; tensor input_363_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_260_to_fp16_quantized, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165720320))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166770048))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166768960)))]; tensor module_layers_6_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166772160)))]; tensor x_155_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv2_bias_to_fp16, dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_365_cast_fp16)[name = tensor("x_155_cast_fp16")]; tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_155_cast_fp16)[name = tensor("transpose_264")]; tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166774272)))]; tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166776384)))]; tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; tensor module_layers_6_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166778496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170977024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170972864)))]; tensor module_layers_6_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170985280)))]; tensor linear_62_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear1_bias_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16_quantized, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; tensor module_layers_6_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170993536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175188992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175187904)))]; tensor module_layers_6_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175191104)))]; tensor linear_63_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear2_bias_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16_quantized, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; tensor var_1423_to_fp16 = const()[name = tensor("op_1423_to_fp16"), val = tensor(0x1p-1)]; tensor var_1424_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1423_to_fp16)[name = tensor("op_1424_cast_fp16")]; tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1424_cast_fp16)[name = tensor("input_381_cast_fp16")]; tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175193216)))]; tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175195328)))]; tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175197440)))]; tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175199552)))]; tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; tensor module_layers_7_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175201664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179400192))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179396032)))]; tensor module_layers_7_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179408448)))]; tensor linear_64_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear1_bias_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16_quantized, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; tensor module_layers_7_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179416704))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183612160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183611072)))]; tensor module_layers_7_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183614272)))]; tensor linear_65_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear2_bias_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16_quantized, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; tensor var_1454_to_fp16 = const()[name = tensor("op_1454_to_fp16"), val = tensor(0x1p-1)]; tensor var_1455_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1454_to_fp16)[name = tensor("op_1455_cast_fp16")]; tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1455_cast_fp16)[name = tensor("input_395_cast_fp16")]; tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183616384)))]; tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183618496)))]; tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; tensor module_layers_7_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183620608))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184670336))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184669248)))]; tensor module_layers_7_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184672448)))]; tensor linear_66_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_q_bias_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16_quantized, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; tensor var_1472 = const()[name = tensor("op_1472"), val = tensor([1, -1, 8, 128])]; tensor q_43_cast_fp16 = reshape(shape = var_1472, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; tensor module_layers_7_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184674560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185724288))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185723200)))]; tensor module_layers_7_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185726400)))]; tensor linear_67_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_k_bias_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16_quantized, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; tensor var_1477 = const()[name = tensor("op_1477"), val = tensor([1, -1, 8, 128])]; tensor k_29_cast_fp16 = reshape(shape = var_1477, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; tensor module_layers_7_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185728512))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186778240))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186777152)))]; tensor module_layers_7_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186780352)))]; tensor linear_68_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_v_bias_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16_quantized, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; tensor var_1482 = const()[name = tensor("op_1482"), val = tensor([1, -1, 8, 128])]; tensor v_15_cast_fp16 = reshape(shape = var_1482, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186782464)))]; tensor var_1494_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1494_cast_fp16")]; tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186784576)))]; tensor var_1496_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1496_cast_fp16")]; tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; tensor op_1498_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1498_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186786688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187171200))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187170752)))]; tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1496_cast_fp16)[name = tensor("transpose_263")]; tensor x_163_cast_fp16 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = op_1498_to_fp16_quantized)[name = tensor("x_163_cast_fp16")]; tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; tensor const_84_to_fp16 = const()[name = tensor("const_84_to_fp16"), val = tensor(0x0p+0)]; tensor x_165_cast_fp16 = pad(constant_val = const_84_to_fp16, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; tensor var_1506 = const()[name = tensor("op_1506"), val = tensor([1, 8, -1, 188])]; tensor x_167_cast_fp16 = reshape(shape = var_1506, x = x_165_cast_fp16)[name = tensor("x_167_cast_fp16")]; tensor var_1510_begin_0 = const()[name = tensor("op_1510_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1510_end_0 = const()[name = tensor("op_1510_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1510_end_mask_0 = const()[name = tensor("op_1510_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1510_cast_fp16 = slice_by_index(begin = var_1510_begin_0, end = var_1510_end_0, end_mask = var_1510_end_mask_0, x = x_167_cast_fp16)[name = tensor("op_1510_cast_fp16")]; tensor var_1511 = const()[name = tensor("op_1511"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1511, x = var_1510_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; tensor transpose_110_perm_0 = const()[name = tensor("transpose_110_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_111_perm_0 = const()[name = tensor("transpose_111_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_111 = transpose(perm = transpose_111_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_261")]; tensor transpose_110 = transpose(perm = transpose_110_perm_0, x = var_1494_cast_fp16)[name = tensor("transpose_262")]; tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_110, y = transpose_111)[name = tensor("matrix_ac_15_cast_fp16")]; tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; tensor var_1520_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1520_cast_fp16")]; tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_29_cast_fp16 = mul(x = var_1520_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_3)[name = tensor("scores_31_cast_fp16")]; tensor var_1526_cast_fp16 = softmax(axis = var_30, x = scores_31_cast_fp16)[name = tensor("op_1526_cast_fp16")]; tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1526_cast_fp16, cond = mask_3)[name = tensor("input_397_cast_fp16")]; tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_260")]; tensor x_169_cast_fp16 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_169_cast_fp16")]; tensor var_1530_perm_0 = const()[name = tensor("op_1530_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1531 = const()[name = tensor("op_1531"), val = tensor([1, -1, 1024])]; tensor var_1530_cast_fp16 = transpose(perm = var_1530_perm_0, x = x_169_cast_fp16)[name = tensor("transpose_259")]; tensor input_399_cast_fp16 = reshape(shape = var_1531, x = var_1530_cast_fp16)[name = tensor("input_399_cast_fp16")]; tensor module_layers_7_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187172032))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188221760))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188220672)))]; tensor module_layers_7_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188223872)))]; tensor linear_70_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_out_bias_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16_quantized, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188225984)))]; tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188228096)))]; tensor x_173_cast_fp16 = layer_norm(axes = x_173_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_173_cast_fp16")]; tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188230208))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190329536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190327424)))]; tensor module_layers_7_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190333696)))]; tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_173_cast_fp16)[name = tensor("transpose_258")]; tensor input_407_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv1_bias_to_fp16, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; tensor x_175_split_cast_fp16_0, tensor x_175_split_cast_fp16_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_175_split_cast_fp16")]; tensor x_175_split_1_sigmoid_cast_fp16 = sigmoid(x = x_175_split_cast_fp16_1)[name = tensor("x_175_split_1_sigmoid_cast_fp16")]; tensor x_175_cast_fp16 = mul(x = x_175_split_cast_fp16_0, y = x_175_split_1_sigmoid_cast_fp16)[name = tensor("x_175_cast_fp16")]; tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_175_cast_fp16, cond = var_337)[name = tensor("input_409_cast_fp16")]; tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; tensor const_87_to_fp16 = const()[name = tensor("const_87_to_fp16"), val = tensor(0x0p+0)]; tensor input_411_cast_fp16 = pad(constant_val = const_87_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(1024)]; tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; tensor const_262_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_262_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190337856))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190348224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190347136)))]; tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190350336)))]; tensor input_415_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_262_to_fp16_quantized, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(1)]; tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190352448))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191402176))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191401088)))]; tensor module_layers_7_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191404288)))]; tensor x_177_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv2_bias_to_fp16, dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_417_cast_fp16)[name = tensor("x_177_cast_fp16")]; tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_177_cast_fp16)[name = tensor("transpose_257")]; tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191406400)))]; tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191408512)))]; tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; tensor module_layers_7_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191410624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195609152))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195604992)))]; tensor module_layers_7_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195617408)))]; tensor linear_71_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear1_bias_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16_quantized, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; tensor module_layers_7_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195625664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199821120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199820032)))]; tensor module_layers_7_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199823232)))]; tensor linear_72_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear2_bias_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16_quantized, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; tensor var_1597_to_fp16 = const()[name = tensor("op_1597_to_fp16"), val = tensor(0x1p-1)]; tensor var_1598_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1597_to_fp16)[name = tensor("op_1598_cast_fp16")]; tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1598_cast_fp16)[name = tensor("input_433_cast_fp16")]; tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199825344)))]; tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199827456)))]; tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199829568)))]; tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199831680)))]; tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; tensor module_layers_8_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199833792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204032320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204028160)))]; tensor module_layers_8_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204040576)))]; tensor linear_73_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear1_bias_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16_quantized, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; tensor module_layers_8_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204048832))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208244288))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208243200)))]; tensor module_layers_8_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208246400)))]; tensor linear_74_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear2_bias_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16_quantized, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; tensor var_1628_to_fp16 = const()[name = tensor("op_1628_to_fp16"), val = tensor(0x1p-1)]; tensor var_1629_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1628_to_fp16)[name = tensor("op_1629_cast_fp16")]; tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1629_cast_fp16)[name = tensor("input_447_cast_fp16")]; tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208248512)))]; tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208250624)))]; tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; tensor module_layers_8_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208252736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209302464))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209301376)))]; tensor module_layers_8_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209304576)))]; tensor linear_75_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_q_bias_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16_quantized, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; tensor var_1646 = const()[name = tensor("op_1646"), val = tensor([1, -1, 8, 128])]; tensor q_49_cast_fp16 = reshape(shape = var_1646, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; tensor module_layers_8_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209306688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210356416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210355328)))]; tensor module_layers_8_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210358528)))]; tensor linear_76_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_k_bias_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16_quantized, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; tensor var_1651 = const()[name = tensor("op_1651"), val = tensor([1, -1, 8, 128])]; tensor k_33_cast_fp16 = reshape(shape = var_1651, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; tensor module_layers_8_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210360640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211410368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211409280)))]; tensor module_layers_8_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211412480)))]; tensor linear_77_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_v_bias_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16_quantized, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; tensor var_1656 = const()[name = tensor("op_1656"), val = tensor([1, -1, 8, 128])]; tensor v_17_cast_fp16 = reshape(shape = var_1656, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211414592)))]; tensor var_1668_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1668_cast_fp16")]; tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211416704)))]; tensor var_1670_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1670_cast_fp16")]; tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_185_transpose_x_0 = const()[name = tensor("x_185_transpose_x_0"), val = tensor(false)]; tensor x_185_transpose_y_0 = const()[name = tensor("x_185_transpose_y_0"), val = tensor(false)]; tensor op_1672_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1672_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211418816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211803328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211802880)))]; tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1670_cast_fp16)[name = tensor("transpose_256")]; tensor x_185_cast_fp16 = matmul(transpose_x = x_185_transpose_x_0, transpose_y = x_185_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = op_1672_to_fp16_quantized)[name = tensor("x_185_cast_fp16")]; tensor x_187_pad_0 = const()[name = tensor("x_187_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_187_mode_0 = const()[name = tensor("x_187_mode_0"), val = tensor("constant")]; tensor const_94_to_fp16 = const()[name = tensor("const_94_to_fp16"), val = tensor(0x0p+0)]; tensor x_187_cast_fp16 = pad(constant_val = const_94_to_fp16, mode = x_187_mode_0, pad = x_187_pad_0, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; tensor var_1680 = const()[name = tensor("op_1680"), val = tensor([1, 8, -1, 188])]; tensor x_189_cast_fp16 = reshape(shape = var_1680, x = x_187_cast_fp16)[name = tensor("x_189_cast_fp16")]; tensor var_1684_begin_0 = const()[name = tensor("op_1684_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1684_end_0 = const()[name = tensor("op_1684_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1684_end_mask_0 = const()[name = tensor("op_1684_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1684_cast_fp16 = slice_by_index(begin = var_1684_begin_0, end = var_1684_end_0, end_mask = var_1684_end_mask_0, x = x_189_cast_fp16)[name = tensor("op_1684_cast_fp16")]; tensor var_1685 = const()[name = tensor("op_1685"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1685, x = var_1684_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; tensor transpose_112_perm_0 = const()[name = tensor("transpose_112_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_113_perm_0 = const()[name = tensor("transpose_113_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_113 = transpose(perm = transpose_113_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_254")]; tensor transpose_112 = transpose(perm = transpose_112_perm_0, x = var_1668_cast_fp16)[name = tensor("transpose_255")]; tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_112, y = transpose_113)[name = tensor("matrix_ac_17_cast_fp16")]; tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; tensor var_1694_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1694_cast_fp16")]; tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_33_cast_fp16 = mul(x = var_1694_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_3)[name = tensor("scores_35_cast_fp16")]; tensor var_1700_cast_fp16 = softmax(axis = var_30, x = scores_35_cast_fp16)[name = tensor("op_1700_cast_fp16")]; tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1700_cast_fp16, cond = mask_3)[name = tensor("input_449_cast_fp16")]; tensor x_191_transpose_x_0 = const()[name = tensor("x_191_transpose_x_0"), val = tensor(false)]; tensor x_191_transpose_y_0 = const()[name = tensor("x_191_transpose_y_0"), val = tensor(false)]; tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_253")]; tensor x_191_cast_fp16 = matmul(transpose_x = x_191_transpose_x_0, transpose_y = x_191_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_191_cast_fp16")]; tensor var_1704_perm_0 = const()[name = tensor("op_1704_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1705 = const()[name = tensor("op_1705"), val = tensor([1, -1, 1024])]; tensor var_1704_cast_fp16 = transpose(perm = var_1704_perm_0, x = x_191_cast_fp16)[name = tensor("transpose_252")]; tensor input_451_cast_fp16 = reshape(shape = var_1705, x = var_1704_cast_fp16)[name = tensor("input_451_cast_fp16")]; tensor module_layers_8_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211804160))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212853888))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212852800)))]; tensor module_layers_8_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212856000)))]; tensor linear_79_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_out_bias_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16_quantized, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; tensor x_195_axes_0 = const()[name = tensor("x_195_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212858112)))]; tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212860224)))]; tensor x_195_cast_fp16 = layer_norm(axes = x_195_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_195_cast_fp16")]; tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212862336))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214961664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214959552)))]; tensor module_layers_8_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214965824)))]; tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_195_cast_fp16)[name = tensor("transpose_251")]; tensor input_459_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv1_bias_to_fp16, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; tensor x_197_split_num_splits_0 = const()[name = tensor("x_197_split_num_splits_0"), val = tensor(2)]; tensor x_197_split_axis_0 = const()[name = tensor("x_197_split_axis_0"), val = tensor(1)]; tensor x_197_split_cast_fp16_0, tensor x_197_split_cast_fp16_1 = split(axis = x_197_split_axis_0, num_splits = x_197_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_197_split_cast_fp16")]; tensor x_197_split_1_sigmoid_cast_fp16 = sigmoid(x = x_197_split_cast_fp16_1)[name = tensor("x_197_split_1_sigmoid_cast_fp16")]; tensor x_197_cast_fp16 = mul(x = x_197_split_cast_fp16_0, y = x_197_split_1_sigmoid_cast_fp16)[name = tensor("x_197_cast_fp16")]; tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_197_cast_fp16, cond = var_337)[name = tensor("input_461_cast_fp16")]; tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; tensor const_97_to_fp16 = const()[name = tensor("const_97_to_fp16"), val = tensor(0x0p+0)]; tensor input_463_cast_fp16 = pad(constant_val = const_97_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(1024)]; tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; tensor const_264_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_264_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214969984))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214980352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214979264)))]; tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214982464)))]; tensor input_467_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_264_to_fp16_quantized, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; tensor x_199_pad_type_0 = const()[name = tensor("x_199_pad_type_0"), val = tensor("valid")]; tensor x_199_strides_0 = const()[name = tensor("x_199_strides_0"), val = tensor([1])]; tensor x_199_pad_0 = const()[name = tensor("x_199_pad_0"), val = tensor([0, 0])]; tensor x_199_dilations_0 = const()[name = tensor("x_199_dilations_0"), val = tensor([1])]; tensor x_199_groups_0 = const()[name = tensor("x_199_groups_0"), val = tensor(1)]; tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214984576))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216034304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216033216)))]; tensor module_layers_8_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216036416)))]; tensor x_199_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv2_bias_to_fp16, dilations = x_199_dilations_0, groups = x_199_groups_0, pad = x_199_pad_0, pad_type = x_199_pad_type_0, strides = x_199_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_469_cast_fp16)[name = tensor("x_199_cast_fp16")]; tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_199_cast_fp16)[name = tensor("transpose_250")]; tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216038528)))]; tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216040640)))]; tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; tensor module_layers_8_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216042752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220241280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220237120)))]; tensor module_layers_8_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220249536)))]; tensor linear_80_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear1_bias_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16_quantized, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; tensor module_layers_8_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220257792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224453248))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224452160)))]; tensor module_layers_8_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224455360)))]; tensor linear_81_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear2_bias_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16_quantized, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; tensor var_1771_to_fp16 = const()[name = tensor("op_1771_to_fp16"), val = tensor(0x1p-1)]; tensor var_1772_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1771_to_fp16)[name = tensor("op_1772_cast_fp16")]; tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1772_cast_fp16)[name = tensor("input_485_cast_fp16")]; tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224457472)))]; tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224459584)))]; tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224461696)))]; tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224463808)))]; tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; tensor module_layers_9_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224465920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(228664448))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(228660288)))]; tensor module_layers_9_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(228672704)))]; tensor linear_82_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear1_bias_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16_quantized, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; tensor module_layers_9_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(228680960))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232876416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232875328)))]; tensor module_layers_9_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232878528)))]; tensor linear_83_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear2_bias_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16_quantized, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; tensor var_1802_to_fp16 = const()[name = tensor("op_1802_to_fp16"), val = tensor(0x1p-1)]; tensor var_1803_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1802_to_fp16)[name = tensor("op_1803_cast_fp16")]; tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1803_cast_fp16)[name = tensor("input_499_cast_fp16")]; tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232880640)))]; tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232882752)))]; tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; tensor module_layers_9_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232884864))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233934592))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233933504)))]; tensor module_layers_9_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233936704)))]; tensor linear_84_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_q_bias_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16_quantized, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; tensor var_1820 = const()[name = tensor("op_1820"), val = tensor([1, -1, 8, 128])]; tensor q_55_cast_fp16 = reshape(shape = var_1820, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; tensor module_layers_9_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233938816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234988544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234987456)))]; tensor module_layers_9_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234990656)))]; tensor linear_85_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_k_bias_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16_quantized, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; tensor var_1825 = const()[name = tensor("op_1825"), val = tensor([1, -1, 8, 128])]; tensor k_37_cast_fp16 = reshape(shape = var_1825, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; tensor module_layers_9_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234992768))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236042496))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236041408)))]; tensor module_layers_9_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236044608)))]; tensor linear_86_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_v_bias_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16_quantized, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; tensor var_1830 = const()[name = tensor("op_1830"), val = tensor([1, -1, 8, 128])]; tensor v_19_cast_fp16 = reshape(shape = var_1830, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236046720)))]; tensor var_1842_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1842_cast_fp16")]; tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236048832)))]; tensor var_1844_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1844_cast_fp16")]; tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_207_transpose_x_0 = const()[name = tensor("x_207_transpose_x_0"), val = tensor(false)]; tensor x_207_transpose_y_0 = const()[name = tensor("x_207_transpose_y_0"), val = tensor(false)]; tensor op_1846_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1846_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236050944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236435456))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236435008)))]; tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1844_cast_fp16)[name = tensor("transpose_249")]; tensor x_207_cast_fp16 = matmul(transpose_x = x_207_transpose_x_0, transpose_y = x_207_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = op_1846_to_fp16_quantized)[name = tensor("x_207_cast_fp16")]; tensor x_209_pad_0 = const()[name = tensor("x_209_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_209_mode_0 = const()[name = tensor("x_209_mode_0"), val = tensor("constant")]; tensor const_104_to_fp16 = const()[name = tensor("const_104_to_fp16"), val = tensor(0x0p+0)]; tensor x_209_cast_fp16 = pad(constant_val = const_104_to_fp16, mode = x_209_mode_0, pad = x_209_pad_0, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; tensor var_1854 = const()[name = tensor("op_1854"), val = tensor([1, 8, -1, 188])]; tensor x_211_cast_fp16 = reshape(shape = var_1854, x = x_209_cast_fp16)[name = tensor("x_211_cast_fp16")]; tensor var_1858_begin_0 = const()[name = tensor("op_1858_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1858_end_0 = const()[name = tensor("op_1858_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1858_end_mask_0 = const()[name = tensor("op_1858_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1858_cast_fp16 = slice_by_index(begin = var_1858_begin_0, end = var_1858_end_0, end_mask = var_1858_end_mask_0, x = x_211_cast_fp16)[name = tensor("op_1858_cast_fp16")]; tensor var_1859 = const()[name = tensor("op_1859"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1859, x = var_1858_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; tensor transpose_114_perm_0 = const()[name = tensor("transpose_114_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_115_perm_0 = const()[name = tensor("transpose_115_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_115 = transpose(perm = transpose_115_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_247")]; tensor transpose_114 = transpose(perm = transpose_114_perm_0, x = var_1842_cast_fp16)[name = tensor("transpose_248")]; tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_114, y = transpose_115)[name = tensor("matrix_ac_19_cast_fp16")]; tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; tensor var_1868_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1868_cast_fp16")]; tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_37_cast_fp16 = mul(x = var_1868_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_3)[name = tensor("scores_39_cast_fp16")]; tensor var_1874_cast_fp16 = softmax(axis = var_30, x = scores_39_cast_fp16)[name = tensor("op_1874_cast_fp16")]; tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1874_cast_fp16, cond = mask_3)[name = tensor("input_501_cast_fp16")]; tensor x_213_transpose_x_0 = const()[name = tensor("x_213_transpose_x_0"), val = tensor(false)]; tensor x_213_transpose_y_0 = const()[name = tensor("x_213_transpose_y_0"), val = tensor(false)]; tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_246")]; tensor x_213_cast_fp16 = matmul(transpose_x = x_213_transpose_x_0, transpose_y = x_213_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_213_cast_fp16")]; tensor var_1878_perm_0 = const()[name = tensor("op_1878_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1879 = const()[name = tensor("op_1879"), val = tensor([1, -1, 1024])]; tensor var_1878_cast_fp16 = transpose(perm = var_1878_perm_0, x = x_213_cast_fp16)[name = tensor("transpose_245")]; tensor input_503_cast_fp16 = reshape(shape = var_1879, x = var_1878_cast_fp16)[name = tensor("input_503_cast_fp16")]; tensor module_layers_9_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236436288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237486016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237484928)))]; tensor module_layers_9_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237488128)))]; tensor linear_88_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_out_bias_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16_quantized, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; tensor x_217_axes_0 = const()[name = tensor("x_217_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237490240)))]; tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237492352)))]; tensor x_217_cast_fp16 = layer_norm(axes = x_217_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_217_cast_fp16")]; tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237494464))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239593792))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239591680)))]; tensor module_layers_9_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239597952)))]; tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_217_cast_fp16)[name = tensor("transpose_244")]; tensor input_511_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv1_bias_to_fp16, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; tensor x_219_split_num_splits_0 = const()[name = tensor("x_219_split_num_splits_0"), val = tensor(2)]; tensor x_219_split_axis_0 = const()[name = tensor("x_219_split_axis_0"), val = tensor(1)]; tensor x_219_split_cast_fp16_0, tensor x_219_split_cast_fp16_1 = split(axis = x_219_split_axis_0, num_splits = x_219_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_219_split_cast_fp16")]; tensor x_219_split_1_sigmoid_cast_fp16 = sigmoid(x = x_219_split_cast_fp16_1)[name = tensor("x_219_split_1_sigmoid_cast_fp16")]; tensor x_219_cast_fp16 = mul(x = x_219_split_cast_fp16_0, y = x_219_split_1_sigmoid_cast_fp16)[name = tensor("x_219_cast_fp16")]; tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_219_cast_fp16, cond = var_337)[name = tensor("input_513_cast_fp16")]; tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; tensor const_107_to_fp16 = const()[name = tensor("const_107_to_fp16"), val = tensor(0x0p+0)]; tensor input_515_cast_fp16 = pad(constant_val = const_107_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(1024)]; tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; tensor const_266_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_266_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239602112))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239612480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239611392)))]; tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239614592)))]; tensor input_519_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_266_to_fp16_quantized, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; tensor x_221_pad_type_0 = const()[name = tensor("x_221_pad_type_0"), val = tensor("valid")]; tensor x_221_strides_0 = const()[name = tensor("x_221_strides_0"), val = tensor([1])]; tensor x_221_pad_0 = const()[name = tensor("x_221_pad_0"), val = tensor([0, 0])]; tensor x_221_dilations_0 = const()[name = tensor("x_221_dilations_0"), val = tensor([1])]; tensor x_221_groups_0 = const()[name = tensor("x_221_groups_0"), val = tensor(1)]; tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239616704))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(240666432))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(240665344)))]; tensor module_layers_9_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(240668544)))]; tensor x_221_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv2_bias_to_fp16, dilations = x_221_dilations_0, groups = x_221_groups_0, pad = x_221_pad_0, pad_type = x_221_pad_type_0, strides = x_221_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_521_cast_fp16)[name = tensor("x_221_cast_fp16")]; tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_221_cast_fp16)[name = tensor("transpose_243")]; tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(240670656)))]; tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(240672768)))]; tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; tensor module_layers_9_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(240674880))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244873408))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244869248)))]; tensor module_layers_9_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244881664)))]; tensor linear_89_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear1_bias_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16_quantized, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; tensor module_layers_9_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244889920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249085376))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249084288)))]; tensor module_layers_9_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249087488)))]; tensor linear_90_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear2_bias_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16_quantized, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; tensor var_1945_to_fp16 = const()[name = tensor("op_1945_to_fp16"), val = tensor(0x1p-1)]; tensor var_1946_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1945_to_fp16)[name = tensor("op_1946_cast_fp16")]; tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_1946_cast_fp16)[name = tensor("input_537_cast_fp16")]; tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249089600)))]; tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249091712)))]; tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249093824)))]; tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249095936)))]; tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; tensor module_layers_10_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249098048))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253296576))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253292416)))]; tensor module_layers_10_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253304832)))]; tensor linear_91_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear1_bias_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16_quantized, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; tensor module_layers_10_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253313088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257508544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257507456)))]; tensor module_layers_10_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257510656)))]; tensor linear_92_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear2_bias_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16_quantized, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; tensor var_1976_to_fp16 = const()[name = tensor("op_1976_to_fp16"), val = tensor(0x1p-1)]; tensor var_1977_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_1976_to_fp16)[name = tensor("op_1977_cast_fp16")]; tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_1977_cast_fp16)[name = tensor("input_551_cast_fp16")]; tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257512768)))]; tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257514880)))]; tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; tensor module_layers_10_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257516992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258566720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258565632)))]; tensor module_layers_10_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258568832)))]; tensor linear_93_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_q_bias_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16_quantized, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; tensor var_1994 = const()[name = tensor("op_1994"), val = tensor([1, -1, 8, 128])]; tensor q_61_cast_fp16 = reshape(shape = var_1994, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; tensor module_layers_10_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258570944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259620672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259619584)))]; tensor module_layers_10_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259622784)))]; tensor linear_94_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_k_bias_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16_quantized, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; tensor var_1999 = const()[name = tensor("op_1999"), val = tensor([1, -1, 8, 128])]; tensor k_41_cast_fp16 = reshape(shape = var_1999, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; tensor module_layers_10_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259624896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260674624))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260673536)))]; tensor module_layers_10_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260676736)))]; tensor linear_95_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_v_bias_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16_quantized, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; tensor var_2004 = const()[name = tensor("op_2004"), val = tensor([1, -1, 8, 128])]; tensor v_21_cast_fp16 = reshape(shape = var_2004, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260678848)))]; tensor var_2016_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2016_cast_fp16")]; tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260680960)))]; tensor var_2018_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2018_cast_fp16")]; tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_229_transpose_x_0 = const()[name = tensor("x_229_transpose_x_0"), val = tensor(false)]; tensor x_229_transpose_y_0 = const()[name = tensor("x_229_transpose_y_0"), val = tensor(false)]; tensor op_2020_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2020_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260683072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261067584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261067136)))]; tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2018_cast_fp16)[name = tensor("transpose_242")]; tensor x_229_cast_fp16 = matmul(transpose_x = x_229_transpose_x_0, transpose_y = x_229_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = op_2020_to_fp16_quantized)[name = tensor("x_229_cast_fp16")]; tensor x_231_pad_0 = const()[name = tensor("x_231_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_231_mode_0 = const()[name = tensor("x_231_mode_0"), val = tensor("constant")]; tensor const_114_to_fp16 = const()[name = tensor("const_114_to_fp16"), val = tensor(0x0p+0)]; tensor x_231_cast_fp16 = pad(constant_val = const_114_to_fp16, mode = x_231_mode_0, pad = x_231_pad_0, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; tensor var_2028 = const()[name = tensor("op_2028"), val = tensor([1, 8, -1, 188])]; tensor x_233_cast_fp16 = reshape(shape = var_2028, x = x_231_cast_fp16)[name = tensor("x_233_cast_fp16")]; tensor var_2032_begin_0 = const()[name = tensor("op_2032_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2032_end_0 = const()[name = tensor("op_2032_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2032_end_mask_0 = const()[name = tensor("op_2032_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2032_cast_fp16 = slice_by_index(begin = var_2032_begin_0, end = var_2032_end_0, end_mask = var_2032_end_mask_0, x = x_233_cast_fp16)[name = tensor("op_2032_cast_fp16")]; tensor var_2033 = const()[name = tensor("op_2033"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2033, x = var_2032_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; tensor transpose_116_perm_0 = const()[name = tensor("transpose_116_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_117_perm_0 = const()[name = tensor("transpose_117_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_117 = transpose(perm = transpose_117_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_240")]; tensor transpose_116 = transpose(perm = transpose_116_perm_0, x = var_2016_cast_fp16)[name = tensor("transpose_241")]; tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_116, y = transpose_117)[name = tensor("matrix_ac_21_cast_fp16")]; tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; tensor var_2042_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2042_cast_fp16")]; tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_41_cast_fp16 = mul(x = var_2042_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_3)[name = tensor("scores_43_cast_fp16")]; tensor var_2048_cast_fp16 = softmax(axis = var_30, x = scores_43_cast_fp16)[name = tensor("op_2048_cast_fp16")]; tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_2048_cast_fp16, cond = mask_3)[name = tensor("input_553_cast_fp16")]; tensor x_235_transpose_x_0 = const()[name = tensor("x_235_transpose_x_0"), val = tensor(false)]; tensor x_235_transpose_y_0 = const()[name = tensor("x_235_transpose_y_0"), val = tensor(false)]; tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_239")]; tensor x_235_cast_fp16 = matmul(transpose_x = x_235_transpose_x_0, transpose_y = x_235_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_235_cast_fp16")]; tensor var_2052_perm_0 = const()[name = tensor("op_2052_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2053 = const()[name = tensor("op_2053"), val = tensor([1, -1, 1024])]; tensor var_2052_cast_fp16 = transpose(perm = var_2052_perm_0, x = x_235_cast_fp16)[name = tensor("transpose_238")]; tensor input_555_cast_fp16 = reshape(shape = var_2053, x = var_2052_cast_fp16)[name = tensor("input_555_cast_fp16")]; tensor module_layers_10_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261068416))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262118144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262117056)))]; tensor module_layers_10_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262120256)))]; tensor linear_97_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_out_bias_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16_quantized, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; tensor x_239_axes_0 = const()[name = tensor("x_239_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262122368)))]; tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262124480)))]; tensor x_239_cast_fp16 = layer_norm(axes = x_239_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_239_cast_fp16")]; tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262126592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264225920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264223808)))]; tensor module_layers_10_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264230080)))]; tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_239_cast_fp16)[name = tensor("transpose_237")]; tensor input_563_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv1_bias_to_fp16, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; tensor x_241_split_num_splits_0 = const()[name = tensor("x_241_split_num_splits_0"), val = tensor(2)]; tensor x_241_split_axis_0 = const()[name = tensor("x_241_split_axis_0"), val = tensor(1)]; tensor x_241_split_cast_fp16_0, tensor x_241_split_cast_fp16_1 = split(axis = x_241_split_axis_0, num_splits = x_241_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_241_split_cast_fp16")]; tensor x_241_split_1_sigmoid_cast_fp16 = sigmoid(x = x_241_split_cast_fp16_1)[name = tensor("x_241_split_1_sigmoid_cast_fp16")]; tensor x_241_cast_fp16 = mul(x = x_241_split_cast_fp16_0, y = x_241_split_1_sigmoid_cast_fp16)[name = tensor("x_241_cast_fp16")]; tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_241_cast_fp16, cond = var_337)[name = tensor("input_565_cast_fp16")]; tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; tensor const_117_to_fp16 = const()[name = tensor("const_117_to_fp16"), val = tensor(0x0p+0)]; tensor input_567_cast_fp16 = pad(constant_val = const_117_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(1024)]; tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; tensor const_268_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_268_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264234240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264244608))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264243520)))]; tensor const_269_to_fp16 = const()[name = tensor("const_269_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264246720)))]; tensor input_571_cast_fp16 = conv(bias = const_269_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_268_to_fp16_quantized, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; tensor x_243_pad_type_0 = const()[name = tensor("x_243_pad_type_0"), val = tensor("valid")]; tensor x_243_strides_0 = const()[name = tensor("x_243_strides_0"), val = tensor([1])]; tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0])]; tensor x_243_dilations_0 = const()[name = tensor("x_243_dilations_0"), val = tensor([1])]; tensor x_243_groups_0 = const()[name = tensor("x_243_groups_0"), val = tensor(1)]; tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264248832))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265298560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265297472)))]; tensor module_layers_10_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265300672)))]; tensor x_243_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv2_bias_to_fp16, dilations = x_243_dilations_0, groups = x_243_groups_0, pad = x_243_pad_0, pad_type = x_243_pad_type_0, strides = x_243_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_573_cast_fp16)[name = tensor("x_243_cast_fp16")]; tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_243_cast_fp16)[name = tensor("transpose_236")]; tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265302784)))]; tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265304896)))]; tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; tensor module_layers_10_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265307008))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269505536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269501376)))]; tensor module_layers_10_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269513792)))]; tensor linear_98_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear1_bias_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16_quantized, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; tensor module_layers_10_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269522048))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273717504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273716416)))]; tensor module_layers_10_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273719616)))]; tensor linear_99_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear2_bias_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16_quantized, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; tensor var_2119_to_fp16 = const()[name = tensor("op_2119_to_fp16"), val = tensor(0x1p-1)]; tensor var_2120_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2119_to_fp16)[name = tensor("op_2120_cast_fp16")]; tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_2120_cast_fp16)[name = tensor("input_589_cast_fp16")]; tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273721728)))]; tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273723840)))]; tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273725952)))]; tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273728064)))]; tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; tensor module_layers_11_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273730176))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277928704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277924544)))]; tensor module_layers_11_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277936960)))]; tensor linear_100_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear1_bias_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16_quantized, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; tensor module_layers_11_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277945216))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282140672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282139584)))]; tensor module_layers_11_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282142784)))]; tensor linear_101_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear2_bias_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16_quantized, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; tensor var_2150_to_fp16 = const()[name = tensor("op_2150_to_fp16"), val = tensor(0x1p-1)]; tensor var_2151_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2150_to_fp16)[name = tensor("op_2151_cast_fp16")]; tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2151_cast_fp16)[name = tensor("input_603_cast_fp16")]; tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282144896)))]; tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282147008)))]; tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; tensor module_layers_11_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282149120))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283198848))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283197760)))]; tensor module_layers_11_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283200960)))]; tensor linear_102_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_q_bias_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16_quantized, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; tensor var_2168 = const()[name = tensor("op_2168"), val = tensor([1, -1, 8, 128])]; tensor q_67_cast_fp16 = reshape(shape = var_2168, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; tensor module_layers_11_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283203072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284252800))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284251712)))]; tensor module_layers_11_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284254912)))]; tensor linear_103_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_k_bias_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16_quantized, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; tensor var_2173 = const()[name = tensor("op_2173"), val = tensor([1, -1, 8, 128])]; tensor k_45_cast_fp16 = reshape(shape = var_2173, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; tensor module_layers_11_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284257024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285306752))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285305664)))]; tensor module_layers_11_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285308864)))]; tensor linear_104_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_v_bias_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16_quantized, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; tensor var_2178 = const()[name = tensor("op_2178"), val = tensor([1, -1, 8, 128])]; tensor v_23_cast_fp16 = reshape(shape = var_2178, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285310976)))]; tensor var_2190_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2190_cast_fp16")]; tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285313088)))]; tensor var_2192_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2192_cast_fp16")]; tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_251_transpose_x_0 = const()[name = tensor("x_251_transpose_x_0"), val = tensor(false)]; tensor x_251_transpose_y_0 = const()[name = tensor("x_251_transpose_y_0"), val = tensor(false)]; tensor op_2194_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2194_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285315200))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285699712))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285699264)))]; tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2192_cast_fp16)[name = tensor("transpose_235")]; tensor x_251_cast_fp16 = matmul(transpose_x = x_251_transpose_x_0, transpose_y = x_251_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = op_2194_to_fp16_quantized)[name = tensor("x_251_cast_fp16")]; tensor x_253_pad_0 = const()[name = tensor("x_253_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_253_mode_0 = const()[name = tensor("x_253_mode_0"), val = tensor("constant")]; tensor const_124_to_fp16 = const()[name = tensor("const_124_to_fp16"), val = tensor(0x0p+0)]; tensor x_253_cast_fp16 = pad(constant_val = const_124_to_fp16, mode = x_253_mode_0, pad = x_253_pad_0, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; tensor var_2202 = const()[name = tensor("op_2202"), val = tensor([1, 8, -1, 188])]; tensor x_255_cast_fp16 = reshape(shape = var_2202, x = x_253_cast_fp16)[name = tensor("x_255_cast_fp16")]; tensor var_2206_begin_0 = const()[name = tensor("op_2206_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2206_end_0 = const()[name = tensor("op_2206_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2206_end_mask_0 = const()[name = tensor("op_2206_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2206_cast_fp16 = slice_by_index(begin = var_2206_begin_0, end = var_2206_end_0, end_mask = var_2206_end_mask_0, x = x_255_cast_fp16)[name = tensor("op_2206_cast_fp16")]; tensor var_2207 = const()[name = tensor("op_2207"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2207, x = var_2206_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; tensor transpose_118_perm_0 = const()[name = tensor("transpose_118_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_119_perm_0 = const()[name = tensor("transpose_119_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_119 = transpose(perm = transpose_119_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_233")]; tensor transpose_118 = transpose(perm = transpose_118_perm_0, x = var_2190_cast_fp16)[name = tensor("transpose_234")]; tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_118, y = transpose_119)[name = tensor("matrix_ac_23_cast_fp16")]; tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; tensor var_2216_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2216_cast_fp16")]; tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_45_cast_fp16 = mul(x = var_2216_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_3)[name = tensor("scores_47_cast_fp16")]; tensor var_2222_cast_fp16 = softmax(axis = var_30, x = scores_47_cast_fp16)[name = tensor("op_2222_cast_fp16")]; tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2222_cast_fp16, cond = mask_3)[name = tensor("input_605_cast_fp16")]; tensor x_257_transpose_x_0 = const()[name = tensor("x_257_transpose_x_0"), val = tensor(false)]; tensor x_257_transpose_y_0 = const()[name = tensor("x_257_transpose_y_0"), val = tensor(false)]; tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_232")]; tensor x_257_cast_fp16 = matmul(transpose_x = x_257_transpose_x_0, transpose_y = x_257_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_257_cast_fp16")]; tensor var_2226_perm_0 = const()[name = tensor("op_2226_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2227 = const()[name = tensor("op_2227"), val = tensor([1, -1, 1024])]; tensor var_2226_cast_fp16 = transpose(perm = var_2226_perm_0, x = x_257_cast_fp16)[name = tensor("transpose_231")]; tensor input_607_cast_fp16 = reshape(shape = var_2227, x = var_2226_cast_fp16)[name = tensor("input_607_cast_fp16")]; tensor module_layers_11_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285700544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286750272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286749184)))]; tensor module_layers_11_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286752384)))]; tensor linear_106_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_out_bias_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16_quantized, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; tensor x_261_axes_0 = const()[name = tensor("x_261_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286754496)))]; tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286756608)))]; tensor x_261_cast_fp16 = layer_norm(axes = x_261_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_261_cast_fp16")]; tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286758720))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288858048))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288855936)))]; tensor module_layers_11_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288862208)))]; tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_261_cast_fp16)[name = tensor("transpose_230")]; tensor input_615_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv1_bias_to_fp16, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; tensor x_263_split_num_splits_0 = const()[name = tensor("x_263_split_num_splits_0"), val = tensor(2)]; tensor x_263_split_axis_0 = const()[name = tensor("x_263_split_axis_0"), val = tensor(1)]; tensor x_263_split_cast_fp16_0, tensor x_263_split_cast_fp16_1 = split(axis = x_263_split_axis_0, num_splits = x_263_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_263_split_cast_fp16")]; tensor x_263_split_1_sigmoid_cast_fp16 = sigmoid(x = x_263_split_cast_fp16_1)[name = tensor("x_263_split_1_sigmoid_cast_fp16")]; tensor x_263_cast_fp16 = mul(x = x_263_split_cast_fp16_0, y = x_263_split_1_sigmoid_cast_fp16)[name = tensor("x_263_cast_fp16")]; tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_263_cast_fp16, cond = var_337)[name = tensor("input_617_cast_fp16")]; tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; tensor const_127_to_fp16 = const()[name = tensor("const_127_to_fp16"), val = tensor(0x0p+0)]; tensor input_619_cast_fp16 = pad(constant_val = const_127_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(1024)]; tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; tensor const_270_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_270_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288866368))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288876736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288875648)))]; tensor const_271_to_fp16 = const()[name = tensor("const_271_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288878848)))]; tensor input_623_cast_fp16 = conv(bias = const_271_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_270_to_fp16_quantized, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; tensor x_265_pad_type_0 = const()[name = tensor("x_265_pad_type_0"), val = tensor("valid")]; tensor x_265_strides_0 = const()[name = tensor("x_265_strides_0"), val = tensor([1])]; tensor x_265_pad_0 = const()[name = tensor("x_265_pad_0"), val = tensor([0, 0])]; tensor x_265_dilations_0 = const()[name = tensor("x_265_dilations_0"), val = tensor([1])]; tensor x_265_groups_0 = const()[name = tensor("x_265_groups_0"), val = tensor(1)]; tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288880960))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289930688))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289929600)))]; tensor module_layers_11_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289932800)))]; tensor x_265_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv2_bias_to_fp16, dilations = x_265_dilations_0, groups = x_265_groups_0, pad = x_265_pad_0, pad_type = x_265_pad_type_0, strides = x_265_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_625_cast_fp16)[name = tensor("x_265_cast_fp16")]; tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_265_cast_fp16)[name = tensor("transpose_229")]; tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289934912)))]; tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289937024)))]; tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; tensor module_layers_11_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289939136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294137664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294133504)))]; tensor module_layers_11_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294145920)))]; tensor linear_107_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear1_bias_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16_quantized, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; tensor module_layers_11_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294154176))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298349632))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298348544)))]; tensor module_layers_11_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298351744)))]; tensor linear_108_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear2_bias_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16_quantized, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; tensor var_2293_to_fp16 = const()[name = tensor("op_2293_to_fp16"), val = tensor(0x1p-1)]; tensor var_2294_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2293_to_fp16)[name = tensor("op_2294_cast_fp16")]; tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2294_cast_fp16)[name = tensor("input_641_cast_fp16")]; tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298353856)))]; tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298355968)))]; tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298358080)))]; tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298360192)))]; tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; tensor module_layers_12_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298362304))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302560832))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302556672)))]; tensor module_layers_12_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302569088)))]; tensor linear_109_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear1_bias_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16_quantized, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; tensor module_layers_12_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302577344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306772800))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306771712)))]; tensor module_layers_12_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306774912)))]; tensor linear_110_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear2_bias_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16_quantized, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; tensor var_2324_to_fp16 = const()[name = tensor("op_2324_to_fp16"), val = tensor(0x1p-1)]; tensor var_2325_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2324_to_fp16)[name = tensor("op_2325_cast_fp16")]; tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2325_cast_fp16)[name = tensor("input_655_cast_fp16")]; tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306777024)))]; tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306779136)))]; tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; tensor module_layers_12_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306781248))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307830976))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307829888)))]; tensor module_layers_12_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307833088)))]; tensor linear_111_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_q_bias_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16_quantized, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; tensor var_2342 = const()[name = tensor("op_2342"), val = tensor([1, -1, 8, 128])]; tensor q_73_cast_fp16 = reshape(shape = var_2342, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; tensor module_layers_12_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307835200))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308884928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308883840)))]; tensor module_layers_12_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308887040)))]; tensor linear_112_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_k_bias_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16_quantized, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; tensor var_2347 = const()[name = tensor("op_2347"), val = tensor([1, -1, 8, 128])]; tensor k_49_cast_fp16 = reshape(shape = var_2347, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; tensor module_layers_12_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308889152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309938880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309937792)))]; tensor module_layers_12_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309940992)))]; tensor linear_113_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_v_bias_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16_quantized, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; tensor var_2352 = const()[name = tensor("op_2352"), val = tensor([1, -1, 8, 128])]; tensor v_25_cast_fp16 = reshape(shape = var_2352, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309943104)))]; tensor var_2364_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2364_cast_fp16")]; tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309945216)))]; tensor var_2366_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2366_cast_fp16")]; tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; tensor op_2368_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2368_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309947328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310331840))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310331392)))]; tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2366_cast_fp16)[name = tensor("transpose_228")]; tensor x_273_cast_fp16 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = op_2368_to_fp16_quantized)[name = tensor("x_273_cast_fp16")]; tensor x_275_pad_0 = const()[name = tensor("x_275_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_275_mode_0 = const()[name = tensor("x_275_mode_0"), val = tensor("constant")]; tensor const_134_to_fp16 = const()[name = tensor("const_134_to_fp16"), val = tensor(0x0p+0)]; tensor x_275_cast_fp16 = pad(constant_val = const_134_to_fp16, mode = x_275_mode_0, pad = x_275_pad_0, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; tensor var_2376 = const()[name = tensor("op_2376"), val = tensor([1, 8, -1, 188])]; tensor x_277_cast_fp16 = reshape(shape = var_2376, x = x_275_cast_fp16)[name = tensor("x_277_cast_fp16")]; tensor var_2380_begin_0 = const()[name = tensor("op_2380_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2380_end_0 = const()[name = tensor("op_2380_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2380_end_mask_0 = const()[name = tensor("op_2380_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2380_cast_fp16 = slice_by_index(begin = var_2380_begin_0, end = var_2380_end_0, end_mask = var_2380_end_mask_0, x = x_277_cast_fp16)[name = tensor("op_2380_cast_fp16")]; tensor var_2381 = const()[name = tensor("op_2381"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2381, x = var_2380_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; tensor transpose_120_perm_0 = const()[name = tensor("transpose_120_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_121_perm_0 = const()[name = tensor("transpose_121_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_121 = transpose(perm = transpose_121_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_226")]; tensor transpose_120 = transpose(perm = transpose_120_perm_0, x = var_2364_cast_fp16)[name = tensor("transpose_227")]; tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_120, y = transpose_121)[name = tensor("matrix_ac_25_cast_fp16")]; tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; tensor var_2390_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2390_cast_fp16")]; tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_49_cast_fp16 = mul(x = var_2390_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_3)[name = tensor("scores_51_cast_fp16")]; tensor var_2396_cast_fp16 = softmax(axis = var_30, x = scores_51_cast_fp16)[name = tensor("op_2396_cast_fp16")]; tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2396_cast_fp16, cond = mask_3)[name = tensor("input_657_cast_fp16")]; tensor x_279_transpose_x_0 = const()[name = tensor("x_279_transpose_x_0"), val = tensor(false)]; tensor x_279_transpose_y_0 = const()[name = tensor("x_279_transpose_y_0"), val = tensor(false)]; tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_225")]; tensor x_279_cast_fp16 = matmul(transpose_x = x_279_transpose_x_0, transpose_y = x_279_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_279_cast_fp16")]; tensor var_2400_perm_0 = const()[name = tensor("op_2400_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2401 = const()[name = tensor("op_2401"), val = tensor([1, -1, 1024])]; tensor var_2400_cast_fp16 = transpose(perm = var_2400_perm_0, x = x_279_cast_fp16)[name = tensor("transpose_224")]; tensor input_659_cast_fp16 = reshape(shape = var_2401, x = var_2400_cast_fp16)[name = tensor("input_659_cast_fp16")]; tensor module_layers_12_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310332672))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311382400))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311381312)))]; tensor module_layers_12_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311384512)))]; tensor linear_115_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_out_bias_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16_quantized, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311386624)))]; tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311388736)))]; tensor x_283_cast_fp16 = layer_norm(axes = x_283_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_283_cast_fp16")]; tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311390848))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313490176))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313488064)))]; tensor module_layers_12_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313494336)))]; tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_283_cast_fp16)[name = tensor("transpose_223")]; tensor input_667_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv1_bias_to_fp16, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; tensor x_285_split_num_splits_0 = const()[name = tensor("x_285_split_num_splits_0"), val = tensor(2)]; tensor x_285_split_axis_0 = const()[name = tensor("x_285_split_axis_0"), val = tensor(1)]; tensor x_285_split_cast_fp16_0, tensor x_285_split_cast_fp16_1 = split(axis = x_285_split_axis_0, num_splits = x_285_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_285_split_cast_fp16")]; tensor x_285_split_1_sigmoid_cast_fp16 = sigmoid(x = x_285_split_cast_fp16_1)[name = tensor("x_285_split_1_sigmoid_cast_fp16")]; tensor x_285_cast_fp16 = mul(x = x_285_split_cast_fp16_0, y = x_285_split_1_sigmoid_cast_fp16)[name = tensor("x_285_cast_fp16")]; tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_285_cast_fp16, cond = var_337)[name = tensor("input_669_cast_fp16")]; tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; tensor const_137_to_fp16 = const()[name = tensor("const_137_to_fp16"), val = tensor(0x0p+0)]; tensor input_671_cast_fp16 = pad(constant_val = const_137_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(1024)]; tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; tensor const_272_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_272_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313498496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313508864))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313507776)))]; tensor const_273_to_fp16 = const()[name = tensor("const_273_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313510976)))]; tensor input_675_cast_fp16 = conv(bias = const_273_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_272_to_fp16_quantized, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; tensor x_287_pad_type_0 = const()[name = tensor("x_287_pad_type_0"), val = tensor("valid")]; tensor x_287_strides_0 = const()[name = tensor("x_287_strides_0"), val = tensor([1])]; tensor x_287_pad_0 = const()[name = tensor("x_287_pad_0"), val = tensor([0, 0])]; tensor x_287_dilations_0 = const()[name = tensor("x_287_dilations_0"), val = tensor([1])]; tensor x_287_groups_0 = const()[name = tensor("x_287_groups_0"), val = tensor(1)]; tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313513088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(314562816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(314561728)))]; tensor module_layers_12_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(314564928)))]; tensor x_287_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv2_bias_to_fp16, dilations = x_287_dilations_0, groups = x_287_groups_0, pad = x_287_pad_0, pad_type = x_287_pad_type_0, strides = x_287_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_677_cast_fp16)[name = tensor("x_287_cast_fp16")]; tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_287_cast_fp16)[name = tensor("transpose_222")]; tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(314567040)))]; tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(314569152)))]; tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; tensor module_layers_12_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(314571264))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318769792))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318765632)))]; tensor module_layers_12_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318778048)))]; tensor linear_116_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear1_bias_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16_quantized, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; tensor module_layers_12_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318786304))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322981760))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322980672)))]; tensor module_layers_12_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322983872)))]; tensor linear_117_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear2_bias_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16_quantized, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; tensor var_2467_to_fp16 = const()[name = tensor("op_2467_to_fp16"), val = tensor(0x1p-1)]; tensor var_2468_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2467_to_fp16)[name = tensor("op_2468_cast_fp16")]; tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2468_cast_fp16)[name = tensor("input_693_cast_fp16")]; tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322985984)))]; tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322988096)))]; tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322990208)))]; tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322992320)))]; tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; tensor module_layers_13_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322994432))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327192960))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327188800)))]; tensor module_layers_13_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327201216)))]; tensor linear_118_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear1_bias_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16_quantized, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; tensor module_layers_13_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327209472))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331404928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331403840)))]; tensor module_layers_13_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331407040)))]; tensor linear_119_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear2_bias_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16_quantized, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; tensor var_2498_to_fp16 = const()[name = tensor("op_2498_to_fp16"), val = tensor(0x1p-1)]; tensor var_2499_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2498_to_fp16)[name = tensor("op_2499_cast_fp16")]; tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2499_cast_fp16)[name = tensor("input_707_cast_fp16")]; tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331409152)))]; tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331411264)))]; tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; tensor module_layers_13_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331413376))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332463104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332462016)))]; tensor module_layers_13_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332465216)))]; tensor linear_120_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_q_bias_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16_quantized, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; tensor var_2516 = const()[name = tensor("op_2516"), val = tensor([1, -1, 8, 128])]; tensor q_79_cast_fp16 = reshape(shape = var_2516, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; tensor module_layers_13_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332467328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333517056))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333515968)))]; tensor module_layers_13_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333519168)))]; tensor linear_121_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_k_bias_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16_quantized, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; tensor var_2521 = const()[name = tensor("op_2521"), val = tensor([1, -1, 8, 128])]; tensor k_53_cast_fp16 = reshape(shape = var_2521, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; tensor module_layers_13_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333521280))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334571008))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334569920)))]; tensor module_layers_13_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334573120)))]; tensor linear_122_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_v_bias_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16_quantized, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; tensor var_2526 = const()[name = tensor("op_2526"), val = tensor([1, -1, 8, 128])]; tensor v_27_cast_fp16 = reshape(shape = var_2526, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334575232)))]; tensor var_2538_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2538_cast_fp16")]; tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334577344)))]; tensor var_2540_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2540_cast_fp16")]; tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_295_transpose_x_0 = const()[name = tensor("x_295_transpose_x_0"), val = tensor(false)]; tensor x_295_transpose_y_0 = const()[name = tensor("x_295_transpose_y_0"), val = tensor(false)]; tensor op_2542_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2542_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334579456))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334963968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334963520)))]; tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2540_cast_fp16)[name = tensor("transpose_221")]; tensor x_295_cast_fp16 = matmul(transpose_x = x_295_transpose_x_0, transpose_y = x_295_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = op_2542_to_fp16_quantized)[name = tensor("x_295_cast_fp16")]; tensor x_297_pad_0 = const()[name = tensor("x_297_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_297_mode_0 = const()[name = tensor("x_297_mode_0"), val = tensor("constant")]; tensor const_144_to_fp16 = const()[name = tensor("const_144_to_fp16"), val = tensor(0x0p+0)]; tensor x_297_cast_fp16 = pad(constant_val = const_144_to_fp16, mode = x_297_mode_0, pad = x_297_pad_0, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; tensor var_2550 = const()[name = tensor("op_2550"), val = tensor([1, 8, -1, 188])]; tensor x_299_cast_fp16 = reshape(shape = var_2550, x = x_297_cast_fp16)[name = tensor("x_299_cast_fp16")]; tensor var_2554_begin_0 = const()[name = tensor("op_2554_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2554_end_0 = const()[name = tensor("op_2554_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2554_end_mask_0 = const()[name = tensor("op_2554_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2554_cast_fp16 = slice_by_index(begin = var_2554_begin_0, end = var_2554_end_0, end_mask = var_2554_end_mask_0, x = x_299_cast_fp16)[name = tensor("op_2554_cast_fp16")]; tensor var_2555 = const()[name = tensor("op_2555"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2555, x = var_2554_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; tensor transpose_122_perm_0 = const()[name = tensor("transpose_122_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_123_perm_0 = const()[name = tensor("transpose_123_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_123 = transpose(perm = transpose_123_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_219")]; tensor transpose_122 = transpose(perm = transpose_122_perm_0, x = var_2538_cast_fp16)[name = tensor("transpose_220")]; tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_122, y = transpose_123)[name = tensor("matrix_ac_27_cast_fp16")]; tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; tensor var_2564_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2564_cast_fp16")]; tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_53_cast_fp16 = mul(x = var_2564_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_3)[name = tensor("scores_55_cast_fp16")]; tensor var_2570_cast_fp16 = softmax(axis = var_30, x = scores_55_cast_fp16)[name = tensor("op_2570_cast_fp16")]; tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2570_cast_fp16, cond = mask_3)[name = tensor("input_709_cast_fp16")]; tensor x_301_transpose_x_0 = const()[name = tensor("x_301_transpose_x_0"), val = tensor(false)]; tensor x_301_transpose_y_0 = const()[name = tensor("x_301_transpose_y_0"), val = tensor(false)]; tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_218")]; tensor x_301_cast_fp16 = matmul(transpose_x = x_301_transpose_x_0, transpose_y = x_301_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_301_cast_fp16")]; tensor var_2574_perm_0 = const()[name = tensor("op_2574_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2575 = const()[name = tensor("op_2575"), val = tensor([1, -1, 1024])]; tensor var_2574_cast_fp16 = transpose(perm = var_2574_perm_0, x = x_301_cast_fp16)[name = tensor("transpose_217")]; tensor input_711_cast_fp16 = reshape(shape = var_2575, x = var_2574_cast_fp16)[name = tensor("input_711_cast_fp16")]; tensor module_layers_13_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334964800))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336014528))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336013440)))]; tensor module_layers_13_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336016640)))]; tensor linear_124_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_out_bias_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16_quantized, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; tensor x_305_axes_0 = const()[name = tensor("x_305_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336018752)))]; tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336020864)))]; tensor x_305_cast_fp16 = layer_norm(axes = x_305_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_305_cast_fp16")]; tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336022976))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338122304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338120192)))]; tensor module_layers_13_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338126464)))]; tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_305_cast_fp16)[name = tensor("transpose_216")]; tensor input_719_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv1_bias_to_fp16, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; tensor x_307_split_num_splits_0 = const()[name = tensor("x_307_split_num_splits_0"), val = tensor(2)]; tensor x_307_split_axis_0 = const()[name = tensor("x_307_split_axis_0"), val = tensor(1)]; tensor x_307_split_cast_fp16_0, tensor x_307_split_cast_fp16_1 = split(axis = x_307_split_axis_0, num_splits = x_307_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_307_split_cast_fp16")]; tensor x_307_split_1_sigmoid_cast_fp16 = sigmoid(x = x_307_split_cast_fp16_1)[name = tensor("x_307_split_1_sigmoid_cast_fp16")]; tensor x_307_cast_fp16 = mul(x = x_307_split_cast_fp16_0, y = x_307_split_1_sigmoid_cast_fp16)[name = tensor("x_307_cast_fp16")]; tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_307_cast_fp16, cond = var_337)[name = tensor("input_721_cast_fp16")]; tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; tensor const_147_to_fp16 = const()[name = tensor("const_147_to_fp16"), val = tensor(0x0p+0)]; tensor input_723_cast_fp16 = pad(constant_val = const_147_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(1024)]; tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; tensor const_274_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_274_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338130624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338140992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338139904)))]; tensor const_275_to_fp16 = const()[name = tensor("const_275_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338143104)))]; tensor input_727_cast_fp16 = conv(bias = const_275_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_274_to_fp16_quantized, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; tensor x_309_pad_type_0 = const()[name = tensor("x_309_pad_type_0"), val = tensor("valid")]; tensor x_309_strides_0 = const()[name = tensor("x_309_strides_0"), val = tensor([1])]; tensor x_309_pad_0 = const()[name = tensor("x_309_pad_0"), val = tensor([0, 0])]; tensor x_309_dilations_0 = const()[name = tensor("x_309_dilations_0"), val = tensor([1])]; tensor x_309_groups_0 = const()[name = tensor("x_309_groups_0"), val = tensor(1)]; tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338145216))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(339194944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(339193856)))]; tensor module_layers_13_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(339197056)))]; tensor x_309_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv2_bias_to_fp16, dilations = x_309_dilations_0, groups = x_309_groups_0, pad = x_309_pad_0, pad_type = x_309_pad_type_0, strides = x_309_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_729_cast_fp16)[name = tensor("x_309_cast_fp16")]; tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_309_cast_fp16)[name = tensor("transpose_215")]; tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(339199168)))]; tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(339201280)))]; tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; tensor module_layers_13_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(339203392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343401920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343397760)))]; tensor module_layers_13_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343410176)))]; tensor linear_125_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear1_bias_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16_quantized, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; tensor module_layers_13_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343418432))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347613888))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347612800)))]; tensor module_layers_13_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347616000)))]; tensor linear_126_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear2_bias_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16_quantized, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; tensor var_2641_to_fp16 = const()[name = tensor("op_2641_to_fp16"), val = tensor(0x1p-1)]; tensor var_2642_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2641_to_fp16)[name = tensor("op_2642_cast_fp16")]; tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2642_cast_fp16)[name = tensor("input_745_cast_fp16")]; tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347618112)))]; tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347620224)))]; tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347622336)))]; tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347624448)))]; tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; tensor module_layers_14_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347626560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351825088))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351820928)))]; tensor module_layers_14_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351833344)))]; tensor linear_127_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear1_bias_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16_quantized, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; tensor module_layers_14_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351841600))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356037056))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356035968)))]; tensor module_layers_14_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356039168)))]; tensor linear_128_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear2_bias_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16_quantized, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; tensor var_2672_to_fp16 = const()[name = tensor("op_2672_to_fp16"), val = tensor(0x1p-1)]; tensor var_2673_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2672_to_fp16)[name = tensor("op_2673_cast_fp16")]; tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2673_cast_fp16)[name = tensor("input_759_cast_fp16")]; tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356041280)))]; tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356043392)))]; tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; tensor module_layers_14_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356045504))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357095232))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357094144)))]; tensor module_layers_14_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357097344)))]; tensor linear_129_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_q_bias_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16_quantized, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; tensor var_2690 = const()[name = tensor("op_2690"), val = tensor([1, -1, 8, 128])]; tensor q_85_cast_fp16 = reshape(shape = var_2690, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; tensor module_layers_14_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357099456))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358149184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358148096)))]; tensor module_layers_14_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358151296)))]; tensor linear_130_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_k_bias_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16_quantized, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; tensor var_2695 = const()[name = tensor("op_2695"), val = tensor([1, -1, 8, 128])]; tensor k_57_cast_fp16 = reshape(shape = var_2695, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; tensor module_layers_14_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358153408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359203136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359202048)))]; tensor module_layers_14_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359205248)))]; tensor linear_131_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_v_bias_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16_quantized, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; tensor var_2700 = const()[name = tensor("op_2700"), val = tensor([1, -1, 8, 128])]; tensor v_29_cast_fp16 = reshape(shape = var_2700, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359207360)))]; tensor var_2712_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2712_cast_fp16")]; tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359209472)))]; tensor var_2714_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2714_cast_fp16")]; tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_317_transpose_x_0 = const()[name = tensor("x_317_transpose_x_0"), val = tensor(false)]; tensor x_317_transpose_y_0 = const()[name = tensor("x_317_transpose_y_0"), val = tensor(false)]; tensor op_2716_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2716_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359211584))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359596096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359595648)))]; tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2714_cast_fp16)[name = tensor("transpose_214")]; tensor x_317_cast_fp16 = matmul(transpose_x = x_317_transpose_x_0, transpose_y = x_317_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = op_2716_to_fp16_quantized)[name = tensor("x_317_cast_fp16")]; tensor x_319_pad_0 = const()[name = tensor("x_319_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_319_mode_0 = const()[name = tensor("x_319_mode_0"), val = tensor("constant")]; tensor const_154_to_fp16 = const()[name = tensor("const_154_to_fp16"), val = tensor(0x0p+0)]; tensor x_319_cast_fp16 = pad(constant_val = const_154_to_fp16, mode = x_319_mode_0, pad = x_319_pad_0, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; tensor var_2724 = const()[name = tensor("op_2724"), val = tensor([1, 8, -1, 188])]; tensor x_321_cast_fp16 = reshape(shape = var_2724, x = x_319_cast_fp16)[name = tensor("x_321_cast_fp16")]; tensor var_2728_begin_0 = const()[name = tensor("op_2728_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2728_end_0 = const()[name = tensor("op_2728_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2728_end_mask_0 = const()[name = tensor("op_2728_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2728_cast_fp16 = slice_by_index(begin = var_2728_begin_0, end = var_2728_end_0, end_mask = var_2728_end_mask_0, x = x_321_cast_fp16)[name = tensor("op_2728_cast_fp16")]; tensor var_2729 = const()[name = tensor("op_2729"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2729, x = var_2728_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; tensor transpose_124_perm_0 = const()[name = tensor("transpose_124_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_125_perm_0 = const()[name = tensor("transpose_125_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_125 = transpose(perm = transpose_125_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_212")]; tensor transpose_124 = transpose(perm = transpose_124_perm_0, x = var_2712_cast_fp16)[name = tensor("transpose_213")]; tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_124, y = transpose_125)[name = tensor("matrix_ac_29_cast_fp16")]; tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; tensor var_2738_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2738_cast_fp16")]; tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_57_cast_fp16 = mul(x = var_2738_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_3)[name = tensor("scores_59_cast_fp16")]; tensor var_2744_cast_fp16 = softmax(axis = var_30, x = scores_59_cast_fp16)[name = tensor("op_2744_cast_fp16")]; tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2744_cast_fp16, cond = mask_3)[name = tensor("input_761_cast_fp16")]; tensor x_323_transpose_x_0 = const()[name = tensor("x_323_transpose_x_0"), val = tensor(false)]; tensor x_323_transpose_y_0 = const()[name = tensor("x_323_transpose_y_0"), val = tensor(false)]; tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_211")]; tensor x_323_cast_fp16 = matmul(transpose_x = x_323_transpose_x_0, transpose_y = x_323_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_323_cast_fp16")]; tensor var_2748_perm_0 = const()[name = tensor("op_2748_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2749 = const()[name = tensor("op_2749"), val = tensor([1, -1, 1024])]; tensor var_2748_cast_fp16 = transpose(perm = var_2748_perm_0, x = x_323_cast_fp16)[name = tensor("transpose_210")]; tensor input_763_cast_fp16 = reshape(shape = var_2749, x = var_2748_cast_fp16)[name = tensor("input_763_cast_fp16")]; tensor module_layers_14_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359596928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360646656))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360645568)))]; tensor module_layers_14_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360648768)))]; tensor linear_133_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_out_bias_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16_quantized, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; tensor x_327_axes_0 = const()[name = tensor("x_327_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360650880)))]; tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360652992)))]; tensor x_327_cast_fp16 = layer_norm(axes = x_327_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_327_cast_fp16")]; tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360655104))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362754432))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362752320)))]; tensor module_layers_14_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362758592)))]; tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_327_cast_fp16)[name = tensor("transpose_209")]; tensor input_771_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv1_bias_to_fp16, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; tensor x_329_split_num_splits_0 = const()[name = tensor("x_329_split_num_splits_0"), val = tensor(2)]; tensor x_329_split_axis_0 = const()[name = tensor("x_329_split_axis_0"), val = tensor(1)]; tensor x_329_split_cast_fp16_0, tensor x_329_split_cast_fp16_1 = split(axis = x_329_split_axis_0, num_splits = x_329_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_329_split_cast_fp16")]; tensor x_329_split_1_sigmoid_cast_fp16 = sigmoid(x = x_329_split_cast_fp16_1)[name = tensor("x_329_split_1_sigmoid_cast_fp16")]; tensor x_329_cast_fp16 = mul(x = x_329_split_cast_fp16_0, y = x_329_split_1_sigmoid_cast_fp16)[name = tensor("x_329_cast_fp16")]; tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_329_cast_fp16, cond = var_337)[name = tensor("input_773_cast_fp16")]; tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; tensor const_157_to_fp16 = const()[name = tensor("const_157_to_fp16"), val = tensor(0x0p+0)]; tensor input_775_cast_fp16 = pad(constant_val = const_157_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(1024)]; tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; tensor const_276_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_276_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362762752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362773120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362772032)))]; tensor const_277_to_fp16 = const()[name = tensor("const_277_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362775232)))]; tensor input_779_cast_fp16 = conv(bias = const_277_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_276_to_fp16_quantized, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; tensor x_331_pad_type_0 = const()[name = tensor("x_331_pad_type_0"), val = tensor("valid")]; tensor x_331_strides_0 = const()[name = tensor("x_331_strides_0"), val = tensor([1])]; tensor x_331_pad_0 = const()[name = tensor("x_331_pad_0"), val = tensor([0, 0])]; tensor x_331_dilations_0 = const()[name = tensor("x_331_dilations_0"), val = tensor([1])]; tensor x_331_groups_0 = const()[name = tensor("x_331_groups_0"), val = tensor(1)]; tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362777344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363827072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363825984)))]; tensor module_layers_14_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363829184)))]; tensor x_331_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv2_bias_to_fp16, dilations = x_331_dilations_0, groups = x_331_groups_0, pad = x_331_pad_0, pad_type = x_331_pad_type_0, strides = x_331_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_781_cast_fp16)[name = tensor("x_331_cast_fp16")]; tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_331_cast_fp16)[name = tensor("transpose_208")]; tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363831296)))]; tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363833408)))]; tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; tensor module_layers_14_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363835520))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368034048))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368029888)))]; tensor module_layers_14_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368042304)))]; tensor linear_134_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear1_bias_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16_quantized, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; tensor module_layers_14_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368050560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372246016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372244928)))]; tensor module_layers_14_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372248128)))]; tensor linear_135_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear2_bias_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16_quantized, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; tensor var_2815_to_fp16 = const()[name = tensor("op_2815_to_fp16"), val = tensor(0x1p-1)]; tensor var_2816_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2815_to_fp16)[name = tensor("op_2816_cast_fp16")]; tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2816_cast_fp16)[name = tensor("input_797_cast_fp16")]; tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372250240)))]; tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372252352)))]; tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372254464)))]; tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372256576)))]; tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; tensor module_layers_15_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372258688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(376457216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(376453056)))]; tensor module_layers_15_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(376465472)))]; tensor linear_136_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear1_bias_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16_quantized, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; tensor module_layers_15_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(376473728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380669184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380668096)))]; tensor module_layers_15_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380671296)))]; tensor linear_137_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear2_bias_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16_quantized, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; tensor var_2846_to_fp16 = const()[name = tensor("op_2846_to_fp16"), val = tensor(0x1p-1)]; tensor var_2847_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2846_to_fp16)[name = tensor("op_2847_cast_fp16")]; tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2847_cast_fp16)[name = tensor("input_811_cast_fp16")]; tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380673408)))]; tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380675520)))]; tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; tensor module_layers_15_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380677632))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381727360))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381726272)))]; tensor module_layers_15_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381729472)))]; tensor linear_138_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_q_bias_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16_quantized, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; tensor var_2864 = const()[name = tensor("op_2864"), val = tensor([1, -1, 8, 128])]; tensor q_91_cast_fp16 = reshape(shape = var_2864, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; tensor module_layers_15_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381731584))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382781312))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382780224)))]; tensor module_layers_15_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382783424)))]; tensor linear_139_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_k_bias_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16_quantized, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; tensor var_2869 = const()[name = tensor("op_2869"), val = tensor([1, -1, 8, 128])]; tensor k_61_cast_fp16 = reshape(shape = var_2869, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; tensor module_layers_15_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382785536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383835264))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383834176)))]; tensor module_layers_15_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383837376)))]; tensor linear_140_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_v_bias_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16_quantized, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; tensor var_2874 = const()[name = tensor("op_2874"), val = tensor([1, -1, 8, 128])]; tensor v_31_cast_fp16 = reshape(shape = var_2874, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383839488)))]; tensor var_2886_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2886_cast_fp16")]; tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383841600)))]; tensor var_2888_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2888_cast_fp16")]; tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_339_transpose_x_0 = const()[name = tensor("x_339_transpose_x_0"), val = tensor(false)]; tensor x_339_transpose_y_0 = const()[name = tensor("x_339_transpose_y_0"), val = tensor(false)]; tensor op_2890_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2890_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383843712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384228224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384227776)))]; tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_2888_cast_fp16)[name = tensor("transpose_207")]; tensor x_339_cast_fp16 = matmul(transpose_x = x_339_transpose_x_0, transpose_y = x_339_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = op_2890_to_fp16_quantized)[name = tensor("x_339_cast_fp16")]; tensor x_341_pad_0 = const()[name = tensor("x_341_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_341_mode_0 = const()[name = tensor("x_341_mode_0"), val = tensor("constant")]; tensor const_164_to_fp16 = const()[name = tensor("const_164_to_fp16"), val = tensor(0x0p+0)]; tensor x_341_cast_fp16 = pad(constant_val = const_164_to_fp16, mode = x_341_mode_0, pad = x_341_pad_0, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; tensor var_2898 = const()[name = tensor("op_2898"), val = tensor([1, 8, -1, 188])]; tensor x_343_cast_fp16 = reshape(shape = var_2898, x = x_341_cast_fp16)[name = tensor("x_343_cast_fp16")]; tensor var_2902_begin_0 = const()[name = tensor("op_2902_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2902_end_0 = const()[name = tensor("op_2902_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2902_end_mask_0 = const()[name = tensor("op_2902_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2902_cast_fp16 = slice_by_index(begin = var_2902_begin_0, end = var_2902_end_0, end_mask = var_2902_end_mask_0, x = x_343_cast_fp16)[name = tensor("op_2902_cast_fp16")]; tensor var_2903 = const()[name = tensor("op_2903"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_61_cast_fp16 = reshape(shape = var_2903, x = var_2902_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; tensor transpose_126_perm_0 = const()[name = tensor("transpose_126_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_127_perm_0 = const()[name = tensor("transpose_127_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_127 = transpose(perm = transpose_127_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_205")]; tensor transpose_126 = transpose(perm = transpose_126_perm_0, x = var_2886_cast_fp16)[name = tensor("transpose_206")]; tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_126, y = transpose_127)[name = tensor("matrix_ac_31_cast_fp16")]; tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; tensor var_2912_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_2912_cast_fp16")]; tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_61_cast_fp16 = mul(x = var_2912_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_3)[name = tensor("scores_63_cast_fp16")]; tensor var_2918_cast_fp16 = softmax(axis = var_30, x = scores_63_cast_fp16)[name = tensor("op_2918_cast_fp16")]; tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_2918_cast_fp16, cond = mask_3)[name = tensor("input_813_cast_fp16")]; tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_204")]; tensor x_345_cast_fp16 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_345_cast_fp16")]; tensor var_2922_perm_0 = const()[name = tensor("op_2922_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2923 = const()[name = tensor("op_2923"), val = tensor([1, -1, 1024])]; tensor var_2922_cast_fp16 = transpose(perm = var_2922_perm_0, x = x_345_cast_fp16)[name = tensor("transpose_203")]; tensor input_815_cast_fp16 = reshape(shape = var_2923, x = var_2922_cast_fp16)[name = tensor("input_815_cast_fp16")]; tensor module_layers_15_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384229056))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(385278784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(385277696)))]; tensor module_layers_15_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(385280896)))]; tensor linear_142_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_out_bias_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16_quantized, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; tensor x_349_axes_0 = const()[name = tensor("x_349_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(385283008)))]; tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(385285120)))]; tensor x_349_cast_fp16 = layer_norm(axes = x_349_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_349_cast_fp16")]; tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(385287232))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387386560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387384448)))]; tensor module_layers_15_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387390720)))]; tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_349_cast_fp16)[name = tensor("transpose_202")]; tensor input_823_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv1_bias_to_fp16, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; tensor x_351_split_num_splits_0 = const()[name = tensor("x_351_split_num_splits_0"), val = tensor(2)]; tensor x_351_split_axis_0 = const()[name = tensor("x_351_split_axis_0"), val = tensor(1)]; tensor x_351_split_cast_fp16_0, tensor x_351_split_cast_fp16_1 = split(axis = x_351_split_axis_0, num_splits = x_351_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_351_split_cast_fp16")]; tensor x_351_split_1_sigmoid_cast_fp16 = sigmoid(x = x_351_split_cast_fp16_1)[name = tensor("x_351_split_1_sigmoid_cast_fp16")]; tensor x_351_cast_fp16 = mul(x = x_351_split_cast_fp16_0, y = x_351_split_1_sigmoid_cast_fp16)[name = tensor("x_351_cast_fp16")]; tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_351_cast_fp16, cond = var_337)[name = tensor("input_825_cast_fp16")]; tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; tensor const_167_to_fp16 = const()[name = tensor("const_167_to_fp16"), val = tensor(0x0p+0)]; tensor input_827_cast_fp16 = pad(constant_val = const_167_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(1024)]; tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; tensor const_278_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_278_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387394880))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387405248))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387404160)))]; tensor const_279_to_fp16 = const()[name = tensor("const_279_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387407360)))]; tensor input_831_cast_fp16 = conv(bias = const_279_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_278_to_fp16_quantized, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; tensor x_353_pad_type_0 = const()[name = tensor("x_353_pad_type_0"), val = tensor("valid")]; tensor x_353_strides_0 = const()[name = tensor("x_353_strides_0"), val = tensor([1])]; tensor x_353_pad_0 = const()[name = tensor("x_353_pad_0"), val = tensor([0, 0])]; tensor x_353_dilations_0 = const()[name = tensor("x_353_dilations_0"), val = tensor([1])]; tensor x_353_groups_0 = const()[name = tensor("x_353_groups_0"), val = tensor(1)]; tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387409472))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388459200))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388458112)))]; tensor module_layers_15_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388461312)))]; tensor x_353_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv2_bias_to_fp16, dilations = x_353_dilations_0, groups = x_353_groups_0, pad = x_353_pad_0, pad_type = x_353_pad_type_0, strides = x_353_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_833_cast_fp16)[name = tensor("x_353_cast_fp16")]; tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_353_cast_fp16)[name = tensor("transpose_201")]; tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388463424)))]; tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388465536)))]; tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; tensor module_layers_15_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388467648))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392666176))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392662016)))]; tensor module_layers_15_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392674432)))]; tensor linear_143_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear1_bias_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16_quantized, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; tensor module_layers_15_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392682688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396878144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396877056)))]; tensor module_layers_15_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396880256)))]; tensor linear_144_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear2_bias_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16_quantized, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; tensor var_2989_to_fp16 = const()[name = tensor("op_2989_to_fp16"), val = tensor(0x1p-1)]; tensor var_2990_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_2989_to_fp16)[name = tensor("op_2990_cast_fp16")]; tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_2990_cast_fp16)[name = tensor("input_849_cast_fp16")]; tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396882368)))]; tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396884480)))]; tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396886592)))]; tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396888704)))]; tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; tensor module_layers_16_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396890816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401089344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401085184)))]; tensor module_layers_16_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401097600)))]; tensor linear_145_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear1_bias_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16_quantized, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; tensor module_layers_16_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401105856))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405301312))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405300224)))]; tensor module_layers_16_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405303424)))]; tensor linear_146_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear2_bias_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16_quantized, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; tensor var_3020_to_fp16 = const()[name = tensor("op_3020_to_fp16"), val = tensor(0x1p-1)]; tensor var_3021_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3020_to_fp16)[name = tensor("op_3021_cast_fp16")]; tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_3021_cast_fp16)[name = tensor("input_863_cast_fp16")]; tensor query_33_axes_0 = const()[name = tensor("query_33_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405305536)))]; tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405307648)))]; tensor query_33_cast_fp16 = layer_norm(axes = query_33_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_33_cast_fp16")]; tensor module_layers_16_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405309760))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(406359488))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(406358400)))]; tensor module_layers_16_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(406361600)))]; tensor linear_147_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_q_bias_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16_quantized, x = query_33_cast_fp16)[name = tensor("linear_147_cast_fp16")]; tensor var_3038 = const()[name = tensor("op_3038"), val = tensor([1, -1, 8, 128])]; tensor q_97_cast_fp16 = reshape(shape = var_3038, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; tensor module_layers_16_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(406363712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407413440))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407412352)))]; tensor module_layers_16_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407415552)))]; tensor linear_148_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_k_bias_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16_quantized, x = query_33_cast_fp16)[name = tensor("linear_148_cast_fp16")]; tensor var_3043 = const()[name = tensor("op_3043"), val = tensor([1, -1, 8, 128])]; tensor k_65_cast_fp16 = reshape(shape = var_3043, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; tensor module_layers_16_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407417664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408467392))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408466304)))]; tensor module_layers_16_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408469504)))]; tensor linear_149_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_v_bias_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16_quantized, x = query_33_cast_fp16)[name = tensor("linear_149_cast_fp16")]; tensor var_3048 = const()[name = tensor("op_3048"), val = tensor([1, -1, 8, 128])]; tensor v_33_cast_fp16 = reshape(shape = var_3048, x = linear_149_cast_fp16)[name = tensor("v_33_cast_fp16")]; tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408471616)))]; tensor var_3060_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3060_cast_fp16")]; tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408473728)))]; tensor var_3062_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3062_cast_fp16")]; tensor q_with_bias_v_33_perm_0 = const()[name = tensor("q_with_bias_v_33_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_361_transpose_x_0 = const()[name = tensor("x_361_transpose_x_0"), val = tensor(false)]; tensor x_361_transpose_y_0 = const()[name = tensor("x_361_transpose_y_0"), val = tensor(false)]; tensor op_3064_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3064_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408475840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408860352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408859904)))]; tensor q_with_bias_v_33_cast_fp16 = transpose(perm = q_with_bias_v_33_perm_0, x = var_3062_cast_fp16)[name = tensor("transpose_200")]; tensor x_361_cast_fp16 = matmul(transpose_x = x_361_transpose_x_0, transpose_y = x_361_transpose_y_0, x = q_with_bias_v_33_cast_fp16, y = op_3064_to_fp16_quantized)[name = tensor("x_361_cast_fp16")]; tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_363_mode_0 = const()[name = tensor("x_363_mode_0"), val = tensor("constant")]; tensor const_174_to_fp16 = const()[name = tensor("const_174_to_fp16"), val = tensor(0x0p+0)]; tensor x_363_cast_fp16 = pad(constant_val = const_174_to_fp16, mode = x_363_mode_0, pad = x_363_pad_0, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; tensor var_3072 = const()[name = tensor("op_3072"), val = tensor([1, 8, -1, 188])]; tensor x_365_cast_fp16 = reshape(shape = var_3072, x = x_363_cast_fp16)[name = tensor("x_365_cast_fp16")]; tensor var_3076_begin_0 = const()[name = tensor("op_3076_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3076_end_0 = const()[name = tensor("op_3076_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3076_end_mask_0 = const()[name = tensor("op_3076_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3076_cast_fp16 = slice_by_index(begin = var_3076_begin_0, end = var_3076_end_0, end_mask = var_3076_end_mask_0, x = x_365_cast_fp16)[name = tensor("op_3076_cast_fp16")]; tensor var_3077 = const()[name = tensor("op_3077"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3077, x = var_3076_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; tensor matrix_ac_33_transpose_x_0 = const()[name = tensor("matrix_ac_33_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_33_transpose_y_0 = const()[name = tensor("matrix_ac_33_transpose_y_0"), val = tensor(false)]; tensor transpose_128_perm_0 = const()[name = tensor("transpose_128_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_129_perm_0 = const()[name = tensor("transpose_129_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_129 = transpose(perm = transpose_129_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_198")]; tensor transpose_128 = transpose(perm = transpose_128_perm_0, x = var_3060_cast_fp16)[name = tensor("transpose_199")]; tensor matrix_ac_33_cast_fp16 = matmul(transpose_x = matrix_ac_33_transpose_x_0, transpose_y = matrix_ac_33_transpose_y_0, x = transpose_128, y = transpose_129)[name = tensor("matrix_ac_33_cast_fp16")]; tensor matrix_bd_67_begin_0 = const()[name = tensor("matrix_bd_67_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_67_end_0 = const()[name = tensor("matrix_bd_67_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_67_end_mask_0 = const()[name = tensor("matrix_bd_67_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_67_cast_fp16 = slice_by_index(begin = matrix_bd_67_begin_0, end = matrix_bd_67_end_0, end_mask = matrix_bd_67_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_67_cast_fp16")]; tensor var_3086_cast_fp16 = add(x = matrix_ac_33_cast_fp16, y = matrix_bd_67_cast_fp16)[name = tensor("op_3086_cast_fp16")]; tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_65_cast_fp16 = mul(x = var_3086_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; tensor scores_67_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_3)[name = tensor("scores_67_cast_fp16")]; tensor var_3092_cast_fp16 = softmax(axis = var_30, x = scores_67_cast_fp16)[name = tensor("op_3092_cast_fp16")]; tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_3092_cast_fp16, cond = mask_3)[name = tensor("input_865_cast_fp16")]; tensor x_367_transpose_x_0 = const()[name = tensor("x_367_transpose_x_0"), val = tensor(false)]; tensor x_367_transpose_y_0 = const()[name = tensor("x_367_transpose_y_0"), val = tensor(false)]; tensor value_35_cast_fp16 = transpose(perm = value_35_perm_0, x = v_33_cast_fp16)[name = tensor("transpose_197")]; tensor x_367_cast_fp16 = matmul(transpose_x = x_367_transpose_x_0, transpose_y = x_367_transpose_y_0, x = input_865_cast_fp16, y = value_35_cast_fp16)[name = tensor("x_367_cast_fp16")]; tensor var_3096_perm_0 = const()[name = tensor("op_3096_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3097 = const()[name = tensor("op_3097"), val = tensor([1, -1, 1024])]; tensor var_3096_cast_fp16 = transpose(perm = var_3096_perm_0, x = x_367_cast_fp16)[name = tensor("transpose_196")]; tensor input_867_cast_fp16 = reshape(shape = var_3097, x = var_3096_cast_fp16)[name = tensor("input_867_cast_fp16")]; tensor module_layers_16_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408861184))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409910912))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409909824)))]; tensor module_layers_16_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409913024)))]; tensor linear_151_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_out_bias_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16_quantized, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; tensor x_371_axes_0 = const()[name = tensor("x_371_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409915136)))]; tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409917248)))]; tensor x_371_cast_fp16 = layer_norm(axes = x_371_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_371_cast_fp16")]; tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409919360))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412018688))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412016576)))]; tensor module_layers_16_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412022848)))]; tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_371_cast_fp16)[name = tensor("transpose_195")]; tensor input_875_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv1_bias_to_fp16, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; tensor x_373_split_num_splits_0 = const()[name = tensor("x_373_split_num_splits_0"), val = tensor(2)]; tensor x_373_split_axis_0 = const()[name = tensor("x_373_split_axis_0"), val = tensor(1)]; tensor x_373_split_cast_fp16_0, tensor x_373_split_cast_fp16_1 = split(axis = x_373_split_axis_0, num_splits = x_373_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_373_split_cast_fp16")]; tensor x_373_split_1_sigmoid_cast_fp16 = sigmoid(x = x_373_split_cast_fp16_1)[name = tensor("x_373_split_1_sigmoid_cast_fp16")]; tensor x_373_cast_fp16 = mul(x = x_373_split_cast_fp16_0, y = x_373_split_1_sigmoid_cast_fp16)[name = tensor("x_373_cast_fp16")]; tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_373_cast_fp16, cond = var_337)[name = tensor("input_877_cast_fp16")]; tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; tensor const_177_to_fp16 = const()[name = tensor("const_177_to_fp16"), val = tensor(0x0p+0)]; tensor input_879_cast_fp16 = pad(constant_val = const_177_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(1024)]; tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; tensor const_280_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_280_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412027008))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412037376))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412036288)))]; tensor const_281_to_fp16 = const()[name = tensor("const_281_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412039488)))]; tensor input_883_cast_fp16 = conv(bias = const_281_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_280_to_fp16_quantized, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; tensor x_375_pad_type_0 = const()[name = tensor("x_375_pad_type_0"), val = tensor("valid")]; tensor x_375_strides_0 = const()[name = tensor("x_375_strides_0"), val = tensor([1])]; tensor x_375_pad_0 = const()[name = tensor("x_375_pad_0"), val = tensor([0, 0])]; tensor x_375_dilations_0 = const()[name = tensor("x_375_dilations_0"), val = tensor([1])]; tensor x_375_groups_0 = const()[name = tensor("x_375_groups_0"), val = tensor(1)]; tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412041600))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413091328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413090240)))]; tensor module_layers_16_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413093440)))]; tensor x_375_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv2_bias_to_fp16, dilations = x_375_dilations_0, groups = x_375_groups_0, pad = x_375_pad_0, pad_type = x_375_pad_type_0, strides = x_375_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_885_cast_fp16)[name = tensor("x_375_cast_fp16")]; tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_375_cast_fp16)[name = tensor("transpose_194")]; tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413095552)))]; tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413097664)))]; tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; tensor module_layers_16_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413099776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417298304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417294144)))]; tensor module_layers_16_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417306560)))]; tensor linear_152_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear1_bias_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16_quantized, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; tensor module_layers_16_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417314816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421510272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421509184)))]; tensor module_layers_16_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421512384)))]; tensor linear_153_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear2_bias_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16_quantized, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; tensor var_3163_to_fp16 = const()[name = tensor("op_3163_to_fp16"), val = tensor(0x1p-1)]; tensor var_3164_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3163_to_fp16)[name = tensor("op_3164_cast_fp16")]; tensor input_901_cast_fp16 = add(x = input_889_cast_fp16, y = var_3164_cast_fp16)[name = tensor("input_901_cast_fp16")]; tensor input_903_axes_0 = const()[name = tensor("input_903_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421514496)))]; tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421516608)))]; tensor input_903_cast_fp16 = layer_norm(axes = input_903_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("input_903_cast_fp16")]; tensor input_905_axes_0 = const()[name = tensor("input_905_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421518720)))]; tensor module_layers_17_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421520832)))]; tensor input_905_cast_fp16 = layer_norm(axes = input_905_axes_0, beta = module_layers_17_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward1_weight_to_fp16, x = input_903_cast_fp16)[name = tensor("input_905_cast_fp16")]; tensor module_layers_17_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421522944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425721472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425717312)))]; tensor module_layers_17_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_17_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425729728)))]; tensor linear_154_cast_fp16 = linear(bias = module_layers_17_feed_forward1_linear1_bias_to_fp16, weight = module_layers_17_feed_forward1_linear1_weight_to_fp16_quantized, x = input_905_cast_fp16)[name = tensor("linear_154_cast_fp16")]; tensor input_909_cast_fp16 = silu(x = linear_154_cast_fp16)[name = tensor("input_909_cast_fp16")]; tensor module_layers_17_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425737984))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429933440))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429932352)))]; tensor module_layers_17_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_17_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429935552)))]; tensor linear_155_cast_fp16 = linear(bias = module_layers_17_feed_forward1_linear2_bias_to_fp16, weight = module_layers_17_feed_forward1_linear2_weight_to_fp16_quantized, x = input_909_cast_fp16)[name = tensor("linear_155_cast_fp16")]; tensor var_3194_to_fp16 = const()[name = tensor("op_3194_to_fp16"), val = tensor(0x1p-1)]; tensor var_3195_cast_fp16 = mul(x = linear_155_cast_fp16, y = var_3194_to_fp16)[name = tensor("op_3195_cast_fp16")]; tensor input_915_cast_fp16 = add(x = input_903_cast_fp16, y = var_3195_cast_fp16)[name = tensor("input_915_cast_fp16")]; tensor query_35_axes_0 = const()[name = tensor("query_35_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429937664)))]; tensor module_layers_17_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429939776)))]; tensor query_35_cast_fp16 = layer_norm(axes = query_35_axes_0, beta = module_layers_17_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_self_att_weight_to_fp16, x = input_915_cast_fp16)[name = tensor("query_35_cast_fp16")]; tensor module_layers_17_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429941888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430991616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430990528)))]; tensor module_layers_17_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430993728)))]; tensor linear_156_cast_fp16 = linear(bias = module_layers_17_self_attn_linear_q_bias_to_fp16, weight = module_layers_17_self_attn_linear_q_weight_to_fp16_quantized, x = query_35_cast_fp16)[name = tensor("linear_156_cast_fp16")]; tensor var_3212 = const()[name = tensor("op_3212"), val = tensor([1, -1, 8, 128])]; tensor q_103_cast_fp16 = reshape(shape = var_3212, x = linear_156_cast_fp16)[name = tensor("q_103_cast_fp16")]; tensor module_layers_17_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430995840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432045568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432044480)))]; tensor module_layers_17_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432047680)))]; tensor linear_157_cast_fp16 = linear(bias = module_layers_17_self_attn_linear_k_bias_to_fp16, weight = module_layers_17_self_attn_linear_k_weight_to_fp16_quantized, x = query_35_cast_fp16)[name = tensor("linear_157_cast_fp16")]; tensor var_3217 = const()[name = tensor("op_3217"), val = tensor([1, -1, 8, 128])]; tensor k_69_cast_fp16 = reshape(shape = var_3217, x = linear_157_cast_fp16)[name = tensor("k_69_cast_fp16")]; tensor module_layers_17_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432049792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433099520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433098432)))]; tensor module_layers_17_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433101632)))]; tensor linear_158_cast_fp16 = linear(bias = module_layers_17_self_attn_linear_v_bias_to_fp16, weight = module_layers_17_self_attn_linear_v_weight_to_fp16_quantized, x = query_35_cast_fp16)[name = tensor("linear_158_cast_fp16")]; tensor var_3222 = const()[name = tensor("op_3222"), val = tensor([1, -1, 8, 128])]; tensor v_35_cast_fp16 = reshape(shape = var_3222, x = linear_158_cast_fp16)[name = tensor("v_35_cast_fp16")]; tensor value_37_perm_0 = const()[name = tensor("value_37_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_17_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433103744)))]; tensor var_3234_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3234_cast_fp16")]; tensor module_layers_17_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433105856)))]; tensor var_3236_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3236_cast_fp16")]; tensor q_with_bias_v_35_perm_0 = const()[name = tensor("q_with_bias_v_35_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_383_transpose_x_0 = const()[name = tensor("x_383_transpose_x_0"), val = tensor(false)]; tensor x_383_transpose_y_0 = const()[name = tensor("x_383_transpose_y_0"), val = tensor(false)]; tensor op_3238_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3238_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433107968))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433492480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433492032)))]; tensor q_with_bias_v_35_cast_fp16 = transpose(perm = q_with_bias_v_35_perm_0, x = var_3236_cast_fp16)[name = tensor("transpose_193")]; tensor x_383_cast_fp16 = matmul(transpose_x = x_383_transpose_x_0, transpose_y = x_383_transpose_y_0, x = q_with_bias_v_35_cast_fp16, y = op_3238_to_fp16_quantized)[name = tensor("x_383_cast_fp16")]; tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_385_mode_0 = const()[name = tensor("x_385_mode_0"), val = tensor("constant")]; tensor const_184_to_fp16 = const()[name = tensor("const_184_to_fp16"), val = tensor(0x0p+0)]; tensor x_385_cast_fp16 = pad(constant_val = const_184_to_fp16, mode = x_385_mode_0, pad = x_385_pad_0, x = x_383_cast_fp16)[name = tensor("x_385_cast_fp16")]; tensor var_3246 = const()[name = tensor("op_3246"), val = tensor([1, 8, -1, 188])]; tensor x_387_cast_fp16 = reshape(shape = var_3246, x = x_385_cast_fp16)[name = tensor("x_387_cast_fp16")]; tensor var_3250_begin_0 = const()[name = tensor("op_3250_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3250_end_0 = const()[name = tensor("op_3250_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3250_end_mask_0 = const()[name = tensor("op_3250_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3250_cast_fp16 = slice_by_index(begin = var_3250_begin_0, end = var_3250_end_0, end_mask = var_3250_end_mask_0, x = x_387_cast_fp16)[name = tensor("op_3250_cast_fp16")]; tensor var_3251 = const()[name = tensor("op_3251"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_69_cast_fp16 = reshape(shape = var_3251, x = var_3250_cast_fp16)[name = tensor("matrix_bd_69_cast_fp16")]; tensor matrix_ac_35_transpose_x_0 = const()[name = tensor("matrix_ac_35_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_35_transpose_y_0 = const()[name = tensor("matrix_ac_35_transpose_y_0"), val = tensor(false)]; tensor transpose_130_perm_0 = const()[name = tensor("transpose_130_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_131_perm_0 = const()[name = tensor("transpose_131_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_131 = transpose(perm = transpose_131_perm_0, x = k_69_cast_fp16)[name = tensor("transpose_191")]; tensor transpose_130 = transpose(perm = transpose_130_perm_0, x = var_3234_cast_fp16)[name = tensor("transpose_192")]; tensor matrix_ac_35_cast_fp16 = matmul(transpose_x = matrix_ac_35_transpose_x_0, transpose_y = matrix_ac_35_transpose_y_0, x = transpose_130, y = transpose_131)[name = tensor("matrix_ac_35_cast_fp16")]; tensor matrix_bd_71_begin_0 = const()[name = tensor("matrix_bd_71_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_71_end_0 = const()[name = tensor("matrix_bd_71_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_71_end_mask_0 = const()[name = tensor("matrix_bd_71_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_71_cast_fp16 = slice_by_index(begin = matrix_bd_71_begin_0, end = matrix_bd_71_end_0, end_mask = matrix_bd_71_end_mask_0, x = matrix_bd_69_cast_fp16)[name = tensor("matrix_bd_71_cast_fp16")]; tensor var_3260_cast_fp16 = add(x = matrix_ac_35_cast_fp16, y = matrix_bd_71_cast_fp16)[name = tensor("op_3260_cast_fp16")]; tensor _inversed_scores_69_y_0_to_fp16 = const()[name = tensor("_inversed_scores_69_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_69_cast_fp16 = mul(x = var_3260_cast_fp16, y = _inversed_scores_69_y_0_to_fp16)[name = tensor("_inversed_scores_69_cast_fp16")]; tensor scores_71_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_69_cast_fp16, cond = mask_3)[name = tensor("scores_71_cast_fp16")]; tensor var_3266_cast_fp16 = softmax(axis = var_30, x = scores_71_cast_fp16)[name = tensor("op_3266_cast_fp16")]; tensor input_917_cast_fp16 = select(a = var_11_to_fp16, b = var_3266_cast_fp16, cond = mask_3)[name = tensor("input_917_cast_fp16")]; tensor x_389_transpose_x_0 = const()[name = tensor("x_389_transpose_x_0"), val = tensor(false)]; tensor x_389_transpose_y_0 = const()[name = tensor("x_389_transpose_y_0"), val = tensor(false)]; tensor value_37_cast_fp16 = transpose(perm = value_37_perm_0, x = v_35_cast_fp16)[name = tensor("transpose_190")]; tensor x_389_cast_fp16 = matmul(transpose_x = x_389_transpose_x_0, transpose_y = x_389_transpose_y_0, x = input_917_cast_fp16, y = value_37_cast_fp16)[name = tensor("x_389_cast_fp16")]; tensor var_3270_perm_0 = const()[name = tensor("op_3270_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3271 = const()[name = tensor("op_3271"), val = tensor([1, -1, 1024])]; tensor var_3270_cast_fp16 = transpose(perm = var_3270_perm_0, x = x_389_cast_fp16)[name = tensor("transpose_189")]; tensor input_919_cast_fp16 = reshape(shape = var_3271, x = var_3270_cast_fp16)[name = tensor("input_919_cast_fp16")]; tensor module_layers_17_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433493312))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434543040))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434541952)))]; tensor module_layers_17_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434545152)))]; tensor linear_160_cast_fp16 = linear(bias = module_layers_17_self_attn_linear_out_bias_to_fp16, weight = module_layers_17_self_attn_linear_out_weight_to_fp16_quantized, x = input_919_cast_fp16)[name = tensor("linear_160_cast_fp16")]; tensor input_923_cast_fp16 = add(x = input_915_cast_fp16, y = linear_160_cast_fp16)[name = tensor("input_923_cast_fp16")]; tensor x_393_axes_0 = const()[name = tensor("x_393_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434547264)))]; tensor module_layers_17_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434549376)))]; tensor x_393_cast_fp16 = layer_norm(axes = x_393_axes_0, beta = module_layers_17_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_conv_weight_to_fp16, x = input_923_cast_fp16)[name = tensor("x_393_cast_fp16")]; tensor input_925_perm_0 = const()[name = tensor("input_925_perm_0"), val = tensor([0, 2, 1])]; tensor input_927_pad_type_0 = const()[name = tensor("input_927_pad_type_0"), val = tensor("valid")]; tensor input_927_strides_0 = const()[name = tensor("input_927_strides_0"), val = tensor([1])]; tensor input_927_pad_0 = const()[name = tensor("input_927_pad_0"), val = tensor([0, 0])]; tensor input_927_dilations_0 = const()[name = tensor("input_927_dilations_0"), val = tensor([1])]; tensor input_927_groups_0 = const()[name = tensor("input_927_groups_0"), val = tensor(1)]; tensor module_layers_17_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434551488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436650816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436648704)))]; tensor module_layers_17_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_17_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436654976)))]; tensor input_925_cast_fp16 = transpose(perm = input_925_perm_0, x = x_393_cast_fp16)[name = tensor("transpose_188")]; tensor input_927_cast_fp16 = conv(bias = module_layers_17_conv_pointwise_conv1_bias_to_fp16, dilations = input_927_dilations_0, groups = input_927_groups_0, pad = input_927_pad_0, pad_type = input_927_pad_type_0, strides = input_927_strides_0, weight = module_layers_17_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_925_cast_fp16)[name = tensor("input_927_cast_fp16")]; tensor x_395_split_num_splits_0 = const()[name = tensor("x_395_split_num_splits_0"), val = tensor(2)]; tensor x_395_split_axis_0 = const()[name = tensor("x_395_split_axis_0"), val = tensor(1)]; tensor x_395_split_cast_fp16_0, tensor x_395_split_cast_fp16_1 = split(axis = x_395_split_axis_0, num_splits = x_395_split_num_splits_0, x = input_927_cast_fp16)[name = tensor("x_395_split_cast_fp16")]; tensor x_395_split_1_sigmoid_cast_fp16 = sigmoid(x = x_395_split_cast_fp16_1)[name = tensor("x_395_split_1_sigmoid_cast_fp16")]; tensor x_395_cast_fp16 = mul(x = x_395_split_cast_fp16_0, y = x_395_split_1_sigmoid_cast_fp16)[name = tensor("x_395_cast_fp16")]; tensor input_929_cast_fp16 = select(a = var_11_to_fp16, b = x_395_cast_fp16, cond = var_337)[name = tensor("input_929_cast_fp16")]; tensor input_931_pad_0 = const()[name = tensor("input_931_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_931_mode_0 = const()[name = tensor("input_931_mode_0"), val = tensor("constant")]; tensor const_187_to_fp16 = const()[name = tensor("const_187_to_fp16"), val = tensor(0x0p+0)]; tensor input_931_cast_fp16 = pad(constant_val = const_187_to_fp16, mode = input_931_mode_0, pad = input_931_pad_0, x = input_929_cast_fp16)[name = tensor("input_931_cast_fp16")]; tensor input_933_pad_type_0 = const()[name = tensor("input_933_pad_type_0"), val = tensor("valid")]; tensor input_933_groups_0 = const()[name = tensor("input_933_groups_0"), val = tensor(1024)]; tensor input_933_strides_0 = const()[name = tensor("input_933_strides_0"), val = tensor([1])]; tensor input_933_pad_0 = const()[name = tensor("input_933_pad_0"), val = tensor([0, 0])]; tensor input_933_dilations_0 = const()[name = tensor("input_933_dilations_0"), val = tensor([1])]; tensor const_282_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_282_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436659136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436669504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436668416)))]; tensor const_283_to_fp16 = const()[name = tensor("const_283_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436671616)))]; tensor input_935_cast_fp16 = conv(bias = const_283_to_fp16, dilations = input_933_dilations_0, groups = input_933_groups_0, pad = input_933_pad_0, pad_type = input_933_pad_type_0, strides = input_933_strides_0, weight = const_282_to_fp16_quantized, x = input_931_cast_fp16)[name = tensor("input_935_cast_fp16")]; tensor input_937_cast_fp16 = silu(x = input_935_cast_fp16)[name = tensor("input_937_cast_fp16")]; tensor x_397_pad_type_0 = const()[name = tensor("x_397_pad_type_0"), val = tensor("valid")]; tensor x_397_strides_0 = const()[name = tensor("x_397_strides_0"), val = tensor([1])]; tensor x_397_pad_0 = const()[name = tensor("x_397_pad_0"), val = tensor([0, 0])]; tensor x_397_dilations_0 = const()[name = tensor("x_397_dilations_0"), val = tensor([1])]; tensor x_397_groups_0 = const()[name = tensor("x_397_groups_0"), val = tensor(1)]; tensor module_layers_17_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436673728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437723456))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437722368)))]; tensor module_layers_17_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_17_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437725568)))]; tensor x_397_cast_fp16 = conv(bias = module_layers_17_conv_pointwise_conv2_bias_to_fp16, dilations = x_397_dilations_0, groups = x_397_groups_0, pad = x_397_pad_0, pad_type = x_397_pad_type_0, strides = x_397_strides_0, weight = module_layers_17_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_937_cast_fp16)[name = tensor("x_397_cast_fp16")]; tensor input_939_perm_0 = const()[name = tensor("input_939_perm_0"), val = tensor([0, 2, 1])]; tensor input_939_cast_fp16 = transpose(perm = input_939_perm_0, x = x_397_cast_fp16)[name = tensor("transpose_187")]; tensor input_941_cast_fp16 = add(x = input_923_cast_fp16, y = input_939_cast_fp16)[name = tensor("input_941_cast_fp16")]; tensor input_943_axes_0 = const()[name = tensor("input_943_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437727680)))]; tensor module_layers_17_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437729792)))]; tensor input_943_cast_fp16 = layer_norm(axes = input_943_axes_0, beta = module_layers_17_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward2_weight_to_fp16, x = input_941_cast_fp16)[name = tensor("input_943_cast_fp16")]; tensor module_layers_17_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437731904))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441930432))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441926272)))]; tensor module_layers_17_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_17_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441938688)))]; tensor linear_161_cast_fp16 = linear(bias = module_layers_17_feed_forward2_linear1_bias_to_fp16, weight = module_layers_17_feed_forward2_linear1_weight_to_fp16_quantized, x = input_943_cast_fp16)[name = tensor("linear_161_cast_fp16")]; tensor input_947_cast_fp16 = silu(x = linear_161_cast_fp16)[name = tensor("input_947_cast_fp16")]; tensor module_layers_17_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441946944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446142400))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446141312)))]; tensor module_layers_17_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_17_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446144512)))]; tensor linear_162_cast_fp16 = linear(bias = module_layers_17_feed_forward2_linear2_bias_to_fp16, weight = module_layers_17_feed_forward2_linear2_weight_to_fp16_quantized, x = input_947_cast_fp16)[name = tensor("linear_162_cast_fp16")]; tensor var_3337_to_fp16 = const()[name = tensor("op_3337_to_fp16"), val = tensor(0x1p-1)]; tensor var_3338_cast_fp16 = mul(x = linear_162_cast_fp16, y = var_3337_to_fp16)[name = tensor("op_3338_cast_fp16")]; tensor input_953_cast_fp16 = add(x = input_941_cast_fp16, y = var_3338_cast_fp16)[name = tensor("input_953_cast_fp16")]; tensor input_955_axes_0 = const()[name = tensor("input_955_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446146624)))]; tensor module_layers_17_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446148736)))]; tensor input_955_cast_fp16 = layer_norm(axes = input_955_axes_0, beta = module_layers_17_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_out_weight_to_fp16, x = input_953_cast_fp16)[name = tensor("input_955_cast_fp16")]; tensor input_957_axes_0 = const()[name = tensor("input_957_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446150848)))]; tensor module_layers_18_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446152960)))]; tensor input_957_cast_fp16 = layer_norm(axes = input_957_axes_0, beta = module_layers_18_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward1_weight_to_fp16, x = input_955_cast_fp16)[name = tensor("input_957_cast_fp16")]; tensor module_layers_18_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446155072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450353600))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450349440)))]; tensor module_layers_18_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_18_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450361856)))]; tensor linear_163_cast_fp16 = linear(bias = module_layers_18_feed_forward1_linear1_bias_to_fp16, weight = module_layers_18_feed_forward1_linear1_weight_to_fp16_quantized, x = input_957_cast_fp16)[name = tensor("linear_163_cast_fp16")]; tensor input_961_cast_fp16 = silu(x = linear_163_cast_fp16)[name = tensor("input_961_cast_fp16")]; tensor module_layers_18_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450370112))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454565568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454564480)))]; tensor module_layers_18_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_18_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454567680)))]; tensor linear_164_cast_fp16 = linear(bias = module_layers_18_feed_forward1_linear2_bias_to_fp16, weight = module_layers_18_feed_forward1_linear2_weight_to_fp16_quantized, x = input_961_cast_fp16)[name = tensor("linear_164_cast_fp16")]; tensor var_3368_to_fp16 = const()[name = tensor("op_3368_to_fp16"), val = tensor(0x1p-1)]; tensor var_3369_cast_fp16 = mul(x = linear_164_cast_fp16, y = var_3368_to_fp16)[name = tensor("op_3369_cast_fp16")]; tensor input_967_cast_fp16 = add(x = input_955_cast_fp16, y = var_3369_cast_fp16)[name = tensor("input_967_cast_fp16")]; tensor query_37_axes_0 = const()[name = tensor("query_37_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454569792)))]; tensor module_layers_18_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454571904)))]; tensor query_37_cast_fp16 = layer_norm(axes = query_37_axes_0, beta = module_layers_18_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_self_att_weight_to_fp16, x = input_967_cast_fp16)[name = tensor("query_37_cast_fp16")]; tensor module_layers_18_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454574016))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455623744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455622656)))]; tensor module_layers_18_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455625856)))]; tensor linear_165_cast_fp16 = linear(bias = module_layers_18_self_attn_linear_q_bias_to_fp16, weight = module_layers_18_self_attn_linear_q_weight_to_fp16_quantized, x = query_37_cast_fp16)[name = tensor("linear_165_cast_fp16")]; tensor var_3386 = const()[name = tensor("op_3386"), val = tensor([1, -1, 8, 128])]; tensor q_109_cast_fp16 = reshape(shape = var_3386, x = linear_165_cast_fp16)[name = tensor("q_109_cast_fp16")]; tensor module_layers_18_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455627968))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456677696))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456676608)))]; tensor module_layers_18_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456679808)))]; tensor linear_166_cast_fp16 = linear(bias = module_layers_18_self_attn_linear_k_bias_to_fp16, weight = module_layers_18_self_attn_linear_k_weight_to_fp16_quantized, x = query_37_cast_fp16)[name = tensor("linear_166_cast_fp16")]; tensor var_3391 = const()[name = tensor("op_3391"), val = tensor([1, -1, 8, 128])]; tensor k_73_cast_fp16 = reshape(shape = var_3391, x = linear_166_cast_fp16)[name = tensor("k_73_cast_fp16")]; tensor module_layers_18_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456681920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457731648))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457730560)))]; tensor module_layers_18_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457733760)))]; tensor linear_167_cast_fp16 = linear(bias = module_layers_18_self_attn_linear_v_bias_to_fp16, weight = module_layers_18_self_attn_linear_v_weight_to_fp16_quantized, x = query_37_cast_fp16)[name = tensor("linear_167_cast_fp16")]; tensor var_3396 = const()[name = tensor("op_3396"), val = tensor([1, -1, 8, 128])]; tensor v_37_cast_fp16 = reshape(shape = var_3396, x = linear_167_cast_fp16)[name = tensor("v_37_cast_fp16")]; tensor value_39_perm_0 = const()[name = tensor("value_39_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_18_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457735872)))]; tensor var_3408_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3408_cast_fp16")]; tensor module_layers_18_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457737984)))]; tensor var_3410_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3410_cast_fp16")]; tensor q_with_bias_v_37_perm_0 = const()[name = tensor("q_with_bias_v_37_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_405_transpose_x_0 = const()[name = tensor("x_405_transpose_x_0"), val = tensor(false)]; tensor x_405_transpose_y_0 = const()[name = tensor("x_405_transpose_y_0"), val = tensor(false)]; tensor op_3412_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3412_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457740096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(458124608))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(458124160)))]; tensor q_with_bias_v_37_cast_fp16 = transpose(perm = q_with_bias_v_37_perm_0, x = var_3410_cast_fp16)[name = tensor("transpose_186")]; tensor x_405_cast_fp16 = matmul(transpose_x = x_405_transpose_x_0, transpose_y = x_405_transpose_y_0, x = q_with_bias_v_37_cast_fp16, y = op_3412_to_fp16_quantized)[name = tensor("x_405_cast_fp16")]; tensor x_407_pad_0 = const()[name = tensor("x_407_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_407_mode_0 = const()[name = tensor("x_407_mode_0"), val = tensor("constant")]; tensor const_194_to_fp16 = const()[name = tensor("const_194_to_fp16"), val = tensor(0x0p+0)]; tensor x_407_cast_fp16 = pad(constant_val = const_194_to_fp16, mode = x_407_mode_0, pad = x_407_pad_0, x = x_405_cast_fp16)[name = tensor("x_407_cast_fp16")]; tensor var_3420 = const()[name = tensor("op_3420"), val = tensor([1, 8, -1, 188])]; tensor x_409_cast_fp16 = reshape(shape = var_3420, x = x_407_cast_fp16)[name = tensor("x_409_cast_fp16")]; tensor var_3424_begin_0 = const()[name = tensor("op_3424_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3424_end_0 = const()[name = tensor("op_3424_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3424_end_mask_0 = const()[name = tensor("op_3424_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3424_cast_fp16 = slice_by_index(begin = var_3424_begin_0, end = var_3424_end_0, end_mask = var_3424_end_mask_0, x = x_409_cast_fp16)[name = tensor("op_3424_cast_fp16")]; tensor var_3425 = const()[name = tensor("op_3425"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_73_cast_fp16 = reshape(shape = var_3425, x = var_3424_cast_fp16)[name = tensor("matrix_bd_73_cast_fp16")]; tensor matrix_ac_37_transpose_x_0 = const()[name = tensor("matrix_ac_37_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_37_transpose_y_0 = const()[name = tensor("matrix_ac_37_transpose_y_0"), val = tensor(false)]; tensor transpose_132_perm_0 = const()[name = tensor("transpose_132_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_133_perm_0 = const()[name = tensor("transpose_133_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_133 = transpose(perm = transpose_133_perm_0, x = k_73_cast_fp16)[name = tensor("transpose_184")]; tensor transpose_132 = transpose(perm = transpose_132_perm_0, x = var_3408_cast_fp16)[name = tensor("transpose_185")]; tensor matrix_ac_37_cast_fp16 = matmul(transpose_x = matrix_ac_37_transpose_x_0, transpose_y = matrix_ac_37_transpose_y_0, x = transpose_132, y = transpose_133)[name = tensor("matrix_ac_37_cast_fp16")]; tensor matrix_bd_75_begin_0 = const()[name = tensor("matrix_bd_75_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_75_end_0 = const()[name = tensor("matrix_bd_75_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_75_end_mask_0 = const()[name = tensor("matrix_bd_75_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_75_cast_fp16 = slice_by_index(begin = matrix_bd_75_begin_0, end = matrix_bd_75_end_0, end_mask = matrix_bd_75_end_mask_0, x = matrix_bd_73_cast_fp16)[name = tensor("matrix_bd_75_cast_fp16")]; tensor var_3434_cast_fp16 = add(x = matrix_ac_37_cast_fp16, y = matrix_bd_75_cast_fp16)[name = tensor("op_3434_cast_fp16")]; tensor _inversed_scores_73_y_0_to_fp16 = const()[name = tensor("_inversed_scores_73_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_73_cast_fp16 = mul(x = var_3434_cast_fp16, y = _inversed_scores_73_y_0_to_fp16)[name = tensor("_inversed_scores_73_cast_fp16")]; tensor scores_75_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_73_cast_fp16, cond = mask_3)[name = tensor("scores_75_cast_fp16")]; tensor var_3440_cast_fp16 = softmax(axis = var_30, x = scores_75_cast_fp16)[name = tensor("op_3440_cast_fp16")]; tensor input_969_cast_fp16 = select(a = var_11_to_fp16, b = var_3440_cast_fp16, cond = mask_3)[name = tensor("input_969_cast_fp16")]; tensor x_411_transpose_x_0 = const()[name = tensor("x_411_transpose_x_0"), val = tensor(false)]; tensor x_411_transpose_y_0 = const()[name = tensor("x_411_transpose_y_0"), val = tensor(false)]; tensor value_39_cast_fp16 = transpose(perm = value_39_perm_0, x = v_37_cast_fp16)[name = tensor("transpose_183")]; tensor x_411_cast_fp16 = matmul(transpose_x = x_411_transpose_x_0, transpose_y = x_411_transpose_y_0, x = input_969_cast_fp16, y = value_39_cast_fp16)[name = tensor("x_411_cast_fp16")]; tensor var_3444_perm_0 = const()[name = tensor("op_3444_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3445 = const()[name = tensor("op_3445"), val = tensor([1, -1, 1024])]; tensor var_3444_cast_fp16 = transpose(perm = var_3444_perm_0, x = x_411_cast_fp16)[name = tensor("transpose_182")]; tensor input_971_cast_fp16 = reshape(shape = var_3445, x = var_3444_cast_fp16)[name = tensor("input_971_cast_fp16")]; tensor module_layers_18_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(458125440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(459175168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(459174080)))]; tensor module_layers_18_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(459177280)))]; tensor linear_169_cast_fp16 = linear(bias = module_layers_18_self_attn_linear_out_bias_to_fp16, weight = module_layers_18_self_attn_linear_out_weight_to_fp16_quantized, x = input_971_cast_fp16)[name = tensor("linear_169_cast_fp16")]; tensor input_975_cast_fp16 = add(x = input_967_cast_fp16, y = linear_169_cast_fp16)[name = tensor("input_975_cast_fp16")]; tensor x_415_axes_0 = const()[name = tensor("x_415_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(459179392)))]; tensor module_layers_18_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(459181504)))]; tensor x_415_cast_fp16 = layer_norm(axes = x_415_axes_0, beta = module_layers_18_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_conv_weight_to_fp16, x = input_975_cast_fp16)[name = tensor("x_415_cast_fp16")]; tensor input_977_perm_0 = const()[name = tensor("input_977_perm_0"), val = tensor([0, 2, 1])]; tensor input_979_pad_type_0 = const()[name = tensor("input_979_pad_type_0"), val = tensor("valid")]; tensor input_979_strides_0 = const()[name = tensor("input_979_strides_0"), val = tensor([1])]; tensor input_979_pad_0 = const()[name = tensor("input_979_pad_0"), val = tensor([0, 0])]; tensor input_979_dilations_0 = const()[name = tensor("input_979_dilations_0"), val = tensor([1])]; tensor input_979_groups_0 = const()[name = tensor("input_979_groups_0"), val = tensor(1)]; tensor module_layers_18_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(459183616))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461282944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461280832)))]; tensor module_layers_18_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_18_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461287104)))]; tensor input_977_cast_fp16 = transpose(perm = input_977_perm_0, x = x_415_cast_fp16)[name = tensor("transpose_181")]; tensor input_979_cast_fp16 = conv(bias = module_layers_18_conv_pointwise_conv1_bias_to_fp16, dilations = input_979_dilations_0, groups = input_979_groups_0, pad = input_979_pad_0, pad_type = input_979_pad_type_0, strides = input_979_strides_0, weight = module_layers_18_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_977_cast_fp16)[name = tensor("input_979_cast_fp16")]; tensor x_417_split_num_splits_0 = const()[name = tensor("x_417_split_num_splits_0"), val = tensor(2)]; tensor x_417_split_axis_0 = const()[name = tensor("x_417_split_axis_0"), val = tensor(1)]; tensor x_417_split_cast_fp16_0, tensor x_417_split_cast_fp16_1 = split(axis = x_417_split_axis_0, num_splits = x_417_split_num_splits_0, x = input_979_cast_fp16)[name = tensor("x_417_split_cast_fp16")]; tensor x_417_split_1_sigmoid_cast_fp16 = sigmoid(x = x_417_split_cast_fp16_1)[name = tensor("x_417_split_1_sigmoid_cast_fp16")]; tensor x_417_cast_fp16 = mul(x = x_417_split_cast_fp16_0, y = x_417_split_1_sigmoid_cast_fp16)[name = tensor("x_417_cast_fp16")]; tensor input_981_cast_fp16 = select(a = var_11_to_fp16, b = x_417_cast_fp16, cond = var_337)[name = tensor("input_981_cast_fp16")]; tensor input_983_pad_0 = const()[name = tensor("input_983_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_983_mode_0 = const()[name = tensor("input_983_mode_0"), val = tensor("constant")]; tensor const_197_to_fp16 = const()[name = tensor("const_197_to_fp16"), val = tensor(0x0p+0)]; tensor input_983_cast_fp16 = pad(constant_val = const_197_to_fp16, mode = input_983_mode_0, pad = input_983_pad_0, x = input_981_cast_fp16)[name = tensor("input_983_cast_fp16")]; tensor input_985_pad_type_0 = const()[name = tensor("input_985_pad_type_0"), val = tensor("valid")]; tensor input_985_groups_0 = const()[name = tensor("input_985_groups_0"), val = tensor(1024)]; tensor input_985_strides_0 = const()[name = tensor("input_985_strides_0"), val = tensor([1])]; tensor input_985_pad_0 = const()[name = tensor("input_985_pad_0"), val = tensor([0, 0])]; tensor input_985_dilations_0 = const()[name = tensor("input_985_dilations_0"), val = tensor([1])]; tensor const_284_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_284_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461291264))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461301632))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461300544)))]; tensor const_285_to_fp16 = const()[name = tensor("const_285_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461303744)))]; tensor input_987_cast_fp16 = conv(bias = const_285_to_fp16, dilations = input_985_dilations_0, groups = input_985_groups_0, pad = input_985_pad_0, pad_type = input_985_pad_type_0, strides = input_985_strides_0, weight = const_284_to_fp16_quantized, x = input_983_cast_fp16)[name = tensor("input_987_cast_fp16")]; tensor input_989_cast_fp16 = silu(x = input_987_cast_fp16)[name = tensor("input_989_cast_fp16")]; tensor x_419_pad_type_0 = const()[name = tensor("x_419_pad_type_0"), val = tensor("valid")]; tensor x_419_strides_0 = const()[name = tensor("x_419_strides_0"), val = tensor([1])]; tensor x_419_pad_0 = const()[name = tensor("x_419_pad_0"), val = tensor([0, 0])]; tensor x_419_dilations_0 = const()[name = tensor("x_419_dilations_0"), val = tensor([1])]; tensor x_419_groups_0 = const()[name = tensor("x_419_groups_0"), val = tensor(1)]; tensor module_layers_18_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461305856))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462355584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462354496)))]; tensor module_layers_18_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_18_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462357696)))]; tensor x_419_cast_fp16 = conv(bias = module_layers_18_conv_pointwise_conv2_bias_to_fp16, dilations = x_419_dilations_0, groups = x_419_groups_0, pad = x_419_pad_0, pad_type = x_419_pad_type_0, strides = x_419_strides_0, weight = module_layers_18_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_989_cast_fp16)[name = tensor("x_419_cast_fp16")]; tensor input_991_perm_0 = const()[name = tensor("input_991_perm_0"), val = tensor([0, 2, 1])]; tensor input_991_cast_fp16 = transpose(perm = input_991_perm_0, x = x_419_cast_fp16)[name = tensor("transpose_180")]; tensor input_993_cast_fp16 = add(x = input_975_cast_fp16, y = input_991_cast_fp16)[name = tensor("input_993_cast_fp16")]; tensor input_995_axes_0 = const()[name = tensor("input_995_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462359808)))]; tensor module_layers_18_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462361920)))]; tensor input_995_cast_fp16 = layer_norm(axes = input_995_axes_0, beta = module_layers_18_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward2_weight_to_fp16, x = input_993_cast_fp16)[name = tensor("input_995_cast_fp16")]; tensor module_layers_18_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462364032))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466562560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466558400)))]; tensor module_layers_18_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_18_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466570816)))]; tensor linear_170_cast_fp16 = linear(bias = module_layers_18_feed_forward2_linear1_bias_to_fp16, weight = module_layers_18_feed_forward2_linear1_weight_to_fp16_quantized, x = input_995_cast_fp16)[name = tensor("linear_170_cast_fp16")]; tensor input_999_cast_fp16 = silu(x = linear_170_cast_fp16)[name = tensor("input_999_cast_fp16")]; tensor module_layers_18_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466579072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470774528))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470773440)))]; tensor module_layers_18_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_18_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470776640)))]; tensor linear_171_cast_fp16 = linear(bias = module_layers_18_feed_forward2_linear2_bias_to_fp16, weight = module_layers_18_feed_forward2_linear2_weight_to_fp16_quantized, x = input_999_cast_fp16)[name = tensor("linear_171_cast_fp16")]; tensor var_3511_to_fp16 = const()[name = tensor("op_3511_to_fp16"), val = tensor(0x1p-1)]; tensor var_3512_cast_fp16 = mul(x = linear_171_cast_fp16, y = var_3511_to_fp16)[name = tensor("op_3512_cast_fp16")]; tensor input_1005_cast_fp16 = add(x = input_993_cast_fp16, y = var_3512_cast_fp16)[name = tensor("input_1005_cast_fp16")]; tensor input_1007_axes_0 = const()[name = tensor("input_1007_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470778752)))]; tensor module_layers_18_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470780864)))]; tensor input_1007_cast_fp16 = layer_norm(axes = input_1007_axes_0, beta = module_layers_18_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_out_weight_to_fp16, x = input_1005_cast_fp16)[name = tensor("input_1007_cast_fp16")]; tensor input_1009_axes_0 = const()[name = tensor("input_1009_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470782976)))]; tensor module_layers_19_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470785088)))]; tensor input_1009_cast_fp16 = layer_norm(axes = input_1009_axes_0, beta = module_layers_19_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward1_weight_to_fp16, x = input_1007_cast_fp16)[name = tensor("input_1009_cast_fp16")]; tensor module_layers_19_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470787200))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474985728))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474981568)))]; tensor module_layers_19_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_19_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474993984)))]; tensor linear_172_cast_fp16 = linear(bias = module_layers_19_feed_forward1_linear1_bias_to_fp16, weight = module_layers_19_feed_forward1_linear1_weight_to_fp16_quantized, x = input_1009_cast_fp16)[name = tensor("linear_172_cast_fp16")]; tensor input_1013_cast_fp16 = silu(x = linear_172_cast_fp16)[name = tensor("input_1013_cast_fp16")]; tensor module_layers_19_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475002240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479197696))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479196608)))]; tensor module_layers_19_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_19_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479199808)))]; tensor linear_173_cast_fp16 = linear(bias = module_layers_19_feed_forward1_linear2_bias_to_fp16, weight = module_layers_19_feed_forward1_linear2_weight_to_fp16_quantized, x = input_1013_cast_fp16)[name = tensor("linear_173_cast_fp16")]; tensor var_3542_to_fp16 = const()[name = tensor("op_3542_to_fp16"), val = tensor(0x1p-1)]; tensor var_3543_cast_fp16 = mul(x = linear_173_cast_fp16, y = var_3542_to_fp16)[name = tensor("op_3543_cast_fp16")]; tensor input_1019_cast_fp16 = add(x = input_1007_cast_fp16, y = var_3543_cast_fp16)[name = tensor("input_1019_cast_fp16")]; tensor query_39_axes_0 = const()[name = tensor("query_39_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479201920)))]; tensor module_layers_19_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479204032)))]; tensor query_39_cast_fp16 = layer_norm(axes = query_39_axes_0, beta = module_layers_19_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_self_att_weight_to_fp16, x = input_1019_cast_fp16)[name = tensor("query_39_cast_fp16")]; tensor module_layers_19_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479206144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480255872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480254784)))]; tensor module_layers_19_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480257984)))]; tensor linear_174_cast_fp16 = linear(bias = module_layers_19_self_attn_linear_q_bias_to_fp16, weight = module_layers_19_self_attn_linear_q_weight_to_fp16_quantized, x = query_39_cast_fp16)[name = tensor("linear_174_cast_fp16")]; tensor var_3560 = const()[name = tensor("op_3560"), val = tensor([1, -1, 8, 128])]; tensor q_115_cast_fp16 = reshape(shape = var_3560, x = linear_174_cast_fp16)[name = tensor("q_115_cast_fp16")]; tensor module_layers_19_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480260096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481309824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481308736)))]; tensor module_layers_19_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481311936)))]; tensor linear_175_cast_fp16 = linear(bias = module_layers_19_self_attn_linear_k_bias_to_fp16, weight = module_layers_19_self_attn_linear_k_weight_to_fp16_quantized, x = query_39_cast_fp16)[name = tensor("linear_175_cast_fp16")]; tensor var_3565 = const()[name = tensor("op_3565"), val = tensor([1, -1, 8, 128])]; tensor k_77_cast_fp16 = reshape(shape = var_3565, x = linear_175_cast_fp16)[name = tensor("k_77_cast_fp16")]; tensor module_layers_19_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481314048))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482363776))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482362688)))]; tensor module_layers_19_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482365888)))]; tensor linear_176_cast_fp16 = linear(bias = module_layers_19_self_attn_linear_v_bias_to_fp16, weight = module_layers_19_self_attn_linear_v_weight_to_fp16_quantized, x = query_39_cast_fp16)[name = tensor("linear_176_cast_fp16")]; tensor var_3570 = const()[name = tensor("op_3570"), val = tensor([1, -1, 8, 128])]; tensor v_39_cast_fp16 = reshape(shape = var_3570, x = linear_176_cast_fp16)[name = tensor("v_39_cast_fp16")]; tensor value_41_perm_0 = const()[name = tensor("value_41_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_19_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482368000)))]; tensor var_3582_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3582_cast_fp16")]; tensor module_layers_19_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482370112)))]; tensor var_3584_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3584_cast_fp16")]; tensor q_with_bias_v_39_perm_0 = const()[name = tensor("q_with_bias_v_39_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_427_transpose_x_0 = const()[name = tensor("x_427_transpose_x_0"), val = tensor(false)]; tensor x_427_transpose_y_0 = const()[name = tensor("x_427_transpose_y_0"), val = tensor(false)]; tensor op_3586_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3586_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482372224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482756736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482756288)))]; tensor q_with_bias_v_39_cast_fp16 = transpose(perm = q_with_bias_v_39_perm_0, x = var_3584_cast_fp16)[name = tensor("transpose_179")]; tensor x_427_cast_fp16 = matmul(transpose_x = x_427_transpose_x_0, transpose_y = x_427_transpose_y_0, x = q_with_bias_v_39_cast_fp16, y = op_3586_to_fp16_quantized)[name = tensor("x_427_cast_fp16")]; tensor x_429_pad_0 = const()[name = tensor("x_429_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_429_mode_0 = const()[name = tensor("x_429_mode_0"), val = tensor("constant")]; tensor const_204_to_fp16 = const()[name = tensor("const_204_to_fp16"), val = tensor(0x0p+0)]; tensor x_429_cast_fp16 = pad(constant_val = const_204_to_fp16, mode = x_429_mode_0, pad = x_429_pad_0, x = x_427_cast_fp16)[name = tensor("x_429_cast_fp16")]; tensor var_3594 = const()[name = tensor("op_3594"), val = tensor([1, 8, -1, 188])]; tensor x_431_cast_fp16 = reshape(shape = var_3594, x = x_429_cast_fp16)[name = tensor("x_431_cast_fp16")]; tensor var_3598_begin_0 = const()[name = tensor("op_3598_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3598_end_0 = const()[name = tensor("op_3598_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3598_end_mask_0 = const()[name = tensor("op_3598_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3598_cast_fp16 = slice_by_index(begin = var_3598_begin_0, end = var_3598_end_0, end_mask = var_3598_end_mask_0, x = x_431_cast_fp16)[name = tensor("op_3598_cast_fp16")]; tensor var_3599 = const()[name = tensor("op_3599"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_77_cast_fp16 = reshape(shape = var_3599, x = var_3598_cast_fp16)[name = tensor("matrix_bd_77_cast_fp16")]; tensor matrix_ac_39_transpose_x_0 = const()[name = tensor("matrix_ac_39_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_39_transpose_y_0 = const()[name = tensor("matrix_ac_39_transpose_y_0"), val = tensor(false)]; tensor transpose_134_perm_0 = const()[name = tensor("transpose_134_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_135_perm_0 = const()[name = tensor("transpose_135_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_135 = transpose(perm = transpose_135_perm_0, x = k_77_cast_fp16)[name = tensor("transpose_177")]; tensor transpose_134 = transpose(perm = transpose_134_perm_0, x = var_3582_cast_fp16)[name = tensor("transpose_178")]; tensor matrix_ac_39_cast_fp16 = matmul(transpose_x = matrix_ac_39_transpose_x_0, transpose_y = matrix_ac_39_transpose_y_0, x = transpose_134, y = transpose_135)[name = tensor("matrix_ac_39_cast_fp16")]; tensor matrix_bd_79_begin_0 = const()[name = tensor("matrix_bd_79_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_79_end_0 = const()[name = tensor("matrix_bd_79_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_79_end_mask_0 = const()[name = tensor("matrix_bd_79_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_79_cast_fp16 = slice_by_index(begin = matrix_bd_79_begin_0, end = matrix_bd_79_end_0, end_mask = matrix_bd_79_end_mask_0, x = matrix_bd_77_cast_fp16)[name = tensor("matrix_bd_79_cast_fp16")]; tensor var_3608_cast_fp16 = add(x = matrix_ac_39_cast_fp16, y = matrix_bd_79_cast_fp16)[name = tensor("op_3608_cast_fp16")]; tensor _inversed_scores_77_y_0_to_fp16 = const()[name = tensor("_inversed_scores_77_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_77_cast_fp16 = mul(x = var_3608_cast_fp16, y = _inversed_scores_77_y_0_to_fp16)[name = tensor("_inversed_scores_77_cast_fp16")]; tensor scores_79_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_77_cast_fp16, cond = mask_3)[name = tensor("scores_79_cast_fp16")]; tensor var_3614_cast_fp16 = softmax(axis = var_30, x = scores_79_cast_fp16)[name = tensor("op_3614_cast_fp16")]; tensor input_1021_cast_fp16 = select(a = var_11_to_fp16, b = var_3614_cast_fp16, cond = mask_3)[name = tensor("input_1021_cast_fp16")]; tensor x_433_transpose_x_0 = const()[name = tensor("x_433_transpose_x_0"), val = tensor(false)]; tensor x_433_transpose_y_0 = const()[name = tensor("x_433_transpose_y_0"), val = tensor(false)]; tensor value_41_cast_fp16 = transpose(perm = value_41_perm_0, x = v_39_cast_fp16)[name = tensor("transpose_176")]; tensor x_433_cast_fp16 = matmul(transpose_x = x_433_transpose_x_0, transpose_y = x_433_transpose_y_0, x = input_1021_cast_fp16, y = value_41_cast_fp16)[name = tensor("x_433_cast_fp16")]; tensor var_3618_perm_0 = const()[name = tensor("op_3618_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3619 = const()[name = tensor("op_3619"), val = tensor([1, -1, 1024])]; tensor var_3618_cast_fp16 = transpose(perm = var_3618_perm_0, x = x_433_cast_fp16)[name = tensor("transpose_175")]; tensor input_1023_cast_fp16 = reshape(shape = var_3619, x = var_3618_cast_fp16)[name = tensor("input_1023_cast_fp16")]; tensor module_layers_19_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482757568))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483807296))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483806208)))]; tensor module_layers_19_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483809408)))]; tensor linear_178_cast_fp16 = linear(bias = module_layers_19_self_attn_linear_out_bias_to_fp16, weight = module_layers_19_self_attn_linear_out_weight_to_fp16_quantized, x = input_1023_cast_fp16)[name = tensor("linear_178_cast_fp16")]; tensor input_1027_cast_fp16 = add(x = input_1019_cast_fp16, y = linear_178_cast_fp16)[name = tensor("input_1027_cast_fp16")]; tensor x_437_axes_0 = const()[name = tensor("x_437_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483811520)))]; tensor module_layers_19_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483813632)))]; tensor x_437_cast_fp16 = layer_norm(axes = x_437_axes_0, beta = module_layers_19_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_conv_weight_to_fp16, x = input_1027_cast_fp16)[name = tensor("x_437_cast_fp16")]; tensor input_1029_perm_0 = const()[name = tensor("input_1029_perm_0"), val = tensor([0, 2, 1])]; tensor input_1031_pad_type_0 = const()[name = tensor("input_1031_pad_type_0"), val = tensor("valid")]; tensor input_1031_strides_0 = const()[name = tensor("input_1031_strides_0"), val = tensor([1])]; tensor input_1031_pad_0 = const()[name = tensor("input_1031_pad_0"), val = tensor([0, 0])]; tensor input_1031_dilations_0 = const()[name = tensor("input_1031_dilations_0"), val = tensor([1])]; tensor input_1031_groups_0 = const()[name = tensor("input_1031_groups_0"), val = tensor(1)]; tensor module_layers_19_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483815744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485915072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485912960)))]; tensor module_layers_19_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_19_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485919232)))]; tensor input_1029_cast_fp16 = transpose(perm = input_1029_perm_0, x = x_437_cast_fp16)[name = tensor("transpose_174")]; tensor input_1031_cast_fp16 = conv(bias = module_layers_19_conv_pointwise_conv1_bias_to_fp16, dilations = input_1031_dilations_0, groups = input_1031_groups_0, pad = input_1031_pad_0, pad_type = input_1031_pad_type_0, strides = input_1031_strides_0, weight = module_layers_19_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_1029_cast_fp16)[name = tensor("input_1031_cast_fp16")]; tensor x_439_split_num_splits_0 = const()[name = tensor("x_439_split_num_splits_0"), val = tensor(2)]; tensor x_439_split_axis_0 = const()[name = tensor("x_439_split_axis_0"), val = tensor(1)]; tensor x_439_split_cast_fp16_0, tensor x_439_split_cast_fp16_1 = split(axis = x_439_split_axis_0, num_splits = x_439_split_num_splits_0, x = input_1031_cast_fp16)[name = tensor("x_439_split_cast_fp16")]; tensor x_439_split_1_sigmoid_cast_fp16 = sigmoid(x = x_439_split_cast_fp16_1)[name = tensor("x_439_split_1_sigmoid_cast_fp16")]; tensor x_439_cast_fp16 = mul(x = x_439_split_cast_fp16_0, y = x_439_split_1_sigmoid_cast_fp16)[name = tensor("x_439_cast_fp16")]; tensor input_1033_cast_fp16 = select(a = var_11_to_fp16, b = x_439_cast_fp16, cond = var_337)[name = tensor("input_1033_cast_fp16")]; tensor input_1035_pad_0 = const()[name = tensor("input_1035_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1035_mode_0 = const()[name = tensor("input_1035_mode_0"), val = tensor("constant")]; tensor const_207_to_fp16 = const()[name = tensor("const_207_to_fp16"), val = tensor(0x0p+0)]; tensor input_1035_cast_fp16 = pad(constant_val = const_207_to_fp16, mode = input_1035_mode_0, pad = input_1035_pad_0, x = input_1033_cast_fp16)[name = tensor("input_1035_cast_fp16")]; tensor input_1037_pad_type_0 = const()[name = tensor("input_1037_pad_type_0"), val = tensor("valid")]; tensor input_1037_groups_0 = const()[name = tensor("input_1037_groups_0"), val = tensor(1024)]; tensor input_1037_strides_0 = const()[name = tensor("input_1037_strides_0"), val = tensor([1])]; tensor input_1037_pad_0 = const()[name = tensor("input_1037_pad_0"), val = tensor([0, 0])]; tensor input_1037_dilations_0 = const()[name = tensor("input_1037_dilations_0"), val = tensor([1])]; tensor const_286_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_286_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485923392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485933760))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485932672)))]; tensor const_287_to_fp16 = const()[name = tensor("const_287_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485935872)))]; tensor input_1039_cast_fp16 = conv(bias = const_287_to_fp16, dilations = input_1037_dilations_0, groups = input_1037_groups_0, pad = input_1037_pad_0, pad_type = input_1037_pad_type_0, strides = input_1037_strides_0, weight = const_286_to_fp16_quantized, x = input_1035_cast_fp16)[name = tensor("input_1039_cast_fp16")]; tensor input_1041_cast_fp16 = silu(x = input_1039_cast_fp16)[name = tensor("input_1041_cast_fp16")]; tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; tensor module_layers_19_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485937984))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486987712))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486986624)))]; tensor module_layers_19_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_19_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486989824)))]; tensor x_441_cast_fp16 = conv(bias = module_layers_19_conv_pointwise_conv2_bias_to_fp16, dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = module_layers_19_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_1041_cast_fp16)[name = tensor("x_441_cast_fp16")]; tensor input_1043_perm_0 = const()[name = tensor("input_1043_perm_0"), val = tensor([0, 2, 1])]; tensor input_1043_cast_fp16 = transpose(perm = input_1043_perm_0, x = x_441_cast_fp16)[name = tensor("transpose_173")]; tensor input_1045_cast_fp16 = add(x = input_1027_cast_fp16, y = input_1043_cast_fp16)[name = tensor("input_1045_cast_fp16")]; tensor input_1047_axes_0 = const()[name = tensor("input_1047_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486991936)))]; tensor module_layers_19_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486994048)))]; tensor input_1047_cast_fp16 = layer_norm(axes = input_1047_axes_0, beta = module_layers_19_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward2_weight_to_fp16, x = input_1045_cast_fp16)[name = tensor("input_1047_cast_fp16")]; tensor module_layers_19_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486996160))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491194688))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491190528)))]; tensor module_layers_19_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_19_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491202944)))]; tensor linear_179_cast_fp16 = linear(bias = module_layers_19_feed_forward2_linear1_bias_to_fp16, weight = module_layers_19_feed_forward2_linear1_weight_to_fp16_quantized, x = input_1047_cast_fp16)[name = tensor("linear_179_cast_fp16")]; tensor input_1051_cast_fp16 = silu(x = linear_179_cast_fp16)[name = tensor("input_1051_cast_fp16")]; tensor module_layers_19_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491211200))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495406656))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495405568)))]; tensor module_layers_19_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_19_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495408768)))]; tensor linear_180_cast_fp16 = linear(bias = module_layers_19_feed_forward2_linear2_bias_to_fp16, weight = module_layers_19_feed_forward2_linear2_weight_to_fp16_quantized, x = input_1051_cast_fp16)[name = tensor("linear_180_cast_fp16")]; tensor var_3685_to_fp16 = const()[name = tensor("op_3685_to_fp16"), val = tensor(0x1p-1)]; tensor var_3686_cast_fp16 = mul(x = linear_180_cast_fp16, y = var_3685_to_fp16)[name = tensor("op_3686_cast_fp16")]; tensor input_1057_cast_fp16 = add(x = input_1045_cast_fp16, y = var_3686_cast_fp16)[name = tensor("input_1057_cast_fp16")]; tensor input_1059_axes_0 = const()[name = tensor("input_1059_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495410880)))]; tensor module_layers_19_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495412992)))]; tensor input_1059_cast_fp16 = layer_norm(axes = input_1059_axes_0, beta = module_layers_19_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_out_weight_to_fp16, x = input_1057_cast_fp16)[name = tensor("input_1059_cast_fp16")]; tensor input_1061_axes_0 = const()[name = tensor("input_1061_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495415104)))]; tensor module_layers_20_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495417216)))]; tensor input_1061_cast_fp16 = layer_norm(axes = input_1061_axes_0, beta = module_layers_20_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward1_weight_to_fp16, x = input_1059_cast_fp16)[name = tensor("input_1061_cast_fp16")]; tensor module_layers_20_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495419328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499617856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499613696)))]; tensor module_layers_20_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_20_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499626112)))]; tensor linear_181_cast_fp16 = linear(bias = module_layers_20_feed_forward1_linear1_bias_to_fp16, weight = module_layers_20_feed_forward1_linear1_weight_to_fp16_quantized, x = input_1061_cast_fp16)[name = tensor("linear_181_cast_fp16")]; tensor input_1065_cast_fp16 = silu(x = linear_181_cast_fp16)[name = tensor("input_1065_cast_fp16")]; tensor module_layers_20_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499634368))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503829824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503828736)))]; tensor module_layers_20_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_20_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503831936)))]; tensor linear_182_cast_fp16 = linear(bias = module_layers_20_feed_forward1_linear2_bias_to_fp16, weight = module_layers_20_feed_forward1_linear2_weight_to_fp16_quantized, x = input_1065_cast_fp16)[name = tensor("linear_182_cast_fp16")]; tensor var_3716_to_fp16 = const()[name = tensor("op_3716_to_fp16"), val = tensor(0x1p-1)]; tensor var_3717_cast_fp16 = mul(x = linear_182_cast_fp16, y = var_3716_to_fp16)[name = tensor("op_3717_cast_fp16")]; tensor input_1071_cast_fp16 = add(x = input_1059_cast_fp16, y = var_3717_cast_fp16)[name = tensor("input_1071_cast_fp16")]; tensor query_41_axes_0 = const()[name = tensor("query_41_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503834048)))]; tensor module_layers_20_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503836160)))]; tensor query_41_cast_fp16 = layer_norm(axes = query_41_axes_0, beta = module_layers_20_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_self_att_weight_to_fp16, x = input_1071_cast_fp16)[name = tensor("query_41_cast_fp16")]; tensor module_layers_20_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503838272))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504888000))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504886912)))]; tensor module_layers_20_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504890112)))]; tensor linear_183_cast_fp16 = linear(bias = module_layers_20_self_attn_linear_q_bias_to_fp16, weight = module_layers_20_self_attn_linear_q_weight_to_fp16_quantized, x = query_41_cast_fp16)[name = tensor("linear_183_cast_fp16")]; tensor var_3734 = const()[name = tensor("op_3734"), val = tensor([1, -1, 8, 128])]; tensor q_121_cast_fp16 = reshape(shape = var_3734, x = linear_183_cast_fp16)[name = tensor("q_121_cast_fp16")]; tensor module_layers_20_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504892224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505941952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505940864)))]; tensor module_layers_20_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505944064)))]; tensor linear_184_cast_fp16 = linear(bias = module_layers_20_self_attn_linear_k_bias_to_fp16, weight = module_layers_20_self_attn_linear_k_weight_to_fp16_quantized, x = query_41_cast_fp16)[name = tensor("linear_184_cast_fp16")]; tensor var_3739 = const()[name = tensor("op_3739"), val = tensor([1, -1, 8, 128])]; tensor k_81_cast_fp16 = reshape(shape = var_3739, x = linear_184_cast_fp16)[name = tensor("k_81_cast_fp16")]; tensor module_layers_20_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505946176))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506995904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506994816)))]; tensor module_layers_20_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506998016)))]; tensor linear_185_cast_fp16 = linear(bias = module_layers_20_self_attn_linear_v_bias_to_fp16, weight = module_layers_20_self_attn_linear_v_weight_to_fp16_quantized, x = query_41_cast_fp16)[name = tensor("linear_185_cast_fp16")]; tensor var_3744 = const()[name = tensor("op_3744"), val = tensor([1, -1, 8, 128])]; tensor v_41_cast_fp16 = reshape(shape = var_3744, x = linear_185_cast_fp16)[name = tensor("v_41_cast_fp16")]; tensor value_43_perm_0 = const()[name = tensor("value_43_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_20_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507000128)))]; tensor var_3756_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3756_cast_fp16")]; tensor module_layers_20_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507002240)))]; tensor var_3758_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3758_cast_fp16")]; tensor q_with_bias_v_41_perm_0 = const()[name = tensor("q_with_bias_v_41_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_449_transpose_x_0 = const()[name = tensor("x_449_transpose_x_0"), val = tensor(false)]; tensor x_449_transpose_y_0 = const()[name = tensor("x_449_transpose_y_0"), val = tensor(false)]; tensor op_3760_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3760_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507004352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507388864))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507388416)))]; tensor q_with_bias_v_41_cast_fp16 = transpose(perm = q_with_bias_v_41_perm_0, x = var_3758_cast_fp16)[name = tensor("transpose_172")]; tensor x_449_cast_fp16 = matmul(transpose_x = x_449_transpose_x_0, transpose_y = x_449_transpose_y_0, x = q_with_bias_v_41_cast_fp16, y = op_3760_to_fp16_quantized)[name = tensor("x_449_cast_fp16")]; tensor x_451_pad_0 = const()[name = tensor("x_451_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_451_mode_0 = const()[name = tensor("x_451_mode_0"), val = tensor("constant")]; tensor const_214_to_fp16 = const()[name = tensor("const_214_to_fp16"), val = tensor(0x0p+0)]; tensor x_451_cast_fp16 = pad(constant_val = const_214_to_fp16, mode = x_451_mode_0, pad = x_451_pad_0, x = x_449_cast_fp16)[name = tensor("x_451_cast_fp16")]; tensor var_3768 = const()[name = tensor("op_3768"), val = tensor([1, 8, -1, 188])]; tensor x_453_cast_fp16 = reshape(shape = var_3768, x = x_451_cast_fp16)[name = tensor("x_453_cast_fp16")]; tensor var_3772_begin_0 = const()[name = tensor("op_3772_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3772_end_0 = const()[name = tensor("op_3772_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3772_end_mask_0 = const()[name = tensor("op_3772_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3772_cast_fp16 = slice_by_index(begin = var_3772_begin_0, end = var_3772_end_0, end_mask = var_3772_end_mask_0, x = x_453_cast_fp16)[name = tensor("op_3772_cast_fp16")]; tensor var_3773 = const()[name = tensor("op_3773"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_81_cast_fp16 = reshape(shape = var_3773, x = var_3772_cast_fp16)[name = tensor("matrix_bd_81_cast_fp16")]; tensor matrix_ac_41_transpose_x_0 = const()[name = tensor("matrix_ac_41_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_41_transpose_y_0 = const()[name = tensor("matrix_ac_41_transpose_y_0"), val = tensor(false)]; tensor transpose_136_perm_0 = const()[name = tensor("transpose_136_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_137_perm_0 = const()[name = tensor("transpose_137_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_137 = transpose(perm = transpose_137_perm_0, x = k_81_cast_fp16)[name = tensor("transpose_170")]; tensor transpose_136 = transpose(perm = transpose_136_perm_0, x = var_3756_cast_fp16)[name = tensor("transpose_171")]; tensor matrix_ac_41_cast_fp16 = matmul(transpose_x = matrix_ac_41_transpose_x_0, transpose_y = matrix_ac_41_transpose_y_0, x = transpose_136, y = transpose_137)[name = tensor("matrix_ac_41_cast_fp16")]; tensor matrix_bd_83_begin_0 = const()[name = tensor("matrix_bd_83_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_83_end_0 = const()[name = tensor("matrix_bd_83_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_83_end_mask_0 = const()[name = tensor("matrix_bd_83_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_83_cast_fp16 = slice_by_index(begin = matrix_bd_83_begin_0, end = matrix_bd_83_end_0, end_mask = matrix_bd_83_end_mask_0, x = matrix_bd_81_cast_fp16)[name = tensor("matrix_bd_83_cast_fp16")]; tensor var_3782_cast_fp16 = add(x = matrix_ac_41_cast_fp16, y = matrix_bd_83_cast_fp16)[name = tensor("op_3782_cast_fp16")]; tensor _inversed_scores_81_y_0_to_fp16 = const()[name = tensor("_inversed_scores_81_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_81_cast_fp16 = mul(x = var_3782_cast_fp16, y = _inversed_scores_81_y_0_to_fp16)[name = tensor("_inversed_scores_81_cast_fp16")]; tensor scores_83_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_81_cast_fp16, cond = mask_3)[name = tensor("scores_83_cast_fp16")]; tensor var_3788_cast_fp16 = softmax(axis = var_30, x = scores_83_cast_fp16)[name = tensor("op_3788_cast_fp16")]; tensor input_1073_cast_fp16 = select(a = var_11_to_fp16, b = var_3788_cast_fp16, cond = mask_3)[name = tensor("input_1073_cast_fp16")]; tensor x_455_transpose_x_0 = const()[name = tensor("x_455_transpose_x_0"), val = tensor(false)]; tensor x_455_transpose_y_0 = const()[name = tensor("x_455_transpose_y_0"), val = tensor(false)]; tensor value_43_cast_fp16 = transpose(perm = value_43_perm_0, x = v_41_cast_fp16)[name = tensor("transpose_169")]; tensor x_455_cast_fp16 = matmul(transpose_x = x_455_transpose_x_0, transpose_y = x_455_transpose_y_0, x = input_1073_cast_fp16, y = value_43_cast_fp16)[name = tensor("x_455_cast_fp16")]; tensor var_3792_perm_0 = const()[name = tensor("op_3792_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3793 = const()[name = tensor("op_3793"), val = tensor([1, -1, 1024])]; tensor var_3792_cast_fp16 = transpose(perm = var_3792_perm_0, x = x_455_cast_fp16)[name = tensor("transpose_168")]; tensor input_1075_cast_fp16 = reshape(shape = var_3793, x = var_3792_cast_fp16)[name = tensor("input_1075_cast_fp16")]; tensor module_layers_20_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507389696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508439424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508438336)))]; tensor module_layers_20_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508441536)))]; tensor linear_187_cast_fp16 = linear(bias = module_layers_20_self_attn_linear_out_bias_to_fp16, weight = module_layers_20_self_attn_linear_out_weight_to_fp16_quantized, x = input_1075_cast_fp16)[name = tensor("linear_187_cast_fp16")]; tensor input_1079_cast_fp16 = add(x = input_1071_cast_fp16, y = linear_187_cast_fp16)[name = tensor("input_1079_cast_fp16")]; tensor x_459_axes_0 = const()[name = tensor("x_459_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508443648)))]; tensor module_layers_20_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508445760)))]; tensor x_459_cast_fp16 = layer_norm(axes = x_459_axes_0, beta = module_layers_20_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_conv_weight_to_fp16, x = input_1079_cast_fp16)[name = tensor("x_459_cast_fp16")]; tensor input_1081_perm_0 = const()[name = tensor("input_1081_perm_0"), val = tensor([0, 2, 1])]; tensor input_1083_pad_type_0 = const()[name = tensor("input_1083_pad_type_0"), val = tensor("valid")]; tensor input_1083_strides_0 = const()[name = tensor("input_1083_strides_0"), val = tensor([1])]; tensor input_1083_pad_0 = const()[name = tensor("input_1083_pad_0"), val = tensor([0, 0])]; tensor input_1083_dilations_0 = const()[name = tensor("input_1083_dilations_0"), val = tensor([1])]; tensor input_1083_groups_0 = const()[name = tensor("input_1083_groups_0"), val = tensor(1)]; tensor module_layers_20_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508447872))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510547200))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510545088)))]; tensor module_layers_20_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_20_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510551360)))]; tensor input_1081_cast_fp16 = transpose(perm = input_1081_perm_0, x = x_459_cast_fp16)[name = tensor("transpose_167")]; tensor input_1083_cast_fp16 = conv(bias = module_layers_20_conv_pointwise_conv1_bias_to_fp16, dilations = input_1083_dilations_0, groups = input_1083_groups_0, pad = input_1083_pad_0, pad_type = input_1083_pad_type_0, strides = input_1083_strides_0, weight = module_layers_20_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_1081_cast_fp16)[name = tensor("input_1083_cast_fp16")]; tensor x_461_split_num_splits_0 = const()[name = tensor("x_461_split_num_splits_0"), val = tensor(2)]; tensor x_461_split_axis_0 = const()[name = tensor("x_461_split_axis_0"), val = tensor(1)]; tensor x_461_split_cast_fp16_0, tensor x_461_split_cast_fp16_1 = split(axis = x_461_split_axis_0, num_splits = x_461_split_num_splits_0, x = input_1083_cast_fp16)[name = tensor("x_461_split_cast_fp16")]; tensor x_461_split_1_sigmoid_cast_fp16 = sigmoid(x = x_461_split_cast_fp16_1)[name = tensor("x_461_split_1_sigmoid_cast_fp16")]; tensor x_461_cast_fp16 = mul(x = x_461_split_cast_fp16_0, y = x_461_split_1_sigmoid_cast_fp16)[name = tensor("x_461_cast_fp16")]; tensor input_1085_cast_fp16 = select(a = var_11_to_fp16, b = x_461_cast_fp16, cond = var_337)[name = tensor("input_1085_cast_fp16")]; tensor input_1087_pad_0 = const()[name = tensor("input_1087_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1087_mode_0 = const()[name = tensor("input_1087_mode_0"), val = tensor("constant")]; tensor const_217_to_fp16 = const()[name = tensor("const_217_to_fp16"), val = tensor(0x0p+0)]; tensor input_1087_cast_fp16 = pad(constant_val = const_217_to_fp16, mode = input_1087_mode_0, pad = input_1087_pad_0, x = input_1085_cast_fp16)[name = tensor("input_1087_cast_fp16")]; tensor input_1089_pad_type_0 = const()[name = tensor("input_1089_pad_type_0"), val = tensor("valid")]; tensor input_1089_groups_0 = const()[name = tensor("input_1089_groups_0"), val = tensor(1024)]; tensor input_1089_strides_0 = const()[name = tensor("input_1089_strides_0"), val = tensor([1])]; tensor input_1089_pad_0 = const()[name = tensor("input_1089_pad_0"), val = tensor([0, 0])]; tensor input_1089_dilations_0 = const()[name = tensor("input_1089_dilations_0"), val = tensor([1])]; tensor const_288_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_288_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510555520))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510565888))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510564800)))]; tensor const_289_to_fp16 = const()[name = tensor("const_289_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510568000)))]; tensor input_1091_cast_fp16 = conv(bias = const_289_to_fp16, dilations = input_1089_dilations_0, groups = input_1089_groups_0, pad = input_1089_pad_0, pad_type = input_1089_pad_type_0, strides = input_1089_strides_0, weight = const_288_to_fp16_quantized, x = input_1087_cast_fp16)[name = tensor("input_1091_cast_fp16")]; tensor input_1093_cast_fp16 = silu(x = input_1091_cast_fp16)[name = tensor("input_1093_cast_fp16")]; tensor x_463_pad_type_0 = const()[name = tensor("x_463_pad_type_0"), val = tensor("valid")]; tensor x_463_strides_0 = const()[name = tensor("x_463_strides_0"), val = tensor([1])]; tensor x_463_pad_0 = const()[name = tensor("x_463_pad_0"), val = tensor([0, 0])]; tensor x_463_dilations_0 = const()[name = tensor("x_463_dilations_0"), val = tensor([1])]; tensor x_463_groups_0 = const()[name = tensor("x_463_groups_0"), val = tensor(1)]; tensor module_layers_20_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510570112))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511619840))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511618752)))]; tensor module_layers_20_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_20_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511621952)))]; tensor x_463_cast_fp16 = conv(bias = module_layers_20_conv_pointwise_conv2_bias_to_fp16, dilations = x_463_dilations_0, groups = x_463_groups_0, pad = x_463_pad_0, pad_type = x_463_pad_type_0, strides = x_463_strides_0, weight = module_layers_20_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_1093_cast_fp16)[name = tensor("x_463_cast_fp16")]; tensor input_1095_perm_0 = const()[name = tensor("input_1095_perm_0"), val = tensor([0, 2, 1])]; tensor input_1095_cast_fp16 = transpose(perm = input_1095_perm_0, x = x_463_cast_fp16)[name = tensor("transpose_166")]; tensor input_1097_cast_fp16 = add(x = input_1079_cast_fp16, y = input_1095_cast_fp16)[name = tensor("input_1097_cast_fp16")]; tensor input_1099_axes_0 = const()[name = tensor("input_1099_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511624064)))]; tensor module_layers_20_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511626176)))]; tensor input_1099_cast_fp16 = layer_norm(axes = input_1099_axes_0, beta = module_layers_20_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward2_weight_to_fp16, x = input_1097_cast_fp16)[name = tensor("input_1099_cast_fp16")]; tensor module_layers_20_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511628288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515826816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515822656)))]; tensor module_layers_20_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_20_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515835072)))]; tensor linear_188_cast_fp16 = linear(bias = module_layers_20_feed_forward2_linear1_bias_to_fp16, weight = module_layers_20_feed_forward2_linear1_weight_to_fp16_quantized, x = input_1099_cast_fp16)[name = tensor("linear_188_cast_fp16")]; tensor input_1103_cast_fp16 = silu(x = linear_188_cast_fp16)[name = tensor("input_1103_cast_fp16")]; tensor module_layers_20_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515843328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520038784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520037696)))]; tensor module_layers_20_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_20_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520040896)))]; tensor linear_189_cast_fp16 = linear(bias = module_layers_20_feed_forward2_linear2_bias_to_fp16, weight = module_layers_20_feed_forward2_linear2_weight_to_fp16_quantized, x = input_1103_cast_fp16)[name = tensor("linear_189_cast_fp16")]; tensor var_3859_to_fp16 = const()[name = tensor("op_3859_to_fp16"), val = tensor(0x1p-1)]; tensor var_3860_cast_fp16 = mul(x = linear_189_cast_fp16, y = var_3859_to_fp16)[name = tensor("op_3860_cast_fp16")]; tensor input_1109_cast_fp16 = add(x = input_1097_cast_fp16, y = var_3860_cast_fp16)[name = tensor("input_1109_cast_fp16")]; tensor input_1111_axes_0 = const()[name = tensor("input_1111_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520043008)))]; tensor module_layers_20_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520045120)))]; tensor input_1111_cast_fp16 = layer_norm(axes = input_1111_axes_0, beta = module_layers_20_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_out_weight_to_fp16, x = input_1109_cast_fp16)[name = tensor("input_1111_cast_fp16")]; tensor input_1113_axes_0 = const()[name = tensor("input_1113_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520047232)))]; tensor module_layers_21_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520049344)))]; tensor input_1113_cast_fp16 = layer_norm(axes = input_1113_axes_0, beta = module_layers_21_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward1_weight_to_fp16, x = input_1111_cast_fp16)[name = tensor("input_1113_cast_fp16")]; tensor module_layers_21_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520051456))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524249984))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524245824)))]; tensor module_layers_21_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_21_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524258240)))]; tensor linear_190_cast_fp16 = linear(bias = module_layers_21_feed_forward1_linear1_bias_to_fp16, weight = module_layers_21_feed_forward1_linear1_weight_to_fp16_quantized, x = input_1113_cast_fp16)[name = tensor("linear_190_cast_fp16")]; tensor input_1117_cast_fp16 = silu(x = linear_190_cast_fp16)[name = tensor("input_1117_cast_fp16")]; tensor module_layers_21_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524266496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528461952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528460864)))]; tensor module_layers_21_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_21_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528464064)))]; tensor linear_191_cast_fp16 = linear(bias = module_layers_21_feed_forward1_linear2_bias_to_fp16, weight = module_layers_21_feed_forward1_linear2_weight_to_fp16_quantized, x = input_1117_cast_fp16)[name = tensor("linear_191_cast_fp16")]; tensor var_3890_to_fp16 = const()[name = tensor("op_3890_to_fp16"), val = tensor(0x1p-1)]; tensor var_3891_cast_fp16 = mul(x = linear_191_cast_fp16, y = var_3890_to_fp16)[name = tensor("op_3891_cast_fp16")]; tensor input_1123_cast_fp16 = add(x = input_1111_cast_fp16, y = var_3891_cast_fp16)[name = tensor("input_1123_cast_fp16")]; tensor query_43_axes_0 = const()[name = tensor("query_43_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528466176)))]; tensor module_layers_21_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528468288)))]; tensor query_43_cast_fp16 = layer_norm(axes = query_43_axes_0, beta = module_layers_21_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_self_att_weight_to_fp16, x = input_1123_cast_fp16)[name = tensor("query_43_cast_fp16")]; tensor module_layers_21_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528470400))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529520128))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529519040)))]; tensor module_layers_21_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529522240)))]; tensor linear_192_cast_fp16 = linear(bias = module_layers_21_self_attn_linear_q_bias_to_fp16, weight = module_layers_21_self_attn_linear_q_weight_to_fp16_quantized, x = query_43_cast_fp16)[name = tensor("linear_192_cast_fp16")]; tensor var_3908 = const()[name = tensor("op_3908"), val = tensor([1, -1, 8, 128])]; tensor q_127_cast_fp16 = reshape(shape = var_3908, x = linear_192_cast_fp16)[name = tensor("q_127_cast_fp16")]; tensor module_layers_21_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529524352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530574080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530572992)))]; tensor module_layers_21_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530576192)))]; tensor linear_193_cast_fp16 = linear(bias = module_layers_21_self_attn_linear_k_bias_to_fp16, weight = module_layers_21_self_attn_linear_k_weight_to_fp16_quantized, x = query_43_cast_fp16)[name = tensor("linear_193_cast_fp16")]; tensor var_3913 = const()[name = tensor("op_3913"), val = tensor([1, -1, 8, 128])]; tensor k_85_cast_fp16 = reshape(shape = var_3913, x = linear_193_cast_fp16)[name = tensor("k_85_cast_fp16")]; tensor module_layers_21_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530578304))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531628032))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531626944)))]; tensor module_layers_21_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531630144)))]; tensor linear_194_cast_fp16 = linear(bias = module_layers_21_self_attn_linear_v_bias_to_fp16, weight = module_layers_21_self_attn_linear_v_weight_to_fp16_quantized, x = query_43_cast_fp16)[name = tensor("linear_194_cast_fp16")]; tensor var_3918 = const()[name = tensor("op_3918"), val = tensor([1, -1, 8, 128])]; tensor v_43_cast_fp16 = reshape(shape = var_3918, x = linear_194_cast_fp16)[name = tensor("v_43_cast_fp16")]; tensor value_45_perm_0 = const()[name = tensor("value_45_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_21_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531632256)))]; tensor var_3930_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3930_cast_fp16")]; tensor module_layers_21_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531634368)))]; tensor var_3932_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3932_cast_fp16")]; tensor q_with_bias_v_43_perm_0 = const()[name = tensor("q_with_bias_v_43_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_471_transpose_x_0 = const()[name = tensor("x_471_transpose_x_0"), val = tensor(false)]; tensor x_471_transpose_y_0 = const()[name = tensor("x_471_transpose_y_0"), val = tensor(false)]; tensor op_3934_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3934_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531636480))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(532020992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(532020544)))]; tensor q_with_bias_v_43_cast_fp16 = transpose(perm = q_with_bias_v_43_perm_0, x = var_3932_cast_fp16)[name = tensor("transpose_165")]; tensor x_471_cast_fp16 = matmul(transpose_x = x_471_transpose_x_0, transpose_y = x_471_transpose_y_0, x = q_with_bias_v_43_cast_fp16, y = op_3934_to_fp16_quantized)[name = tensor("x_471_cast_fp16")]; tensor x_473_pad_0 = const()[name = tensor("x_473_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_473_mode_0 = const()[name = tensor("x_473_mode_0"), val = tensor("constant")]; tensor const_224_to_fp16 = const()[name = tensor("const_224_to_fp16"), val = tensor(0x0p+0)]; tensor x_473_cast_fp16 = pad(constant_val = const_224_to_fp16, mode = x_473_mode_0, pad = x_473_pad_0, x = x_471_cast_fp16)[name = tensor("x_473_cast_fp16")]; tensor var_3942 = const()[name = tensor("op_3942"), val = tensor([1, 8, -1, 188])]; tensor x_475_cast_fp16 = reshape(shape = var_3942, x = x_473_cast_fp16)[name = tensor("x_475_cast_fp16")]; tensor var_3946_begin_0 = const()[name = tensor("op_3946_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3946_end_0 = const()[name = tensor("op_3946_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3946_end_mask_0 = const()[name = tensor("op_3946_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3946_cast_fp16 = slice_by_index(begin = var_3946_begin_0, end = var_3946_end_0, end_mask = var_3946_end_mask_0, x = x_475_cast_fp16)[name = tensor("op_3946_cast_fp16")]; tensor var_3947 = const()[name = tensor("op_3947"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_85_cast_fp16 = reshape(shape = var_3947, x = var_3946_cast_fp16)[name = tensor("matrix_bd_85_cast_fp16")]; tensor matrix_ac_43_transpose_x_0 = const()[name = tensor("matrix_ac_43_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_43_transpose_y_0 = const()[name = tensor("matrix_ac_43_transpose_y_0"), val = tensor(false)]; tensor transpose_138_perm_0 = const()[name = tensor("transpose_138_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_139_perm_0 = const()[name = tensor("transpose_139_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_139 = transpose(perm = transpose_139_perm_0, x = k_85_cast_fp16)[name = tensor("transpose_163")]; tensor transpose_138 = transpose(perm = transpose_138_perm_0, x = var_3930_cast_fp16)[name = tensor("transpose_164")]; tensor matrix_ac_43_cast_fp16 = matmul(transpose_x = matrix_ac_43_transpose_x_0, transpose_y = matrix_ac_43_transpose_y_0, x = transpose_138, y = transpose_139)[name = tensor("matrix_ac_43_cast_fp16")]; tensor matrix_bd_87_begin_0 = const()[name = tensor("matrix_bd_87_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_87_end_0 = const()[name = tensor("matrix_bd_87_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_87_end_mask_0 = const()[name = tensor("matrix_bd_87_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_87_cast_fp16 = slice_by_index(begin = matrix_bd_87_begin_0, end = matrix_bd_87_end_0, end_mask = matrix_bd_87_end_mask_0, x = matrix_bd_85_cast_fp16)[name = tensor("matrix_bd_87_cast_fp16")]; tensor var_3956_cast_fp16 = add(x = matrix_ac_43_cast_fp16, y = matrix_bd_87_cast_fp16)[name = tensor("op_3956_cast_fp16")]; tensor _inversed_scores_85_y_0_to_fp16 = const()[name = tensor("_inversed_scores_85_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_85_cast_fp16 = mul(x = var_3956_cast_fp16, y = _inversed_scores_85_y_0_to_fp16)[name = tensor("_inversed_scores_85_cast_fp16")]; tensor scores_87_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_85_cast_fp16, cond = mask_3)[name = tensor("scores_87_cast_fp16")]; tensor var_3962_cast_fp16 = softmax(axis = var_30, x = scores_87_cast_fp16)[name = tensor("op_3962_cast_fp16")]; tensor input_1125_cast_fp16 = select(a = var_11_to_fp16, b = var_3962_cast_fp16, cond = mask_3)[name = tensor("input_1125_cast_fp16")]; tensor x_477_transpose_x_0 = const()[name = tensor("x_477_transpose_x_0"), val = tensor(false)]; tensor x_477_transpose_y_0 = const()[name = tensor("x_477_transpose_y_0"), val = tensor(false)]; tensor value_45_cast_fp16 = transpose(perm = value_45_perm_0, x = v_43_cast_fp16)[name = tensor("transpose_162")]; tensor x_477_cast_fp16 = matmul(transpose_x = x_477_transpose_x_0, transpose_y = x_477_transpose_y_0, x = input_1125_cast_fp16, y = value_45_cast_fp16)[name = tensor("x_477_cast_fp16")]; tensor var_3966_perm_0 = const()[name = tensor("op_3966_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3967 = const()[name = tensor("op_3967"), val = tensor([1, -1, 1024])]; tensor var_3966_cast_fp16 = transpose(perm = var_3966_perm_0, x = x_477_cast_fp16)[name = tensor("transpose_161")]; tensor input_1127_cast_fp16 = reshape(shape = var_3967, x = var_3966_cast_fp16)[name = tensor("input_1127_cast_fp16")]; tensor module_layers_21_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(532021824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533071552))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533070464)))]; tensor module_layers_21_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533073664)))]; tensor linear_196_cast_fp16 = linear(bias = module_layers_21_self_attn_linear_out_bias_to_fp16, weight = module_layers_21_self_attn_linear_out_weight_to_fp16_quantized, x = input_1127_cast_fp16)[name = tensor("linear_196_cast_fp16")]; tensor input_1131_cast_fp16 = add(x = input_1123_cast_fp16, y = linear_196_cast_fp16)[name = tensor("input_1131_cast_fp16")]; tensor x_481_axes_0 = const()[name = tensor("x_481_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533075776)))]; tensor module_layers_21_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533077888)))]; tensor x_481_cast_fp16 = layer_norm(axes = x_481_axes_0, beta = module_layers_21_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_conv_weight_to_fp16, x = input_1131_cast_fp16)[name = tensor("x_481_cast_fp16")]; tensor input_1133_perm_0 = const()[name = tensor("input_1133_perm_0"), val = tensor([0, 2, 1])]; tensor input_1135_pad_type_0 = const()[name = tensor("input_1135_pad_type_0"), val = tensor("valid")]; tensor input_1135_strides_0 = const()[name = tensor("input_1135_strides_0"), val = tensor([1])]; tensor input_1135_pad_0 = const()[name = tensor("input_1135_pad_0"), val = tensor([0, 0])]; tensor input_1135_dilations_0 = const()[name = tensor("input_1135_dilations_0"), val = tensor([1])]; tensor input_1135_groups_0 = const()[name = tensor("input_1135_groups_0"), val = tensor(1)]; tensor module_layers_21_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533080000))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535179328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535177216)))]; tensor module_layers_21_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_21_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535183488)))]; tensor input_1133_cast_fp16 = transpose(perm = input_1133_perm_0, x = x_481_cast_fp16)[name = tensor("transpose_160")]; tensor input_1135_cast_fp16 = conv(bias = module_layers_21_conv_pointwise_conv1_bias_to_fp16, dilations = input_1135_dilations_0, groups = input_1135_groups_0, pad = input_1135_pad_0, pad_type = input_1135_pad_type_0, strides = input_1135_strides_0, weight = module_layers_21_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_1133_cast_fp16)[name = tensor("input_1135_cast_fp16")]; tensor x_483_split_num_splits_0 = const()[name = tensor("x_483_split_num_splits_0"), val = tensor(2)]; tensor x_483_split_axis_0 = const()[name = tensor("x_483_split_axis_0"), val = tensor(1)]; tensor x_483_split_cast_fp16_0, tensor x_483_split_cast_fp16_1 = split(axis = x_483_split_axis_0, num_splits = x_483_split_num_splits_0, x = input_1135_cast_fp16)[name = tensor("x_483_split_cast_fp16")]; tensor x_483_split_1_sigmoid_cast_fp16 = sigmoid(x = x_483_split_cast_fp16_1)[name = tensor("x_483_split_1_sigmoid_cast_fp16")]; tensor x_483_cast_fp16 = mul(x = x_483_split_cast_fp16_0, y = x_483_split_1_sigmoid_cast_fp16)[name = tensor("x_483_cast_fp16")]; tensor input_1137_cast_fp16 = select(a = var_11_to_fp16, b = x_483_cast_fp16, cond = var_337)[name = tensor("input_1137_cast_fp16")]; tensor input_1139_pad_0 = const()[name = tensor("input_1139_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1139_mode_0 = const()[name = tensor("input_1139_mode_0"), val = tensor("constant")]; tensor const_227_to_fp16 = const()[name = tensor("const_227_to_fp16"), val = tensor(0x0p+0)]; tensor input_1139_cast_fp16 = pad(constant_val = const_227_to_fp16, mode = input_1139_mode_0, pad = input_1139_pad_0, x = input_1137_cast_fp16)[name = tensor("input_1139_cast_fp16")]; tensor input_1141_pad_type_0 = const()[name = tensor("input_1141_pad_type_0"), val = tensor("valid")]; tensor input_1141_groups_0 = const()[name = tensor("input_1141_groups_0"), val = tensor(1024)]; tensor input_1141_strides_0 = const()[name = tensor("input_1141_strides_0"), val = tensor([1])]; tensor input_1141_pad_0 = const()[name = tensor("input_1141_pad_0"), val = tensor([0, 0])]; tensor input_1141_dilations_0 = const()[name = tensor("input_1141_dilations_0"), val = tensor([1])]; tensor const_290_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_290_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535187648))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535198016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535196928)))]; tensor const_291_to_fp16 = const()[name = tensor("const_291_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535200128)))]; tensor input_1143_cast_fp16 = conv(bias = const_291_to_fp16, dilations = input_1141_dilations_0, groups = input_1141_groups_0, pad = input_1141_pad_0, pad_type = input_1141_pad_type_0, strides = input_1141_strides_0, weight = const_290_to_fp16_quantized, x = input_1139_cast_fp16)[name = tensor("input_1143_cast_fp16")]; tensor input_1145_cast_fp16 = silu(x = input_1143_cast_fp16)[name = tensor("input_1145_cast_fp16")]; tensor x_485_pad_type_0 = const()[name = tensor("x_485_pad_type_0"), val = tensor("valid")]; tensor x_485_strides_0 = const()[name = tensor("x_485_strides_0"), val = tensor([1])]; tensor x_485_pad_0 = const()[name = tensor("x_485_pad_0"), val = tensor([0, 0])]; tensor x_485_dilations_0 = const()[name = tensor("x_485_dilations_0"), val = tensor([1])]; tensor x_485_groups_0 = const()[name = tensor("x_485_groups_0"), val = tensor(1)]; tensor module_layers_21_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535202240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(536251968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(536250880)))]; tensor module_layers_21_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_21_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(536254080)))]; tensor x_485_cast_fp16 = conv(bias = module_layers_21_conv_pointwise_conv2_bias_to_fp16, dilations = x_485_dilations_0, groups = x_485_groups_0, pad = x_485_pad_0, pad_type = x_485_pad_type_0, strides = x_485_strides_0, weight = module_layers_21_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_1145_cast_fp16)[name = tensor("x_485_cast_fp16")]; tensor input_1147_perm_0 = const()[name = tensor("input_1147_perm_0"), val = tensor([0, 2, 1])]; tensor input_1147_cast_fp16 = transpose(perm = input_1147_perm_0, x = x_485_cast_fp16)[name = tensor("transpose_159")]; tensor input_1149_cast_fp16 = add(x = input_1131_cast_fp16, y = input_1147_cast_fp16)[name = tensor("input_1149_cast_fp16")]; tensor input_1151_axes_0 = const()[name = tensor("input_1151_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(536256192)))]; tensor module_layers_21_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(536258304)))]; tensor input_1151_cast_fp16 = layer_norm(axes = input_1151_axes_0, beta = module_layers_21_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward2_weight_to_fp16, x = input_1149_cast_fp16)[name = tensor("input_1151_cast_fp16")]; tensor module_layers_21_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(536260416))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540458944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540454784)))]; tensor module_layers_21_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_21_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540467200)))]; tensor linear_197_cast_fp16 = linear(bias = module_layers_21_feed_forward2_linear1_bias_to_fp16, weight = module_layers_21_feed_forward2_linear1_weight_to_fp16_quantized, x = input_1151_cast_fp16)[name = tensor("linear_197_cast_fp16")]; tensor input_1155_cast_fp16 = silu(x = linear_197_cast_fp16)[name = tensor("input_1155_cast_fp16")]; tensor module_layers_21_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540475456))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544670912))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544669824)))]; tensor module_layers_21_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_21_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544673024)))]; tensor linear_198_cast_fp16 = linear(bias = module_layers_21_feed_forward2_linear2_bias_to_fp16, weight = module_layers_21_feed_forward2_linear2_weight_to_fp16_quantized, x = input_1155_cast_fp16)[name = tensor("linear_198_cast_fp16")]; tensor var_4033_to_fp16 = const()[name = tensor("op_4033_to_fp16"), val = tensor(0x1p-1)]; tensor var_4034_cast_fp16 = mul(x = linear_198_cast_fp16, y = var_4033_to_fp16)[name = tensor("op_4034_cast_fp16")]; tensor input_1161_cast_fp16 = add(x = input_1149_cast_fp16, y = var_4034_cast_fp16)[name = tensor("input_1161_cast_fp16")]; tensor input_1163_axes_0 = const()[name = tensor("input_1163_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544675136)))]; tensor module_layers_21_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544677248)))]; tensor input_1163_cast_fp16 = layer_norm(axes = input_1163_axes_0, beta = module_layers_21_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_out_weight_to_fp16, x = input_1161_cast_fp16)[name = tensor("input_1163_cast_fp16")]; tensor input_1165_axes_0 = const()[name = tensor("input_1165_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544679360)))]; tensor module_layers_22_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544681472)))]; tensor input_1165_cast_fp16 = layer_norm(axes = input_1165_axes_0, beta = module_layers_22_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward1_weight_to_fp16, x = input_1163_cast_fp16)[name = tensor("input_1165_cast_fp16")]; tensor module_layers_22_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544683584))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548882112))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548877952)))]; tensor module_layers_22_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_22_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548890368)))]; tensor linear_199_cast_fp16 = linear(bias = module_layers_22_feed_forward1_linear1_bias_to_fp16, weight = module_layers_22_feed_forward1_linear1_weight_to_fp16_quantized, x = input_1165_cast_fp16)[name = tensor("linear_199_cast_fp16")]; tensor input_1169_cast_fp16 = silu(x = linear_199_cast_fp16)[name = tensor("input_1169_cast_fp16")]; tensor module_layers_22_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548898624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553094080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553092992)))]; tensor module_layers_22_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_22_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553096192)))]; tensor linear_200_cast_fp16 = linear(bias = module_layers_22_feed_forward1_linear2_bias_to_fp16, weight = module_layers_22_feed_forward1_linear2_weight_to_fp16_quantized, x = input_1169_cast_fp16)[name = tensor("linear_200_cast_fp16")]; tensor var_4064_to_fp16 = const()[name = tensor("op_4064_to_fp16"), val = tensor(0x1p-1)]; tensor var_4065_cast_fp16 = mul(x = linear_200_cast_fp16, y = var_4064_to_fp16)[name = tensor("op_4065_cast_fp16")]; tensor input_1175_cast_fp16 = add(x = input_1163_cast_fp16, y = var_4065_cast_fp16)[name = tensor("input_1175_cast_fp16")]; tensor query_45_axes_0 = const()[name = tensor("query_45_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553098304)))]; tensor module_layers_22_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553100416)))]; tensor query_45_cast_fp16 = layer_norm(axes = query_45_axes_0, beta = module_layers_22_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_self_att_weight_to_fp16, x = input_1175_cast_fp16)[name = tensor("query_45_cast_fp16")]; tensor module_layers_22_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553102528))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554152256))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554151168)))]; tensor module_layers_22_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554154368)))]; tensor linear_201_cast_fp16 = linear(bias = module_layers_22_self_attn_linear_q_bias_to_fp16, weight = module_layers_22_self_attn_linear_q_weight_to_fp16_quantized, x = query_45_cast_fp16)[name = tensor("linear_201_cast_fp16")]; tensor var_4082 = const()[name = tensor("op_4082"), val = tensor([1, -1, 8, 128])]; tensor q_133_cast_fp16 = reshape(shape = var_4082, x = linear_201_cast_fp16)[name = tensor("q_133_cast_fp16")]; tensor module_layers_22_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554156480))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555206208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555205120)))]; tensor module_layers_22_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555208320)))]; tensor linear_202_cast_fp16 = linear(bias = module_layers_22_self_attn_linear_k_bias_to_fp16, weight = module_layers_22_self_attn_linear_k_weight_to_fp16_quantized, x = query_45_cast_fp16)[name = tensor("linear_202_cast_fp16")]; tensor var_4087 = const()[name = tensor("op_4087"), val = tensor([1, -1, 8, 128])]; tensor k_89_cast_fp16 = reshape(shape = var_4087, x = linear_202_cast_fp16)[name = tensor("k_89_cast_fp16")]; tensor module_layers_22_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555210432))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556260160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556259072)))]; tensor module_layers_22_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556262272)))]; tensor linear_203_cast_fp16 = linear(bias = module_layers_22_self_attn_linear_v_bias_to_fp16, weight = module_layers_22_self_attn_linear_v_weight_to_fp16_quantized, x = query_45_cast_fp16)[name = tensor("linear_203_cast_fp16")]; tensor var_4092 = const()[name = tensor("op_4092"), val = tensor([1, -1, 8, 128])]; tensor v_45_cast_fp16 = reshape(shape = var_4092, x = linear_203_cast_fp16)[name = tensor("v_45_cast_fp16")]; tensor value_47_perm_0 = const()[name = tensor("value_47_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_22_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556264384)))]; tensor var_4104_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_u_to_fp16)[name = tensor("op_4104_cast_fp16")]; tensor module_layers_22_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556266496)))]; tensor var_4106_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_v_to_fp16)[name = tensor("op_4106_cast_fp16")]; tensor q_with_bias_v_45_perm_0 = const()[name = tensor("q_with_bias_v_45_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_493_transpose_x_0 = const()[name = tensor("x_493_transpose_x_0"), val = tensor(false)]; tensor x_493_transpose_y_0 = const()[name = tensor("x_493_transpose_y_0"), val = tensor(false)]; tensor op_4108_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4108_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556268608))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556653120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556652672)))]; tensor q_with_bias_v_45_cast_fp16 = transpose(perm = q_with_bias_v_45_perm_0, x = var_4106_cast_fp16)[name = tensor("transpose_158")]; tensor x_493_cast_fp16 = matmul(transpose_x = x_493_transpose_x_0, transpose_y = x_493_transpose_y_0, x = q_with_bias_v_45_cast_fp16, y = op_4108_to_fp16_quantized)[name = tensor("x_493_cast_fp16")]; tensor x_495_pad_0 = const()[name = tensor("x_495_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_495_mode_0 = const()[name = tensor("x_495_mode_0"), val = tensor("constant")]; tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(0x0p+0)]; tensor x_495_cast_fp16 = pad(constant_val = const_234_to_fp16, mode = x_495_mode_0, pad = x_495_pad_0, x = x_493_cast_fp16)[name = tensor("x_495_cast_fp16")]; tensor var_4116 = const()[name = tensor("op_4116"), val = tensor([1, 8, -1, 188])]; tensor x_497_cast_fp16 = reshape(shape = var_4116, x = x_495_cast_fp16)[name = tensor("x_497_cast_fp16")]; tensor var_4120_begin_0 = const()[name = tensor("op_4120_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_4120_end_0 = const()[name = tensor("op_4120_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_4120_end_mask_0 = const()[name = tensor("op_4120_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_4120_cast_fp16 = slice_by_index(begin = var_4120_begin_0, end = var_4120_end_0, end_mask = var_4120_end_mask_0, x = x_497_cast_fp16)[name = tensor("op_4120_cast_fp16")]; tensor var_4121 = const()[name = tensor("op_4121"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_89_cast_fp16 = reshape(shape = var_4121, x = var_4120_cast_fp16)[name = tensor("matrix_bd_89_cast_fp16")]; tensor matrix_ac_45_transpose_x_0 = const()[name = tensor("matrix_ac_45_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_45_transpose_y_0 = const()[name = tensor("matrix_ac_45_transpose_y_0"), val = tensor(false)]; tensor transpose_140_perm_0 = const()[name = tensor("transpose_140_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_141_perm_0 = const()[name = tensor("transpose_141_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_141 = transpose(perm = transpose_141_perm_0, x = k_89_cast_fp16)[name = tensor("transpose_156")]; tensor transpose_140 = transpose(perm = transpose_140_perm_0, x = var_4104_cast_fp16)[name = tensor("transpose_157")]; tensor matrix_ac_45_cast_fp16 = matmul(transpose_x = matrix_ac_45_transpose_x_0, transpose_y = matrix_ac_45_transpose_y_0, x = transpose_140, y = transpose_141)[name = tensor("matrix_ac_45_cast_fp16")]; tensor matrix_bd_91_begin_0 = const()[name = tensor("matrix_bd_91_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_91_end_0 = const()[name = tensor("matrix_bd_91_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_91_end_mask_0 = const()[name = tensor("matrix_bd_91_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_91_cast_fp16 = slice_by_index(begin = matrix_bd_91_begin_0, end = matrix_bd_91_end_0, end_mask = matrix_bd_91_end_mask_0, x = matrix_bd_89_cast_fp16)[name = tensor("matrix_bd_91_cast_fp16")]; tensor var_4130_cast_fp16 = add(x = matrix_ac_45_cast_fp16, y = matrix_bd_91_cast_fp16)[name = tensor("op_4130_cast_fp16")]; tensor _inversed_scores_89_y_0_to_fp16 = const()[name = tensor("_inversed_scores_89_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_89_cast_fp16 = mul(x = var_4130_cast_fp16, y = _inversed_scores_89_y_0_to_fp16)[name = tensor("_inversed_scores_89_cast_fp16")]; tensor scores_91_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_89_cast_fp16, cond = mask_3)[name = tensor("scores_91_cast_fp16")]; tensor var_4136_cast_fp16 = softmax(axis = var_30, x = scores_91_cast_fp16)[name = tensor("op_4136_cast_fp16")]; tensor input_1177_cast_fp16 = select(a = var_11_to_fp16, b = var_4136_cast_fp16, cond = mask_3)[name = tensor("input_1177_cast_fp16")]; tensor x_499_transpose_x_0 = const()[name = tensor("x_499_transpose_x_0"), val = tensor(false)]; tensor x_499_transpose_y_0 = const()[name = tensor("x_499_transpose_y_0"), val = tensor(false)]; tensor value_47_cast_fp16 = transpose(perm = value_47_perm_0, x = v_45_cast_fp16)[name = tensor("transpose_155")]; tensor x_499_cast_fp16 = matmul(transpose_x = x_499_transpose_x_0, transpose_y = x_499_transpose_y_0, x = input_1177_cast_fp16, y = value_47_cast_fp16)[name = tensor("x_499_cast_fp16")]; tensor var_4140_perm_0 = const()[name = tensor("op_4140_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4141 = const()[name = tensor("op_4141"), val = tensor([1, -1, 1024])]; tensor var_4140_cast_fp16 = transpose(perm = var_4140_perm_0, x = x_499_cast_fp16)[name = tensor("transpose_154")]; tensor input_1179_cast_fp16 = reshape(shape = var_4141, x = var_4140_cast_fp16)[name = tensor("input_1179_cast_fp16")]; tensor module_layers_22_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556653952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557703680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557702592)))]; tensor module_layers_22_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557705792)))]; tensor linear_205_cast_fp16 = linear(bias = module_layers_22_self_attn_linear_out_bias_to_fp16, weight = module_layers_22_self_attn_linear_out_weight_to_fp16_quantized, x = input_1179_cast_fp16)[name = tensor("linear_205_cast_fp16")]; tensor input_1183_cast_fp16 = add(x = input_1175_cast_fp16, y = linear_205_cast_fp16)[name = tensor("input_1183_cast_fp16")]; tensor x_503_axes_0 = const()[name = tensor("x_503_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557707904)))]; tensor module_layers_22_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557710016)))]; tensor x_503_cast_fp16 = layer_norm(axes = x_503_axes_0, beta = module_layers_22_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_conv_weight_to_fp16, x = input_1183_cast_fp16)[name = tensor("x_503_cast_fp16")]; tensor input_1185_perm_0 = const()[name = tensor("input_1185_perm_0"), val = tensor([0, 2, 1])]; tensor input_1187_pad_type_0 = const()[name = tensor("input_1187_pad_type_0"), val = tensor("valid")]; tensor input_1187_strides_0 = const()[name = tensor("input_1187_strides_0"), val = tensor([1])]; tensor input_1187_pad_0 = const()[name = tensor("input_1187_pad_0"), val = tensor([0, 0])]; tensor input_1187_dilations_0 = const()[name = tensor("input_1187_dilations_0"), val = tensor([1])]; tensor input_1187_groups_0 = const()[name = tensor("input_1187_groups_0"), val = tensor(1)]; tensor module_layers_22_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557712128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559811456))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559809344)))]; tensor module_layers_22_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_22_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559815616)))]; tensor input_1185_cast_fp16 = transpose(perm = input_1185_perm_0, x = x_503_cast_fp16)[name = tensor("transpose_153")]; tensor input_1187_cast_fp16 = conv(bias = module_layers_22_conv_pointwise_conv1_bias_to_fp16, dilations = input_1187_dilations_0, groups = input_1187_groups_0, pad = input_1187_pad_0, pad_type = input_1187_pad_type_0, strides = input_1187_strides_0, weight = module_layers_22_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_1185_cast_fp16)[name = tensor("input_1187_cast_fp16")]; tensor x_505_split_num_splits_0 = const()[name = tensor("x_505_split_num_splits_0"), val = tensor(2)]; tensor x_505_split_axis_0 = const()[name = tensor("x_505_split_axis_0"), val = tensor(1)]; tensor x_505_split_cast_fp16_0, tensor x_505_split_cast_fp16_1 = split(axis = x_505_split_axis_0, num_splits = x_505_split_num_splits_0, x = input_1187_cast_fp16)[name = tensor("x_505_split_cast_fp16")]; tensor x_505_split_1_sigmoid_cast_fp16 = sigmoid(x = x_505_split_cast_fp16_1)[name = tensor("x_505_split_1_sigmoid_cast_fp16")]; tensor x_505_cast_fp16 = mul(x = x_505_split_cast_fp16_0, y = x_505_split_1_sigmoid_cast_fp16)[name = tensor("x_505_cast_fp16")]; tensor input_1189_cast_fp16 = select(a = var_11_to_fp16, b = x_505_cast_fp16, cond = var_337)[name = tensor("input_1189_cast_fp16")]; tensor input_1191_pad_0 = const()[name = tensor("input_1191_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1191_mode_0 = const()[name = tensor("input_1191_mode_0"), val = tensor("constant")]; tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(0x0p+0)]; tensor input_1191_cast_fp16 = pad(constant_val = const_237_to_fp16, mode = input_1191_mode_0, pad = input_1191_pad_0, x = input_1189_cast_fp16)[name = tensor("input_1191_cast_fp16")]; tensor input_1193_pad_type_0 = const()[name = tensor("input_1193_pad_type_0"), val = tensor("valid")]; tensor input_1193_groups_0 = const()[name = tensor("input_1193_groups_0"), val = tensor(1024)]; tensor input_1193_strides_0 = const()[name = tensor("input_1193_strides_0"), val = tensor([1])]; tensor input_1193_pad_0 = const()[name = tensor("input_1193_pad_0"), val = tensor([0, 0])]; tensor input_1193_dilations_0 = const()[name = tensor("input_1193_dilations_0"), val = tensor([1])]; tensor const_292_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_292_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559819776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559830144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559829056)))]; tensor const_293_to_fp16 = const()[name = tensor("const_293_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559832256)))]; tensor input_1195_cast_fp16 = conv(bias = const_293_to_fp16, dilations = input_1193_dilations_0, groups = input_1193_groups_0, pad = input_1193_pad_0, pad_type = input_1193_pad_type_0, strides = input_1193_strides_0, weight = const_292_to_fp16_quantized, x = input_1191_cast_fp16)[name = tensor("input_1195_cast_fp16")]; tensor input_1197_cast_fp16 = silu(x = input_1195_cast_fp16)[name = tensor("input_1197_cast_fp16")]; tensor x_507_pad_type_0 = const()[name = tensor("x_507_pad_type_0"), val = tensor("valid")]; tensor x_507_strides_0 = const()[name = tensor("x_507_strides_0"), val = tensor([1])]; tensor x_507_pad_0 = const()[name = tensor("x_507_pad_0"), val = tensor([0, 0])]; tensor x_507_dilations_0 = const()[name = tensor("x_507_dilations_0"), val = tensor([1])]; tensor x_507_groups_0 = const()[name = tensor("x_507_groups_0"), val = tensor(1)]; tensor module_layers_22_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559834368))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560884096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560883008)))]; tensor module_layers_22_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_22_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560886208)))]; tensor x_507_cast_fp16 = conv(bias = module_layers_22_conv_pointwise_conv2_bias_to_fp16, dilations = x_507_dilations_0, groups = x_507_groups_0, pad = x_507_pad_0, pad_type = x_507_pad_type_0, strides = x_507_strides_0, weight = module_layers_22_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_1197_cast_fp16)[name = tensor("x_507_cast_fp16")]; tensor input_1199_perm_0 = const()[name = tensor("input_1199_perm_0"), val = tensor([0, 2, 1])]; tensor input_1199_cast_fp16 = transpose(perm = input_1199_perm_0, x = x_507_cast_fp16)[name = tensor("transpose_152")]; tensor input_1201_cast_fp16 = add(x = input_1183_cast_fp16, y = input_1199_cast_fp16)[name = tensor("input_1201_cast_fp16")]; tensor input_1203_axes_0 = const()[name = tensor("input_1203_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560888320)))]; tensor module_layers_22_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560890432)))]; tensor input_1203_cast_fp16 = layer_norm(axes = input_1203_axes_0, beta = module_layers_22_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward2_weight_to_fp16, x = input_1201_cast_fp16)[name = tensor("input_1203_cast_fp16")]; tensor module_layers_22_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560892544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565091072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565086912)))]; tensor module_layers_22_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_22_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565099328)))]; tensor linear_206_cast_fp16 = linear(bias = module_layers_22_feed_forward2_linear1_bias_to_fp16, weight = module_layers_22_feed_forward2_linear1_weight_to_fp16_quantized, x = input_1203_cast_fp16)[name = tensor("linear_206_cast_fp16")]; tensor input_1207_cast_fp16 = silu(x = linear_206_cast_fp16)[name = tensor("input_1207_cast_fp16")]; tensor module_layers_22_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565107584))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569303040))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569301952)))]; tensor module_layers_22_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_22_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569305152)))]; tensor linear_207_cast_fp16 = linear(bias = module_layers_22_feed_forward2_linear2_bias_to_fp16, weight = module_layers_22_feed_forward2_linear2_weight_to_fp16_quantized, x = input_1207_cast_fp16)[name = tensor("linear_207_cast_fp16")]; tensor var_4207_to_fp16 = const()[name = tensor("op_4207_to_fp16"), val = tensor(0x1p-1)]; tensor var_4208_cast_fp16 = mul(x = linear_207_cast_fp16, y = var_4207_to_fp16)[name = tensor("op_4208_cast_fp16")]; tensor input_1213_cast_fp16 = add(x = input_1201_cast_fp16, y = var_4208_cast_fp16)[name = tensor("input_1213_cast_fp16")]; tensor input_1215_axes_0 = const()[name = tensor("input_1215_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569307264)))]; tensor module_layers_22_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569309376)))]; tensor input_1215_cast_fp16 = layer_norm(axes = input_1215_axes_0, beta = module_layers_22_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_out_weight_to_fp16, x = input_1213_cast_fp16)[name = tensor("input_1215_cast_fp16")]; tensor input_1217_axes_0 = const()[name = tensor("input_1217_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569311488)))]; tensor module_layers_23_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569313600)))]; tensor input_1217_cast_fp16 = layer_norm(axes = input_1217_axes_0, beta = module_layers_23_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward1_weight_to_fp16, x = input_1215_cast_fp16)[name = tensor("input_1217_cast_fp16")]; tensor module_layers_23_feed_forward1_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward1_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569315712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573514240))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573510080)))]; tensor module_layers_23_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_23_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573522496)))]; tensor linear_208_cast_fp16 = linear(bias = module_layers_23_feed_forward1_linear1_bias_to_fp16, weight = module_layers_23_feed_forward1_linear1_weight_to_fp16_quantized, x = input_1217_cast_fp16)[name = tensor("linear_208_cast_fp16")]; tensor input_1221_cast_fp16 = silu(x = linear_208_cast_fp16)[name = tensor("input_1221_cast_fp16")]; tensor module_layers_23_feed_forward1_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward1_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573530752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577726208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577725120)))]; tensor module_layers_23_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_23_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577728320)))]; tensor linear_209_cast_fp16 = linear(bias = module_layers_23_feed_forward1_linear2_bias_to_fp16, weight = module_layers_23_feed_forward1_linear2_weight_to_fp16_quantized, x = input_1221_cast_fp16)[name = tensor("linear_209_cast_fp16")]; tensor var_4238_to_fp16 = const()[name = tensor("op_4238_to_fp16"), val = tensor(0x1p-1)]; tensor var_4239_cast_fp16 = mul(x = linear_209_cast_fp16, y = var_4238_to_fp16)[name = tensor("op_4239_cast_fp16")]; tensor input_1227_cast_fp16 = add(x = input_1215_cast_fp16, y = var_4239_cast_fp16)[name = tensor("input_1227_cast_fp16")]; tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577730432)))]; tensor module_layers_23_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577732544)))]; tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_23_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_self_att_weight_to_fp16, x = input_1227_cast_fp16)[name = tensor("query_cast_fp16")]; tensor module_layers_23_self_attn_linear_q_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_q_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577734656))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578784384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578783296)))]; tensor module_layers_23_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578786496)))]; tensor linear_210_cast_fp16 = linear(bias = module_layers_23_self_attn_linear_q_bias_to_fp16, weight = module_layers_23_self_attn_linear_q_weight_to_fp16_quantized, x = query_cast_fp16)[name = tensor("linear_210_cast_fp16")]; tensor var_4256 = const()[name = tensor("op_4256"), val = tensor([1, -1, 8, 128])]; tensor q_139_cast_fp16 = reshape(shape = var_4256, x = linear_210_cast_fp16)[name = tensor("q_139_cast_fp16")]; tensor module_layers_23_self_attn_linear_k_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_k_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578788608))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579838336))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579837248)))]; tensor module_layers_23_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579840448)))]; tensor linear_211_cast_fp16 = linear(bias = module_layers_23_self_attn_linear_k_bias_to_fp16, weight = module_layers_23_self_attn_linear_k_weight_to_fp16_quantized, x = query_cast_fp16)[name = tensor("linear_211_cast_fp16")]; tensor var_4261 = const()[name = tensor("op_4261"), val = tensor([1, -1, 8, 128])]; tensor k_93_cast_fp16 = reshape(shape = var_4261, x = linear_211_cast_fp16)[name = tensor("k_93_cast_fp16")]; tensor module_layers_23_self_attn_linear_v_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_v_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579842560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580892288))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580891200)))]; tensor module_layers_23_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580894400)))]; tensor linear_212_cast_fp16 = linear(bias = module_layers_23_self_attn_linear_v_bias_to_fp16, weight = module_layers_23_self_attn_linear_v_weight_to_fp16_quantized, x = query_cast_fp16)[name = tensor("linear_212_cast_fp16")]; tensor var_4266 = const()[name = tensor("op_4266"), val = tensor([1, -1, 8, 128])]; tensor v_cast_fp16 = reshape(shape = var_4266, x = linear_212_cast_fp16)[name = tensor("v_cast_fp16")]; tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, -3, -1])]; tensor module_layers_23_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580896512)))]; tensor var_4278_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_u_to_fp16)[name = tensor("op_4278_cast_fp16")]; tensor module_layers_23_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580898624)))]; tensor var_4280_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_v_to_fp16)[name = tensor("op_4280_cast_fp16")]; tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, -3, -1])]; tensor x_515_transpose_x_0 = const()[name = tensor("x_515_transpose_x_0"), val = tensor(false)]; tensor x_515_transpose_y_0 = const()[name = tensor("x_515_transpose_y_0"), val = tensor(false)]; tensor op_4282_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4282_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580900736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(581285248))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(581284800)))]; tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_4280_cast_fp16)[name = tensor("transpose_151")]; tensor x_515_cast_fp16 = matmul(transpose_x = x_515_transpose_x_0, transpose_y = x_515_transpose_y_0, x = q_with_bias_v_cast_fp16, y = op_4282_to_fp16_quantized)[name = tensor("x_515_cast_fp16")]; tensor x_517_pad_0 = const()[name = tensor("x_517_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_517_mode_0 = const()[name = tensor("x_517_mode_0"), val = tensor("constant")]; tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(0x0p+0)]; tensor x_517_cast_fp16 = pad(constant_val = const_244_to_fp16, mode = x_517_mode_0, pad = x_517_pad_0, x = x_515_cast_fp16)[name = tensor("x_517_cast_fp16")]; tensor var_4290 = const()[name = tensor("op_4290"), val = tensor([1, 8, -1, 188])]; tensor x_519_cast_fp16 = reshape(shape = var_4290, x = x_517_cast_fp16)[name = tensor("x_519_cast_fp16")]; tensor var_4294_begin_0 = const()[name = tensor("op_4294_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_4294_end_0 = const()[name = tensor("op_4294_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_4294_end_mask_0 = const()[name = tensor("op_4294_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_4294_cast_fp16 = slice_by_index(begin = var_4294_begin_0, end = var_4294_end_0, end_mask = var_4294_end_mask_0, x = x_519_cast_fp16)[name = tensor("op_4294_cast_fp16")]; tensor var_4295 = const()[name = tensor("op_4295"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_93_cast_fp16 = reshape(shape = var_4295, x = var_4294_cast_fp16)[name = tensor("matrix_bd_93_cast_fp16")]; tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; tensor transpose_142_perm_0 = const()[name = tensor("transpose_142_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_143_perm_0 = const()[name = tensor("transpose_143_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_143 = transpose(perm = transpose_143_perm_0, x = k_93_cast_fp16)[name = tensor("transpose_149")]; tensor transpose_142 = transpose(perm = transpose_142_perm_0, x = var_4278_cast_fp16)[name = tensor("transpose_150")]; tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_142, y = transpose_143)[name = tensor("matrix_ac_cast_fp16")]; tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_93_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; tensor var_4304_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_4304_cast_fp16")]; tensor _inversed_scores_93_y_0_to_fp16 = const()[name = tensor("_inversed_scores_93_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_93_cast_fp16 = mul(x = var_4304_cast_fp16, y = _inversed_scores_93_y_0_to_fp16)[name = tensor("_inversed_scores_93_cast_fp16")]; tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_93_cast_fp16, cond = mask_3)[name = tensor("scores_cast_fp16")]; tensor var_4310_cast_fp16 = softmax(axis = var_30, x = scores_cast_fp16)[name = tensor("op_4310_cast_fp16")]; tensor input_1229_cast_fp16 = select(a = var_11_to_fp16, b = var_4310_cast_fp16, cond = mask_3)[name = tensor("input_1229_cast_fp16")]; tensor x_521_transpose_x_0 = const()[name = tensor("x_521_transpose_x_0"), val = tensor(false)]; tensor x_521_transpose_y_0 = const()[name = tensor("x_521_transpose_y_0"), val = tensor(false)]; tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_148")]; tensor x_521_cast_fp16 = matmul(transpose_x = x_521_transpose_x_0, transpose_y = x_521_transpose_y_0, x = input_1229_cast_fp16, y = value_cast_fp16)[name = tensor("x_521_cast_fp16")]; tensor var_4314_perm_0 = const()[name = tensor("op_4314_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4315 = const()[name = tensor("op_4315"), val = tensor([1, -1, 1024])]; tensor var_4314_cast_fp16 = transpose(perm = var_4314_perm_0, x = x_521_cast_fp16)[name = tensor("transpose_147")]; tensor input_1231_cast_fp16 = reshape(shape = var_4315, x = var_4314_cast_fp16)[name = tensor("input_1231_cast_fp16")]; tensor module_layers_23_self_attn_linear_out_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_out_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(581286080))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582335808))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582334720)))]; tensor module_layers_23_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582337920)))]; tensor linear_214_cast_fp16 = linear(bias = module_layers_23_self_attn_linear_out_bias_to_fp16, weight = module_layers_23_self_attn_linear_out_weight_to_fp16_quantized, x = input_1231_cast_fp16)[name = tensor("linear_214_cast_fp16")]; tensor input_1235_cast_fp16 = add(x = input_1227_cast_fp16, y = linear_214_cast_fp16)[name = tensor("input_1235_cast_fp16")]; tensor x_525_axes_0 = const()[name = tensor("x_525_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582340032)))]; tensor module_layers_23_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582342144)))]; tensor x_525_cast_fp16 = layer_norm(axes = x_525_axes_0, beta = module_layers_23_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_conv_weight_to_fp16, x = input_1235_cast_fp16)[name = tensor("x_525_cast_fp16")]; tensor input_1237_perm_0 = const()[name = tensor("input_1237_perm_0"), val = tensor([0, 2, 1])]; tensor input_1239_pad_type_0 = const()[name = tensor("input_1239_pad_type_0"), val = tensor("valid")]; tensor input_1239_strides_0 = const()[name = tensor("input_1239_strides_0"), val = tensor([1])]; tensor input_1239_pad_0 = const()[name = tensor("input_1239_pad_0"), val = tensor([0, 0])]; tensor input_1239_dilations_0 = const()[name = tensor("input_1239_dilations_0"), val = tensor([1])]; tensor input_1239_groups_0 = const()[name = tensor("input_1239_groups_0"), val = tensor(1)]; tensor module_layers_23_conv_pointwise_conv1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_conv_pointwise_conv1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582344256))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584443584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584441472)))]; tensor module_layers_23_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_23_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584447744)))]; tensor input_1237_cast_fp16 = transpose(perm = input_1237_perm_0, x = x_525_cast_fp16)[name = tensor("transpose_146")]; tensor input_1239_cast_fp16 = conv(bias = module_layers_23_conv_pointwise_conv1_bias_to_fp16, dilations = input_1239_dilations_0, groups = input_1239_groups_0, pad = input_1239_pad_0, pad_type = input_1239_pad_type_0, strides = input_1239_strides_0, weight = module_layers_23_conv_pointwise_conv1_weight_to_fp16_quantized, x = input_1237_cast_fp16)[name = tensor("input_1239_cast_fp16")]; tensor x_527_split_num_splits_0 = const()[name = tensor("x_527_split_num_splits_0"), val = tensor(2)]; tensor x_527_split_axis_0 = const()[name = tensor("x_527_split_axis_0"), val = tensor(1)]; tensor x_527_split_cast_fp16_0, tensor x_527_split_cast_fp16_1 = split(axis = x_527_split_axis_0, num_splits = x_527_split_num_splits_0, x = input_1239_cast_fp16)[name = tensor("x_527_split_cast_fp16")]; tensor x_527_split_1_sigmoid_cast_fp16 = sigmoid(x = x_527_split_cast_fp16_1)[name = tensor("x_527_split_1_sigmoid_cast_fp16")]; tensor x_527_cast_fp16 = mul(x = x_527_split_cast_fp16_0, y = x_527_split_1_sigmoid_cast_fp16)[name = tensor("x_527_cast_fp16")]; tensor input_1241_cast_fp16 = select(a = var_11_to_fp16, b = x_527_cast_fp16, cond = var_337)[name = tensor("input_1241_cast_fp16")]; tensor input_1243_pad_0 = const()[name = tensor("input_1243_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1243_mode_0 = const()[name = tensor("input_1243_mode_0"), val = tensor("constant")]; tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(0x0p+0)]; tensor input_1243_cast_fp16 = pad(constant_val = const_247_to_fp16, mode = input_1243_mode_0, pad = input_1243_pad_0, x = input_1241_cast_fp16)[name = tensor("input_1243_cast_fp16")]; tensor input_1245_pad_type_0 = const()[name = tensor("input_1245_pad_type_0"), val = tensor("valid")]; tensor input_1245_groups_0 = const()[name = tensor("input_1245_groups_0"), val = tensor(1024)]; tensor input_1245_strides_0 = const()[name = tensor("input_1245_strides_0"), val = tensor([1])]; tensor input_1245_pad_0 = const()[name = tensor("input_1245_pad_0"), val = tensor([0, 0])]; tensor input_1245_dilations_0 = const()[name = tensor("input_1245_dilations_0"), val = tensor([1])]; tensor const_294_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_294_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584451904))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584462272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584461184)))]; tensor const_295_to_fp16 = const()[name = tensor("const_295_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584464384)))]; tensor input_1247_cast_fp16 = conv(bias = const_295_to_fp16, dilations = input_1245_dilations_0, groups = input_1245_groups_0, pad = input_1245_pad_0, pad_type = input_1245_pad_type_0, strides = input_1245_strides_0, weight = const_294_to_fp16_quantized, x = input_1243_cast_fp16)[name = tensor("input_1247_cast_fp16")]; tensor input_1249_cast_fp16 = silu(x = input_1247_cast_fp16)[name = tensor("input_1249_cast_fp16")]; tensor x_529_pad_type_0 = const()[name = tensor("x_529_pad_type_0"), val = tensor("valid")]; tensor x_529_strides_0 = const()[name = tensor("x_529_strides_0"), val = tensor([1])]; tensor x_529_pad_0 = const()[name = tensor("x_529_pad_0"), val = tensor([0, 0])]; tensor x_529_dilations_0 = const()[name = tensor("x_529_dilations_0"), val = tensor([1])]; tensor x_529_groups_0 = const()[name = tensor("x_529_groups_0"), val = tensor(1)]; tensor module_layers_23_conv_pointwise_conv2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_conv_pointwise_conv2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584466496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585516224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585515136)))]; tensor module_layers_23_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_23_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585518336)))]; tensor x_529_cast_fp16 = conv(bias = module_layers_23_conv_pointwise_conv2_bias_to_fp16, dilations = x_529_dilations_0, groups = x_529_groups_0, pad = x_529_pad_0, pad_type = x_529_pad_type_0, strides = x_529_strides_0, weight = module_layers_23_conv_pointwise_conv2_weight_to_fp16_quantized, x = input_1249_cast_fp16)[name = tensor("x_529_cast_fp16")]; tensor input_1251_perm_0 = const()[name = tensor("input_1251_perm_0"), val = tensor([0, 2, 1])]; tensor input_1251_cast_fp16 = transpose(perm = input_1251_perm_0, x = x_529_cast_fp16)[name = tensor("transpose_145")]; tensor input_1253_cast_fp16 = add(x = input_1235_cast_fp16, y = input_1251_cast_fp16)[name = tensor("input_1253_cast_fp16")]; tensor input_1255_axes_0 = const()[name = tensor("input_1255_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585520448)))]; tensor module_layers_23_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585522560)))]; tensor input_1255_cast_fp16 = layer_norm(axes = input_1255_axes_0, beta = module_layers_23_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward2_weight_to_fp16, x = input_1253_cast_fp16)[name = tensor("input_1255_cast_fp16")]; tensor module_layers_23_feed_forward2_linear1_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward2_linear1_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585524672))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589723200))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589719040)))]; tensor module_layers_23_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_23_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589731456)))]; tensor linear_215_cast_fp16 = linear(bias = module_layers_23_feed_forward2_linear1_bias_to_fp16, weight = module_layers_23_feed_forward2_linear1_weight_to_fp16_quantized, x = input_1255_cast_fp16)[name = tensor("linear_215_cast_fp16")]; tensor input_1259_cast_fp16 = silu(x = linear_215_cast_fp16)[name = tensor("input_1259_cast_fp16")]; tensor module_layers_23_feed_forward2_linear2_weight_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward2_linear2_weight_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589739712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593935168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593934080)))]; tensor module_layers_23_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_23_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593937280)))]; tensor linear_216_cast_fp16 = linear(bias = module_layers_23_feed_forward2_linear2_bias_to_fp16, weight = module_layers_23_feed_forward2_linear2_weight_to_fp16_quantized, x = input_1259_cast_fp16)[name = tensor("linear_216_cast_fp16")]; tensor var_4381_to_fp16 = const()[name = tensor("op_4381_to_fp16"), val = tensor(0x1p-1)]; tensor var_4382_cast_fp16 = mul(x = linear_216_cast_fp16, y = var_4381_to_fp16)[name = tensor("op_4382_cast_fp16")]; tensor input_cast_fp16 = add(x = input_1253_cast_fp16, y = var_4382_cast_fp16)[name = tensor("input_cast_fp16")]; tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593939392)))]; tensor module_layers_23_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593941504)))]; tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_23_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_144")]; tensor encoder = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_0")]; } -> (encoder, encoder_length); }