program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0b1"}})] { func main(tensor mel, tensor mel_length) { tensor var_30 = const()[name = tensor("op_30"), val = tensor(-1)]; tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; tensor mel_to_fp16_dtype_0 = const()[name = tensor("mel_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_86_to_fp16_dtype_0 = const()[name = tensor("op_86_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_87_promoted_to_fp16 = const()[name = tensor("op_87_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor mel_length_to_fp16 = cast(dtype = var_86_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_229")]; tensor var_88_cast_fp16 = add(x = mel_length_to_fp16, y = var_87_promoted_to_fp16)[name = tensor("op_88_cast_fp16")]; tensor _inversed_90_y_0_to_fp16 = const()[name = tensor("_inversed_90_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_90_cast_fp16 = mul(x = var_88_cast_fp16, y = _inversed_90_y_0_to_fp16)[name = tensor("_inversed_90_cast_fp16")]; tensor var_91_to_fp16 = const()[name = tensor("op_91_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_1_cast_fp16 = add(x = _inversed_90_cast_fp16, y = var_91_to_fp16)[name = tensor("lengths_1_cast_fp16")]; tensor lengths_3_cast_fp16 = floor(x = lengths_1_cast_fp16)[name = tensor("lengths_3_cast_fp16")]; tensor var_95_promoted_to_fp16 = const()[name = tensor("op_95_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor var_96_cast_fp16 = add(x = lengths_3_cast_fp16, y = var_95_promoted_to_fp16)[name = tensor("op_96_cast_fp16")]; tensor _inversed_98_y_0_to_fp16 = const()[name = tensor("_inversed_98_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_98_cast_fp16 = mul(x = var_96_cast_fp16, y = _inversed_98_y_0_to_fp16)[name = tensor("_inversed_98_cast_fp16")]; tensor var_99_to_fp16 = const()[name = tensor("op_99_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_7_cast_fp16 = add(x = _inversed_98_cast_fp16, y = var_99_to_fp16)[name = tensor("lengths_7_cast_fp16")]; tensor lengths_9_cast_fp16 = floor(x = lengths_7_cast_fp16)[name = tensor("lengths_9_cast_fp16")]; tensor var_103_promoted_to_fp16 = const()[name = tensor("op_103_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor var_104_cast_fp16 = add(x = lengths_9_cast_fp16, y = var_103_promoted_to_fp16)[name = tensor("op_104_cast_fp16")]; tensor _inversed_106_y_0_to_fp16 = const()[name = tensor("_inversed_106_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_106_cast_fp16 = mul(x = var_104_cast_fp16, y = _inversed_106_y_0_to_fp16)[name = tensor("_inversed_106_cast_fp16")]; tensor var_107_to_fp16 = const()[name = tensor("op_107_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_13_cast_fp16 = add(x = _inversed_106_cast_fp16, y = var_107_to_fp16)[name = tensor("lengths_13_cast_fp16")]; tensor lengths_cast_fp16 = floor(x = lengths_13_cast_fp16)[name = tensor("lengths_cast_fp16")]; tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; tensor mel_to_fp16 = cast(dtype = mel_to_fp16_dtype_0, x = mel)[name = tensor("cast_230")]; tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_to_fp16)[name = tensor("transpose_291")]; tensor input_1_cast_fp16 = expand_dims(axes = input_1_axes_0, x = x_1_cast_fp16)[name = tensor("input_1_cast_fp16")]; tensor input_3_pad_type_0 = const()[name = tensor("input_3_pad_type_0"), val = tensor("custom")]; tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_3_strides_0 = const()[name = tensor("input_3_strides_0"), val = tensor([2, 2])]; tensor input_3_dilations_0 = const()[name = tensor("input_3_dilations_0"), val = tensor([1, 1])]; tensor input_3_groups_0 = const()[name = tensor("input_3_groups_0"), val = tensor(1)]; tensor module_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; tensor input_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = input_3_dilations_0, groups = input_3_groups_0, pad = input_3_pad_0, pad_type = input_3_pad_type_0, strides = input_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; tensor input_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("custom")]; tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_7_strides_0 = const()[name = tensor("input_7_strides_0"), val = tensor([2, 2])]; tensor input_7_groups_0 = const()[name = tensor("input_7_groups_0"), val = tensor(256)]; tensor input_7_dilations_0 = const()[name = tensor("input_7_dilations_0"), val = tensor([1, 1])]; tensor module_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; tensor input_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; tensor input_9_pad_type_0 = const()[name = tensor("input_9_pad_type_0"), val = tensor("valid")]; tensor input_9_strides_0 = const()[name = tensor("input_9_strides_0"), val = tensor([1, 1])]; tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_9_dilations_0 = const()[name = tensor("input_9_dilations_0"), val = tensor([1, 1])]; tensor input_9_groups_0 = const()[name = tensor("input_9_groups_0"), val = tensor(1)]; tensor module_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; tensor input_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; tensor input_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("custom")]; tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([2, 2])]; tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(256)]; tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; tensor module_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; tensor input_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("valid")]; tensor input_15_strides_0 = const()[name = tensor("input_15_strides_0"), val = tensor([1, 1])]; tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_15_dilations_0 = const()[name = tensor("input_15_dilations_0"), val = tensor([1, 1])]; tensor input_15_groups_0 = const()[name = tensor("input_15_groups_0"), val = tensor(1)]; tensor module_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; tensor input_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = input_15_dilations_0, groups = input_15_groups_0, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = input_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; tensor x_3_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("x_3_cast_fp16")]; tensor var_157_perm_0 = const()[name = tensor("op_157_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_158 = const()[name = tensor("op_158"), val = tensor([1, 188, -1])]; tensor var_157_cast_fp16 = transpose(perm = var_157_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_290")]; tensor input_17_cast_fp16 = reshape(shape = var_158, x = var_157_cast_fp16)[name = tensor("input_17_cast_fp16")]; tensor module_pre_encode_out_weight_to_fp16 = const()[name = tensor("module_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8667904)))]; tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; tensor padding_length_dtype_0 = const()[name = tensor("padding_length_dtype_0"), val = tensor("int32")]; tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; tensor var_196_axes_0 = const()[name = tensor("op_196_axes_0"), val = tensor([-1])]; tensor encoder_length = cast(dtype = padding_length_dtype_0, x = lengths_cast_fp16)[name = tensor("cast_228")]; tensor var_196 = expand_dims(axes = var_196_axes_0, x = encoder_length)[name = tensor("op_196")]; tensor pad_mask_1 = less(x = expand_dims_0, y = var_196)[name = tensor("pad_mask_1")]; tensor var_198_axes_0 = const()[name = tensor("op_198_axes_0"), val = tensor([1])]; tensor var_198 = expand_dims(axes = var_198_axes_0, x = pad_mask_1)[name = tensor("op_198")]; tensor var_199 = const()[name = tensor("op_199"), val = tensor([1, 188, 1])]; tensor pad_mask_for_att_mask_1 = tile(reps = var_199, x = var_198)[name = tensor("pad_mask_for_att_mask_1")]; tensor var_201_perm_0 = const()[name = tensor("op_201_perm_0"), val = tensor([0, 2, 1])]; tensor var_201 = transpose(perm = var_201_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_289")]; tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_201)[name = tensor("pad_mask_for_att_mask")]; tensor const_7 = const()[name = tensor("const_7"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_7)[name = tensor("att_mask")]; tensor mask_1 = logical_not(x = att_mask)[name = tensor("mask_1")]; tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8670016)))]; tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8672128)))]; tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; tensor module_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8674240)))]; tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17062912)))]; tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; tensor module_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17071168)))]; tensor linear_2_bias_0_to_fp16 = const()[name = tensor("linear_2_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25459840)))]; tensor linear_2_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; tensor var_232_to_fp16 = const()[name = tensor("op_232_to_fp16"), val = tensor(0x1p-1)]; tensor var_233_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_232_to_fp16)[name = tensor("op_233_cast_fp16")]; tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_233_cast_fp16)[name = tensor("input_31_cast_fp16")]; tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25461952)))]; tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25464064)))]; tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; tensor module_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25466176)))]; tensor linear_3_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; tensor var_249 = const()[name = tensor("op_249"), val = tensor([1, -1, 8, 128])]; tensor q_1_cast_fp16 = reshape(shape = var_249, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; tensor module_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27563392)))]; tensor linear_4_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; tensor var_253 = const()[name = tensor("op_253"), val = tensor([1, -1, 8, 128])]; tensor k_1_cast_fp16 = reshape(shape = var_253, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; tensor module_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29660608)))]; tensor linear_5_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; tensor var_257 = const()[name = tensor("op_257"), val = tensor([1, -1, 8, 128])]; tensor v_1_cast_fp16 = reshape(shape = var_257, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31757824)))]; tensor var_269_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_269_cast_fp16")]; tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31759936)))]; tensor var_271_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_271_cast_fp16")]; tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; tensor var_273_to_fp16 = const()[name = tensor("op_273_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31762048)))]; tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_271_cast_fp16)[name = tensor("transpose_287")]; tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_273_to_fp16)[name = tensor("x_7_cast_fp16")]; tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; tensor const_14_to_fp16 = const()[name = tensor("const_14_to_fp16"), val = tensor(0x0p+0)]; tensor x_9_cast_fp16 = pad(constant_val = const_14_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; tensor var_281 = const()[name = tensor("op_281"), val = tensor([1, 8, -1, 188])]; tensor x_11_cast_fp16 = reshape(shape = var_281, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; tensor var_285_begin_0 = const()[name = tensor("op_285_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_285_end_0 = const()[name = tensor("op_285_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_285_end_mask_0 = const()[name = tensor("op_285_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_285_cast_fp16 = slice_by_index(begin = var_285_begin_0, end = var_285_end_0, end_mask = var_285_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_285_cast_fp16")]; tensor var_286 = const()[name = tensor("op_286"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_1_cast_fp16 = reshape(shape = var_286, x = var_285_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_285")]; tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = var_269_cast_fp16)[name = tensor("transpose_286")]; tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_72, y = transpose_73)[name = tensor("matrix_ac_1_cast_fp16")]; tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; tensor var_295_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_295_cast_fp16")]; tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_1_cast_fp16 = mul(x = var_295_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_3)[name = tensor("scores_3_cast_fp16")]; tensor var_301_cast_fp16 = softmax(axis = var_30, x = scores_3_cast_fp16)[name = tensor("op_301_cast_fp16")]; tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_301_cast_fp16, cond = mask_3)[name = tensor("input_33_cast_fp16")]; tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_288")]; tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; tensor var_305_perm_0 = const()[name = tensor("op_305_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_306 = const()[name = tensor("op_306"), val = tensor([1, -1, 1024])]; tensor var_305_cast_fp16 = transpose(perm = var_305_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_284")]; tensor input_35_cast_fp16 = reshape(shape = var_306, x = var_305_cast_fp16)[name = tensor("input_35_cast_fp16")]; tensor module_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32530112)))]; tensor linear_7_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34627328)))]; tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34629440)))]; tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34631552)))]; tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_283")]; tensor input_43_cast_fp16 = conv(dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; tensor var_328_axes_0 = const()[name = tensor("op_328_axes_0"), val = tensor([1])]; tensor var_328 = expand_dims(axes = var_328_axes_0, x = pad_mask)[name = tensor("op_328")]; tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_328)[name = tensor("input_45_cast_fp16")]; tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; tensor const_17_to_fp16 = const()[name = tensor("const_17_to_fp16"), val = tensor(0x0p+0)]; tensor input_47_cast_fp16 = pad(constant_val = const_17_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(1024)]; tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38825920)))]; tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38844416)))]; tensor input_51_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_248_to_fp16, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38846528)))]; tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_282")]; tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40943744)))]; tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40945856)))]; tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; tensor module_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40947968)))]; tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; tensor module_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49336640)))]; tensor linear_9_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; tensor var_366_to_fp16 = const()[name = tensor("op_366_to_fp16"), val = tensor(0x1p-1)]; tensor var_367_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_366_to_fp16)[name = tensor("op_367_cast_fp16")]; tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_367_cast_fp16)[name = tensor("input_69_cast_fp16")]; tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57725312)))]; tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57727424)))]; tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57729536)))]; tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57731648)))]; tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; tensor module_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57733760)))]; tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; tensor module_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66122432)))]; tensor linear_11_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; tensor var_395_to_fp16 = const()[name = tensor("op_395_to_fp16"), val = tensor(0x1p-1)]; tensor var_396_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_395_to_fp16)[name = tensor("op_396_cast_fp16")]; tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_396_cast_fp16)[name = tensor("input_83_cast_fp16")]; tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74511104)))]; tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74513216)))]; tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; tensor module_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74515328)))]; tensor linear_12_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; tensor var_412 = const()[name = tensor("op_412"), val = tensor([1, -1, 8, 128])]; tensor q_7_cast_fp16 = reshape(shape = var_412, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; tensor module_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76612544)))]; tensor linear_13_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; tensor var_416 = const()[name = tensor("op_416"), val = tensor([1, -1, 8, 128])]; tensor k_5_cast_fp16 = reshape(shape = var_416, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; tensor module_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78709760)))]; tensor linear_14_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; tensor var_420 = const()[name = tensor("op_420"), val = tensor([1, -1, 8, 128])]; tensor v_3_cast_fp16 = reshape(shape = var_420, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80806976)))]; tensor var_432_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_432_cast_fp16")]; tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80809088)))]; tensor var_434_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_434_cast_fp16")]; tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; tensor var_436_to_fp16 = const()[name = tensor("op_436_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80811200)))]; tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_434_cast_fp16)[name = tensor("transpose_280")]; tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_436_to_fp16)[name = tensor("x_29_cast_fp16")]; tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; tensor const_24_to_fp16 = const()[name = tensor("const_24_to_fp16"), val = tensor(0x0p+0)]; tensor x_31_cast_fp16 = pad(constant_val = const_24_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; tensor var_444 = const()[name = tensor("op_444"), val = tensor([1, 8, -1, 188])]; tensor x_33_cast_fp16 = reshape(shape = var_444, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; tensor var_448_begin_0 = const()[name = tensor("op_448_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_448_end_0 = const()[name = tensor("op_448_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_448_end_mask_0 = const()[name = tensor("op_448_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_448_cast_fp16 = slice_by_index(begin = var_448_begin_0, end = var_448_end_0, end_mask = var_448_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_448_cast_fp16")]; tensor var_449 = const()[name = tensor("op_449"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_5_cast_fp16 = reshape(shape = var_449, x = var_448_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_278")]; tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = var_432_cast_fp16)[name = tensor("transpose_279")]; tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_74, y = transpose_75)[name = tensor("matrix_ac_3_cast_fp16")]; tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; tensor var_458_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_458_cast_fp16")]; tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_5_cast_fp16 = mul(x = var_458_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_3)[name = tensor("scores_7_cast_fp16")]; tensor var_464_cast_fp16 = softmax(axis = var_30, x = scores_7_cast_fp16)[name = tensor("op_464_cast_fp16")]; tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_464_cast_fp16, cond = mask_3)[name = tensor("input_85_cast_fp16")]; tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_281")]; tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; tensor var_468_perm_0 = const()[name = tensor("op_468_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_469 = const()[name = tensor("op_469"), val = tensor([1, -1, 1024])]; tensor var_468_cast_fp16 = transpose(perm = var_468_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_277")]; tensor input_87_cast_fp16 = reshape(shape = var_469, x = var_468_cast_fp16)[name = tensor("input_87_cast_fp16")]; tensor module_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81579264)))]; tensor linear_16_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83676480)))]; tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83678592)))]; tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83680704)))]; tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_276")]; tensor input_95_cast_fp16 = conv(dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_328)[name = tensor("input_97_cast_fp16")]; tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; tensor const_27_to_fp16 = const()[name = tensor("const_27_to_fp16"), val = tensor(0x0p+0)]; tensor input_99_cast_fp16 = pad(constant_val = const_27_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(1024)]; tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; tensor const_250_to_fp16 = const()[name = tensor("const_250_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87875072)))]; tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87893568)))]; tensor input_103_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_250_to_fp16, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87895680)))]; tensor x_43_cast_fp16 = conv(dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_275")]; tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89992896)))]; tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89995008)))]; tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; tensor module_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89997120)))]; tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; tensor module_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98385792)))]; tensor linear_18_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; tensor var_529_to_fp16 = const()[name = tensor("op_529_to_fp16"), val = tensor(0x1p-1)]; tensor var_530_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_529_to_fp16)[name = tensor("op_530_cast_fp16")]; tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_530_cast_fp16)[name = tensor("input_121_cast_fp16")]; tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106774464)))]; tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106776576)))]; tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106778688)))]; tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106780800)))]; tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; tensor module_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106782912)))]; tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; tensor module_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115171584)))]; tensor linear_20_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; tensor var_558_to_fp16 = const()[name = tensor("op_558_to_fp16"), val = tensor(0x1p-1)]; tensor var_559_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_558_to_fp16)[name = tensor("op_559_cast_fp16")]; tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_559_cast_fp16)[name = tensor("input_135_cast_fp16")]; tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123560256)))]; tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123562368)))]; tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; tensor module_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123564480)))]; tensor linear_21_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; tensor var_575 = const()[name = tensor("op_575"), val = tensor([1, -1, 8, 128])]; tensor q_13_cast_fp16 = reshape(shape = var_575, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; tensor module_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125661696)))]; tensor linear_22_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; tensor var_579 = const()[name = tensor("op_579"), val = tensor([1, -1, 8, 128])]; tensor k_9_cast_fp16 = reshape(shape = var_579, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; tensor module_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127758912)))]; tensor linear_23_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, -1, 8, 128])]; tensor v_5_cast_fp16 = reshape(shape = var_583, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129856128)))]; tensor var_595_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_595_cast_fp16")]; tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129858240)))]; tensor var_597_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_597_cast_fp16")]; tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; tensor var_599_to_fp16 = const()[name = tensor("op_599_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129860352)))]; tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_597_cast_fp16)[name = tensor("transpose_273")]; tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_599_to_fp16)[name = tensor("x_51_cast_fp16")]; tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; tensor const_34_to_fp16 = const()[name = tensor("const_34_to_fp16"), val = tensor(0x0p+0)]; tensor x_53_cast_fp16 = pad(constant_val = const_34_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; tensor var_607 = const()[name = tensor("op_607"), val = tensor([1, 8, -1, 188])]; tensor x_55_cast_fp16 = reshape(shape = var_607, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; tensor var_611_begin_0 = const()[name = tensor("op_611_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_611_end_0 = const()[name = tensor("op_611_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_611_end_mask_0 = const()[name = tensor("op_611_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_611_cast_fp16 = slice_by_index(begin = var_611_begin_0, end = var_611_end_0, end_mask = var_611_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_611_cast_fp16")]; tensor var_612 = const()[name = tensor("op_612"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_9_cast_fp16 = reshape(shape = var_612, x = var_611_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_271")]; tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = var_595_cast_fp16)[name = tensor("transpose_272")]; tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_76, y = transpose_77)[name = tensor("matrix_ac_5_cast_fp16")]; tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; tensor var_621_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_621_cast_fp16")]; tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_9_cast_fp16 = mul(x = var_621_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_3)[name = tensor("scores_11_cast_fp16")]; tensor var_627_cast_fp16 = softmax(axis = var_30, x = scores_11_cast_fp16)[name = tensor("op_627_cast_fp16")]; tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_627_cast_fp16, cond = mask_3)[name = tensor("input_137_cast_fp16")]; tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_274")]; tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; tensor var_631_perm_0 = const()[name = tensor("op_631_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_632 = const()[name = tensor("op_632"), val = tensor([1, -1, 1024])]; tensor var_631_cast_fp16 = transpose(perm = var_631_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_270")]; tensor input_139_cast_fp16 = reshape(shape = var_632, x = var_631_cast_fp16)[name = tensor("input_139_cast_fp16")]; tensor module_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130628416)))]; tensor linear_25_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132725632)))]; tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132727744)))]; tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132729856)))]; tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_269")]; tensor input_147_cast_fp16 = conv(dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_328)[name = tensor("input_149_cast_fp16")]; tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; tensor const_37_to_fp16 = const()[name = tensor("const_37_to_fp16"), val = tensor(0x0p+0)]; tensor input_151_cast_fp16 = pad(constant_val = const_37_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(1024)]; tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; tensor const_252_to_fp16 = const()[name = tensor("const_252_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136924224)))]; tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136942720)))]; tensor input_155_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_252_to_fp16, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136944832)))]; tensor x_65_cast_fp16 = conv(dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_268")]; tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139042048)))]; tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139044160)))]; tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; tensor module_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139046272)))]; tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; tensor module_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147434944)))]; tensor linear_27_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; tensor var_692_to_fp16 = const()[name = tensor("op_692_to_fp16"), val = tensor(0x1p-1)]; tensor var_693_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_692_to_fp16)[name = tensor("op_693_cast_fp16")]; tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_693_cast_fp16)[name = tensor("input_173_cast_fp16")]; tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155823616)))]; tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155825728)))]; tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155827840)))]; tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155829952)))]; tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; tensor module_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155832064)))]; tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; tensor module_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164220736)))]; tensor linear_29_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; tensor var_721_to_fp16 = const()[name = tensor("op_721_to_fp16"), val = tensor(0x1p-1)]; tensor var_722_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_721_to_fp16)[name = tensor("op_722_cast_fp16")]; tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_722_cast_fp16)[name = tensor("input_187_cast_fp16")]; tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172609408)))]; tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172611520)))]; tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; tensor module_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172613632)))]; tensor linear_30_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; tensor var_738 = const()[name = tensor("op_738"), val = tensor([1, -1, 8, 128])]; tensor q_19_cast_fp16 = reshape(shape = var_738, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; tensor module_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174710848)))]; tensor linear_31_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; tensor var_742 = const()[name = tensor("op_742"), val = tensor([1, -1, 8, 128])]; tensor k_13_cast_fp16 = reshape(shape = var_742, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; tensor module_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176808064)))]; tensor linear_32_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; tensor var_746 = const()[name = tensor("op_746"), val = tensor([1, -1, 8, 128])]; tensor v_7_cast_fp16 = reshape(shape = var_746, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178905280)))]; tensor var_758_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_758_cast_fp16")]; tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178907392)))]; tensor var_760_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_760_cast_fp16")]; tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; tensor var_762_to_fp16 = const()[name = tensor("op_762_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178909504)))]; tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_760_cast_fp16)[name = tensor("transpose_266")]; tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_762_to_fp16)[name = tensor("x_73_cast_fp16")]; tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; tensor const_44_to_fp16 = const()[name = tensor("const_44_to_fp16"), val = tensor(0x0p+0)]; tensor x_75_cast_fp16 = pad(constant_val = const_44_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; tensor var_770 = const()[name = tensor("op_770"), val = tensor([1, 8, -1, 188])]; tensor x_77_cast_fp16 = reshape(shape = var_770, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; tensor var_774_begin_0 = const()[name = tensor("op_774_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_774_end_0 = const()[name = tensor("op_774_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_774_end_mask_0 = const()[name = tensor("op_774_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_774_cast_fp16 = slice_by_index(begin = var_774_begin_0, end = var_774_end_0, end_mask = var_774_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_774_cast_fp16")]; tensor var_775 = const()[name = tensor("op_775"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_13_cast_fp16 = reshape(shape = var_775, x = var_774_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_264")]; tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = var_758_cast_fp16)[name = tensor("transpose_265")]; tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_78, y = transpose_79)[name = tensor("matrix_ac_7_cast_fp16")]; tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; tensor var_784_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_784_cast_fp16")]; tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_13_cast_fp16 = mul(x = var_784_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_3)[name = tensor("scores_15_cast_fp16")]; tensor var_790_cast_fp16 = softmax(axis = var_30, x = scores_15_cast_fp16)[name = tensor("op_790_cast_fp16")]; tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_790_cast_fp16, cond = mask_3)[name = tensor("input_189_cast_fp16")]; tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_267")]; tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; tensor var_794_perm_0 = const()[name = tensor("op_794_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_795 = const()[name = tensor("op_795"), val = tensor([1, -1, 1024])]; tensor var_794_cast_fp16 = transpose(perm = var_794_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_263")]; tensor input_191_cast_fp16 = reshape(shape = var_795, x = var_794_cast_fp16)[name = tensor("input_191_cast_fp16")]; tensor module_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179677568)))]; tensor linear_34_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181774784)))]; tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181776896)))]; tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181779008)))]; tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_262")]; tensor input_199_cast_fp16 = conv(dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_328)[name = tensor("input_201_cast_fp16")]; tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; tensor const_47_to_fp16 = const()[name = tensor("const_47_to_fp16"), val = tensor(0x0p+0)]; tensor input_203_cast_fp16 = pad(constant_val = const_47_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(1024)]; tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; tensor const_254_to_fp16 = const()[name = tensor("const_254_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185973376)))]; tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185991872)))]; tensor input_207_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_254_to_fp16, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185993984)))]; tensor x_87_cast_fp16 = conv(dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_261")]; tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188091200)))]; tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188093312)))]; tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; tensor module_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188095424)))]; tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; tensor module_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196484096)))]; tensor linear_36_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; tensor var_855_to_fp16 = const()[name = tensor("op_855_to_fp16"), val = tensor(0x1p-1)]; tensor var_856_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_855_to_fp16)[name = tensor("op_856_cast_fp16")]; tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_856_cast_fp16)[name = tensor("input_225_cast_fp16")]; tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204872768)))]; tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204874880)))]; tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204876992)))]; tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204879104)))]; tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; tensor module_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204881216)))]; tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; tensor module_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213269888)))]; tensor linear_38_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; tensor var_884_to_fp16 = const()[name = tensor("op_884_to_fp16"), val = tensor(0x1p-1)]; tensor var_885_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_884_to_fp16)[name = tensor("op_885_cast_fp16")]; tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_885_cast_fp16)[name = tensor("input_239_cast_fp16")]; tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(221658560)))]; tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(221660672)))]; tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; tensor module_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(221662784)))]; tensor linear_39_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; tensor var_901 = const()[name = tensor("op_901"), val = tensor([1, -1, 8, 128])]; tensor q_25_cast_fp16 = reshape(shape = var_901, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; tensor module_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223760000)))]; tensor linear_40_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; tensor var_905 = const()[name = tensor("op_905"), val = tensor([1, -1, 8, 128])]; tensor k_17_cast_fp16 = reshape(shape = var_905, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; tensor module_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225857216)))]; tensor linear_41_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; tensor var_909 = const()[name = tensor("op_909"), val = tensor([1, -1, 8, 128])]; tensor v_9_cast_fp16 = reshape(shape = var_909, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227954432)))]; tensor var_921_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_921_cast_fp16")]; tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227956544)))]; tensor var_923_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_923_cast_fp16")]; tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; tensor var_925_to_fp16 = const()[name = tensor("op_925_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227958656)))]; tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_923_cast_fp16)[name = tensor("transpose_259")]; tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_925_to_fp16)[name = tensor("x_95_cast_fp16")]; tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; tensor const_54_to_fp16 = const()[name = tensor("const_54_to_fp16"), val = tensor(0x0p+0)]; tensor x_97_cast_fp16 = pad(constant_val = const_54_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; tensor var_933 = const()[name = tensor("op_933"), val = tensor([1, 8, -1, 188])]; tensor x_99_cast_fp16 = reshape(shape = var_933, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; tensor var_937_begin_0 = const()[name = tensor("op_937_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_937_end_0 = const()[name = tensor("op_937_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_937_end_mask_0 = const()[name = tensor("op_937_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_937_cast_fp16 = slice_by_index(begin = var_937_begin_0, end = var_937_end_0, end_mask = var_937_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_937_cast_fp16")]; tensor var_938 = const()[name = tensor("op_938"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_17_cast_fp16 = reshape(shape = var_938, x = var_937_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_257")]; tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = var_921_cast_fp16)[name = tensor("transpose_258")]; tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_80, y = transpose_81)[name = tensor("matrix_ac_9_cast_fp16")]; tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; tensor var_947_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_947_cast_fp16")]; tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_17_cast_fp16 = mul(x = var_947_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_3)[name = tensor("scores_19_cast_fp16")]; tensor var_953_cast_fp16 = softmax(axis = var_30, x = scores_19_cast_fp16)[name = tensor("op_953_cast_fp16")]; tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_953_cast_fp16, cond = mask_3)[name = tensor("input_241_cast_fp16")]; tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_260")]; tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; tensor var_957_perm_0 = const()[name = tensor("op_957_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_958 = const()[name = tensor("op_958"), val = tensor([1, -1, 1024])]; tensor var_957_cast_fp16 = transpose(perm = var_957_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_256")]; tensor input_243_cast_fp16 = reshape(shape = var_958, x = var_957_cast_fp16)[name = tensor("input_243_cast_fp16")]; tensor module_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(228726720)))]; tensor linear_43_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230823936)))]; tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230826048)))]; tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230828160)))]; tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_255")]; tensor input_251_cast_fp16 = conv(dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_328)[name = tensor("input_253_cast_fp16")]; tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; tensor const_57_to_fp16 = const()[name = tensor("const_57_to_fp16"), val = tensor(0x0p+0)]; tensor input_255_cast_fp16 = pad(constant_val = const_57_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(1024)]; tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; tensor const_256_to_fp16 = const()[name = tensor("const_256_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235022528)))]; tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235041024)))]; tensor input_259_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_256_to_fp16, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235043136)))]; tensor x_109_cast_fp16 = conv(dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_254")]; tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237140352)))]; tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237142464)))]; tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; tensor module_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237144576)))]; tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; tensor module_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245533248)))]; tensor linear_45_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; tensor var_1018_to_fp16 = const()[name = tensor("op_1018_to_fp16"), val = tensor(0x1p-1)]; tensor var_1019_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1018_to_fp16)[name = tensor("op_1019_cast_fp16")]; tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1019_cast_fp16)[name = tensor("input_277_cast_fp16")]; tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253921920)))]; tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253924032)))]; tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253926144)))]; tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253928256)))]; tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; tensor module_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253930368)))]; tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; tensor module_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262319040)))]; tensor linear_47_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; tensor var_1047_to_fp16 = const()[name = tensor("op_1047_to_fp16"), val = tensor(0x1p-1)]; tensor var_1048_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1047_to_fp16)[name = tensor("op_1048_cast_fp16")]; tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1048_cast_fp16)[name = tensor("input_291_cast_fp16")]; tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270707712)))]; tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270709824)))]; tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; tensor module_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270711936)))]; tensor linear_48_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; tensor var_1064 = const()[name = tensor("op_1064"), val = tensor([1, -1, 8, 128])]; tensor q_31_cast_fp16 = reshape(shape = var_1064, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; tensor module_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272809152)))]; tensor linear_49_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; tensor var_1068 = const()[name = tensor("op_1068"), val = tensor([1, -1, 8, 128])]; tensor k_21_cast_fp16 = reshape(shape = var_1068, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; tensor module_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274906368)))]; tensor linear_50_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; tensor var_1072 = const()[name = tensor("op_1072"), val = tensor([1, -1, 8, 128])]; tensor v_11_cast_fp16 = reshape(shape = var_1072, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277003584)))]; tensor var_1084_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1084_cast_fp16")]; tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277005696)))]; tensor var_1086_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1086_cast_fp16")]; tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; tensor var_1088_to_fp16 = const()[name = tensor("op_1088_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277007808)))]; tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1086_cast_fp16)[name = tensor("transpose_252")]; tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1088_to_fp16)[name = tensor("x_117_cast_fp16")]; tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; tensor const_64_to_fp16 = const()[name = tensor("const_64_to_fp16"), val = tensor(0x0p+0)]; tensor x_119_cast_fp16 = pad(constant_val = const_64_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; tensor var_1096 = const()[name = tensor("op_1096"), val = tensor([1, 8, -1, 188])]; tensor x_121_cast_fp16 = reshape(shape = var_1096, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; tensor var_1100_begin_0 = const()[name = tensor("op_1100_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1100_end_0 = const()[name = tensor("op_1100_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1100_end_mask_0 = const()[name = tensor("op_1100_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1100_cast_fp16 = slice_by_index(begin = var_1100_begin_0, end = var_1100_end_0, end_mask = var_1100_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1100_cast_fp16")]; tensor var_1101 = const()[name = tensor("op_1101"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1101, x = var_1100_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_250")]; tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = var_1084_cast_fp16)[name = tensor("transpose_251")]; tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_82, y = transpose_83)[name = tensor("matrix_ac_11_cast_fp16")]; tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; tensor var_1110_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1110_cast_fp16")]; tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_21_cast_fp16 = mul(x = var_1110_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_3)[name = tensor("scores_23_cast_fp16")]; tensor var_1116_cast_fp16 = softmax(axis = var_30, x = scores_23_cast_fp16)[name = tensor("op_1116_cast_fp16")]; tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1116_cast_fp16, cond = mask_3)[name = tensor("input_293_cast_fp16")]; tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_253")]; tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; tensor var_1120_perm_0 = const()[name = tensor("op_1120_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1121 = const()[name = tensor("op_1121"), val = tensor([1, -1, 1024])]; tensor var_1120_cast_fp16 = transpose(perm = var_1120_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_249")]; tensor input_295_cast_fp16 = reshape(shape = var_1121, x = var_1120_cast_fp16)[name = tensor("input_295_cast_fp16")]; tensor module_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277775872)))]; tensor linear_52_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279873088)))]; tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279875200)))]; tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279877312)))]; tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_248")]; tensor input_303_cast_fp16 = conv(dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_328)[name = tensor("input_305_cast_fp16")]; tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; tensor const_67_to_fp16 = const()[name = tensor("const_67_to_fp16"), val = tensor(0x0p+0)]; tensor input_307_cast_fp16 = pad(constant_val = const_67_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(1024)]; tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; tensor const_258_to_fp16 = const()[name = tensor("const_258_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284071680)))]; tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284090176)))]; tensor input_311_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_258_to_fp16, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284092288)))]; tensor x_131_cast_fp16 = conv(dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_247")]; tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286189504)))]; tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286191616)))]; tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; tensor module_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286193728)))]; tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; tensor module_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294582400)))]; tensor linear_54_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; tensor var_1181_to_fp16 = const()[name = tensor("op_1181_to_fp16"), val = tensor(0x1p-1)]; tensor var_1182_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1181_to_fp16)[name = tensor("op_1182_cast_fp16")]; tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1182_cast_fp16)[name = tensor("input_329_cast_fp16")]; tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302971072)))]; tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302973184)))]; tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302975296)))]; tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302977408)))]; tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; tensor module_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302979520)))]; tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; tensor module_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311368192)))]; tensor linear_56_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; tensor var_1210_to_fp16 = const()[name = tensor("op_1210_to_fp16"), val = tensor(0x1p-1)]; tensor var_1211_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1210_to_fp16)[name = tensor("op_1211_cast_fp16")]; tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1211_cast_fp16)[name = tensor("input_343_cast_fp16")]; tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319756864)))]; tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319758976)))]; tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; tensor module_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319761088)))]; tensor linear_57_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; tensor var_1227 = const()[name = tensor("op_1227"), val = tensor([1, -1, 8, 128])]; tensor q_37_cast_fp16 = reshape(shape = var_1227, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; tensor module_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321858304)))]; tensor linear_58_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; tensor var_1231 = const()[name = tensor("op_1231"), val = tensor([1, -1, 8, 128])]; tensor k_25_cast_fp16 = reshape(shape = var_1231, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; tensor module_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323955520)))]; tensor linear_59_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; tensor var_1235 = const()[name = tensor("op_1235"), val = tensor([1, -1, 8, 128])]; tensor v_13_cast_fp16 = reshape(shape = var_1235, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326052736)))]; tensor var_1247_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1247_cast_fp16")]; tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326054848)))]; tensor var_1249_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1249_cast_fp16")]; tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; tensor var_1251_to_fp16 = const()[name = tensor("op_1251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326056960)))]; tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1249_cast_fp16)[name = tensor("transpose_245")]; tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1251_to_fp16)[name = tensor("x_139_cast_fp16")]; tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; tensor const_74_to_fp16 = const()[name = tensor("const_74_to_fp16"), val = tensor(0x0p+0)]; tensor x_141_cast_fp16 = pad(constant_val = const_74_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; tensor var_1259 = const()[name = tensor("op_1259"), val = tensor([1, 8, -1, 188])]; tensor x_143_cast_fp16 = reshape(shape = var_1259, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; tensor var_1263_begin_0 = const()[name = tensor("op_1263_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1263_end_0 = const()[name = tensor("op_1263_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1263_end_mask_0 = const()[name = tensor("op_1263_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1263_cast_fp16 = slice_by_index(begin = var_1263_begin_0, end = var_1263_end_0, end_mask = var_1263_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1263_cast_fp16")]; tensor var_1264 = const()[name = tensor("op_1264"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1264, x = var_1263_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_85_perm_0 = const()[name = tensor("transpose_85_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_85 = transpose(perm = transpose_85_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_243")]; tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = var_1247_cast_fp16)[name = tensor("transpose_244")]; tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_84, y = transpose_85)[name = tensor("matrix_ac_13_cast_fp16")]; tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; tensor var_1273_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1273_cast_fp16")]; tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_25_cast_fp16 = mul(x = var_1273_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_3)[name = tensor("scores_27_cast_fp16")]; tensor var_1279_cast_fp16 = softmax(axis = var_30, x = scores_27_cast_fp16)[name = tensor("op_1279_cast_fp16")]; tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1279_cast_fp16, cond = mask_3)[name = tensor("input_345_cast_fp16")]; tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_246")]; tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; tensor var_1283_perm_0 = const()[name = tensor("op_1283_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1284 = const()[name = tensor("op_1284"), val = tensor([1, -1, 1024])]; tensor var_1283_cast_fp16 = transpose(perm = var_1283_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_242")]; tensor input_347_cast_fp16 = reshape(shape = var_1284, x = var_1283_cast_fp16)[name = tensor("input_347_cast_fp16")]; tensor module_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326825024)))]; tensor linear_61_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328922240)))]; tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328924352)))]; tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328926464)))]; tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_241")]; tensor input_355_cast_fp16 = conv(dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_328)[name = tensor("input_357_cast_fp16")]; tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; tensor const_77_to_fp16 = const()[name = tensor("const_77_to_fp16"), val = tensor(0x0p+0)]; tensor input_359_cast_fp16 = pad(constant_val = const_77_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(1024)]; tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; tensor const_260_to_fp16 = const()[name = tensor("const_260_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333120832)))]; tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333139328)))]; tensor input_363_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_260_to_fp16, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333141440)))]; tensor x_153_cast_fp16 = conv(dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_240")]; tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335238656)))]; tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335240768)))]; tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; tensor module_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335242880)))]; tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; tensor module_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343631552)))]; tensor linear_63_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; tensor var_1344_to_fp16 = const()[name = tensor("op_1344_to_fp16"), val = tensor(0x1p-1)]; tensor var_1345_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1344_to_fp16)[name = tensor("op_1345_cast_fp16")]; tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1345_cast_fp16)[name = tensor("input_381_cast_fp16")]; tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352020224)))]; tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352022336)))]; tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352024448)))]; tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352026560)))]; tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; tensor module_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352028672)))]; tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; tensor module_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360417344)))]; tensor linear_65_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; tensor var_1373_to_fp16 = const()[name = tensor("op_1373_to_fp16"), val = tensor(0x1p-1)]; tensor var_1374_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1373_to_fp16)[name = tensor("op_1374_cast_fp16")]; tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1374_cast_fp16)[name = tensor("input_395_cast_fp16")]; tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368806016)))]; tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368808128)))]; tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; tensor module_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368810240)))]; tensor linear_66_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; tensor var_1390 = const()[name = tensor("op_1390"), val = tensor([1, -1, 8, 128])]; tensor q_43_cast_fp16 = reshape(shape = var_1390, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; tensor module_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(370907456)))]; tensor linear_67_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; tensor var_1394 = const()[name = tensor("op_1394"), val = tensor([1, -1, 8, 128])]; tensor k_29_cast_fp16 = reshape(shape = var_1394, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; tensor module_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(373004672)))]; tensor linear_68_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; tensor var_1398 = const()[name = tensor("op_1398"), val = tensor([1, -1, 8, 128])]; tensor v_15_cast_fp16 = reshape(shape = var_1398, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375101888)))]; tensor var_1410_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1410_cast_fp16")]; tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375104000)))]; tensor var_1412_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1412_cast_fp16")]; tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; tensor var_1414_to_fp16 = const()[name = tensor("op_1414_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375106112)))]; tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1412_cast_fp16)[name = tensor("transpose_238")]; tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1414_to_fp16)[name = tensor("x_161_cast_fp16")]; tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; tensor const_84_to_fp16 = const()[name = tensor("const_84_to_fp16"), val = tensor(0x0p+0)]; tensor x_163_cast_fp16 = pad(constant_val = const_84_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; tensor var_1422 = const()[name = tensor("op_1422"), val = tensor([1, 8, -1, 188])]; tensor x_165_cast_fp16 = reshape(shape = var_1422, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; tensor var_1426_begin_0 = const()[name = tensor("op_1426_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1426_end_0 = const()[name = tensor("op_1426_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1426_end_mask_0 = const()[name = tensor("op_1426_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1426_cast_fp16 = slice_by_index(begin = var_1426_begin_0, end = var_1426_end_0, end_mask = var_1426_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1426_cast_fp16")]; tensor var_1427 = const()[name = tensor("op_1427"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1427, x = var_1426_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; tensor transpose_86_perm_0 = const()[name = tensor("transpose_86_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_87_perm_0 = const()[name = tensor("transpose_87_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_87 = transpose(perm = transpose_87_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_236")]; tensor transpose_86 = transpose(perm = transpose_86_perm_0, x = var_1410_cast_fp16)[name = tensor("transpose_237")]; tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_86, y = transpose_87)[name = tensor("matrix_ac_15_cast_fp16")]; tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; tensor var_1436_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1436_cast_fp16")]; tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_29_cast_fp16 = mul(x = var_1436_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_3)[name = tensor("scores_31_cast_fp16")]; tensor var_1442_cast_fp16 = softmax(axis = var_30, x = scores_31_cast_fp16)[name = tensor("op_1442_cast_fp16")]; tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1442_cast_fp16, cond = mask_3)[name = tensor("input_397_cast_fp16")]; tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_239")]; tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; tensor var_1446_perm_0 = const()[name = tensor("op_1446_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1447 = const()[name = tensor("op_1447"), val = tensor([1, -1, 1024])]; tensor var_1446_cast_fp16 = transpose(perm = var_1446_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_235")]; tensor input_399_cast_fp16 = reshape(shape = var_1447, x = var_1446_cast_fp16)[name = tensor("input_399_cast_fp16")]; tensor module_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375874176)))]; tensor linear_70_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377971392)))]; tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377973504)))]; tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377975616)))]; tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_234")]; tensor input_407_cast_fp16 = conv(dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_328)[name = tensor("input_409_cast_fp16")]; tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; tensor const_87_to_fp16 = const()[name = tensor("const_87_to_fp16"), val = tensor(0x0p+0)]; tensor input_411_cast_fp16 = pad(constant_val = const_87_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(1024)]; tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; tensor const_262_to_fp16 = const()[name = tensor("const_262_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382169984)))]; tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382188480)))]; tensor input_415_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_262_to_fp16, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382190592)))]; tensor x_175_cast_fp16 = conv(dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_233")]; tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384287808)))]; tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384289920)))]; tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; tensor module_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384292032)))]; tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; tensor module_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392680704)))]; tensor linear_72_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; tensor var_1507_to_fp16 = const()[name = tensor("op_1507_to_fp16"), val = tensor(0x1p-1)]; tensor var_1508_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1507_to_fp16)[name = tensor("op_1508_cast_fp16")]; tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1508_cast_fp16)[name = tensor("input_433_cast_fp16")]; tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401069376)))]; tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401071488)))]; tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401073600)))]; tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401075712)))]; tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; tensor module_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401077824)))]; tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; tensor module_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409466496)))]; tensor linear_74_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; tensor var_1536_to_fp16 = const()[name = tensor("op_1536_to_fp16"), val = tensor(0x1p-1)]; tensor var_1537_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1536_to_fp16)[name = tensor("op_1537_cast_fp16")]; tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1537_cast_fp16)[name = tensor("input_447_cast_fp16")]; tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417855168)))]; tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417857280)))]; tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; tensor module_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417859392)))]; tensor linear_75_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; tensor var_1553 = const()[name = tensor("op_1553"), val = tensor([1, -1, 8, 128])]; tensor q_49_cast_fp16 = reshape(shape = var_1553, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; tensor module_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419956608)))]; tensor linear_76_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; tensor var_1557 = const()[name = tensor("op_1557"), val = tensor([1, -1, 8, 128])]; tensor k_33_cast_fp16 = reshape(shape = var_1557, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; tensor module_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422053824)))]; tensor linear_77_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; tensor var_1561 = const()[name = tensor("op_1561"), val = tensor([1, -1, 8, 128])]; tensor v_17_cast_fp16 = reshape(shape = var_1561, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424151040)))]; tensor var_1573_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1573_cast_fp16")]; tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424153152)))]; tensor var_1575_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1575_cast_fp16")]; tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; tensor var_1577_to_fp16 = const()[name = tensor("op_1577_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424155264)))]; tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1575_cast_fp16)[name = tensor("transpose_231")]; tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1577_to_fp16)[name = tensor("x_183_cast_fp16")]; tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; tensor const_94_to_fp16 = const()[name = tensor("const_94_to_fp16"), val = tensor(0x0p+0)]; tensor x_185_cast_fp16 = pad(constant_val = const_94_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; tensor var_1585 = const()[name = tensor("op_1585"), val = tensor([1, 8, -1, 188])]; tensor x_187_cast_fp16 = reshape(shape = var_1585, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; tensor var_1589_begin_0 = const()[name = tensor("op_1589_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1589_end_0 = const()[name = tensor("op_1589_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1589_end_mask_0 = const()[name = tensor("op_1589_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1589_cast_fp16 = slice_by_index(begin = var_1589_begin_0, end = var_1589_end_0, end_mask = var_1589_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1589_cast_fp16")]; tensor var_1590 = const()[name = tensor("op_1590"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1590, x = var_1589_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; tensor transpose_88_perm_0 = const()[name = tensor("transpose_88_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_89_perm_0 = const()[name = tensor("transpose_89_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_89 = transpose(perm = transpose_89_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_229")]; tensor transpose_88 = transpose(perm = transpose_88_perm_0, x = var_1573_cast_fp16)[name = tensor("transpose_230")]; tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_88, y = transpose_89)[name = tensor("matrix_ac_17_cast_fp16")]; tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; tensor var_1599_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1599_cast_fp16")]; tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_33_cast_fp16 = mul(x = var_1599_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_3)[name = tensor("scores_35_cast_fp16")]; tensor var_1605_cast_fp16 = softmax(axis = var_30, x = scores_35_cast_fp16)[name = tensor("op_1605_cast_fp16")]; tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1605_cast_fp16, cond = mask_3)[name = tensor("input_449_cast_fp16")]; tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_232")]; tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; tensor var_1609_perm_0 = const()[name = tensor("op_1609_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1610 = const()[name = tensor("op_1610"), val = tensor([1, -1, 1024])]; tensor var_1609_cast_fp16 = transpose(perm = var_1609_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_228")]; tensor input_451_cast_fp16 = reshape(shape = var_1610, x = var_1609_cast_fp16)[name = tensor("input_451_cast_fp16")]; tensor module_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424923328)))]; tensor linear_79_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427020544)))]; tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427022656)))]; tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427024768)))]; tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_227")]; tensor input_459_cast_fp16 = conv(dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_328)[name = tensor("input_461_cast_fp16")]; tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; tensor const_97_to_fp16 = const()[name = tensor("const_97_to_fp16"), val = tensor(0x0p+0)]; tensor input_463_cast_fp16 = pad(constant_val = const_97_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(1024)]; tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; tensor const_264_to_fp16 = const()[name = tensor("const_264_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(431219136)))]; tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(431237632)))]; tensor input_467_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_264_to_fp16, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(431239744)))]; tensor x_197_cast_fp16 = conv(dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_226")]; tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433336960)))]; tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433339072)))]; tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; tensor module_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433341184)))]; tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; tensor module_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441729856)))]; tensor linear_81_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; tensor var_1670_to_fp16 = const()[name = tensor("op_1670_to_fp16"), val = tensor(0x1p-1)]; tensor var_1671_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1670_to_fp16)[name = tensor("op_1671_cast_fp16")]; tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1671_cast_fp16)[name = tensor("input_485_cast_fp16")]; tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450118528)))]; tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450120640)))]; tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450122752)))]; tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450124864)))]; tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; tensor module_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450126976)))]; tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; tensor module_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(458515648)))]; tensor linear_83_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; tensor var_1699_to_fp16 = const()[name = tensor("op_1699_to_fp16"), val = tensor(0x1p-1)]; tensor var_1700_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1699_to_fp16)[name = tensor("op_1700_cast_fp16")]; tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1700_cast_fp16)[name = tensor("input_499_cast_fp16")]; tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466904320)))]; tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466906432)))]; tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; tensor module_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466908544)))]; tensor linear_84_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; tensor var_1716 = const()[name = tensor("op_1716"), val = tensor([1, -1, 8, 128])]; tensor q_55_cast_fp16 = reshape(shape = var_1716, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; tensor module_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(469005760)))]; tensor linear_85_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; tensor var_1720 = const()[name = tensor("op_1720"), val = tensor([1, -1, 8, 128])]; tensor k_37_cast_fp16 = reshape(shape = var_1720, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; tensor module_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471102976)))]; tensor linear_86_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; tensor var_1724 = const()[name = tensor("op_1724"), val = tensor([1, -1, 8, 128])]; tensor v_19_cast_fp16 = reshape(shape = var_1724, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(473200192)))]; tensor var_1736_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1736_cast_fp16")]; tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(473202304)))]; tensor var_1738_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1738_cast_fp16")]; tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; tensor var_1740_to_fp16 = const()[name = tensor("op_1740_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(473204416)))]; tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1738_cast_fp16)[name = tensor("transpose_224")]; tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_1740_to_fp16)[name = tensor("x_205_cast_fp16")]; tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; tensor const_104_to_fp16 = const()[name = tensor("const_104_to_fp16"), val = tensor(0x0p+0)]; tensor x_207_cast_fp16 = pad(constant_val = const_104_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; tensor var_1748 = const()[name = tensor("op_1748"), val = tensor([1, 8, -1, 188])]; tensor x_209_cast_fp16 = reshape(shape = var_1748, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; tensor var_1752_begin_0 = const()[name = tensor("op_1752_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1752_end_0 = const()[name = tensor("op_1752_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1752_end_mask_0 = const()[name = tensor("op_1752_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1752_cast_fp16 = slice_by_index(begin = var_1752_begin_0, end = var_1752_end_0, end_mask = var_1752_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1752_cast_fp16")]; tensor var_1753 = const()[name = tensor("op_1753"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1753, x = var_1752_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; tensor transpose_90_perm_0 = const()[name = tensor("transpose_90_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_91_perm_0 = const()[name = tensor("transpose_91_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_91 = transpose(perm = transpose_91_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_222")]; tensor transpose_90 = transpose(perm = transpose_90_perm_0, x = var_1736_cast_fp16)[name = tensor("transpose_223")]; tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_90, y = transpose_91)[name = tensor("matrix_ac_19_cast_fp16")]; tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; tensor var_1762_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1762_cast_fp16")]; tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_37_cast_fp16 = mul(x = var_1762_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_3)[name = tensor("scores_39_cast_fp16")]; tensor var_1768_cast_fp16 = softmax(axis = var_30, x = scores_39_cast_fp16)[name = tensor("op_1768_cast_fp16")]; tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1768_cast_fp16, cond = mask_3)[name = tensor("input_501_cast_fp16")]; tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_225")]; tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; tensor var_1772_perm_0 = const()[name = tensor("op_1772_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1773 = const()[name = tensor("op_1773"), val = tensor([1, -1, 1024])]; tensor var_1772_cast_fp16 = transpose(perm = var_1772_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_221")]; tensor input_503_cast_fp16 = reshape(shape = var_1773, x = var_1772_cast_fp16)[name = tensor("input_503_cast_fp16")]; tensor module_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(473972480)))]; tensor linear_88_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476069696)))]; tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476071808)))]; tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476073920)))]; tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_220")]; tensor input_511_cast_fp16 = conv(dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_328)[name = tensor("input_513_cast_fp16")]; tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; tensor const_107_to_fp16 = const()[name = tensor("const_107_to_fp16"), val = tensor(0x0p+0)]; tensor input_515_cast_fp16 = pad(constant_val = const_107_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(1024)]; tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; tensor const_266_to_fp16 = const()[name = tensor("const_266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480268288)))]; tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480286784)))]; tensor input_519_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_266_to_fp16, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480288896)))]; tensor x_219_cast_fp16 = conv(dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_219")]; tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482386112)))]; tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482388224)))]; tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; tensor module_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482390336)))]; tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; tensor module_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490779008)))]; tensor linear_90_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; tensor var_1833_to_fp16 = const()[name = tensor("op_1833_to_fp16"), val = tensor(0x1p-1)]; tensor var_1834_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1833_to_fp16)[name = tensor("op_1834_cast_fp16")]; tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_1834_cast_fp16)[name = tensor("input_537_cast_fp16")]; tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499167680)))]; tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499169792)))]; tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499171904)))]; tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499174016)))]; tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; tensor module_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499176128)))]; tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; tensor module_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507564800)))]; tensor linear_92_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; tensor var_1862_to_fp16 = const()[name = tensor("op_1862_to_fp16"), val = tensor(0x1p-1)]; tensor var_1863_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_1862_to_fp16)[name = tensor("op_1863_cast_fp16")]; tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_1863_cast_fp16)[name = tensor("input_551_cast_fp16")]; tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515953472)))]; tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515955584)))]; tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; tensor module_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515957696)))]; tensor linear_93_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; tensor var_1879 = const()[name = tensor("op_1879"), val = tensor([1, -1, 8, 128])]; tensor q_61_cast_fp16 = reshape(shape = var_1879, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; tensor module_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518054912)))]; tensor linear_94_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; tensor var_1883 = const()[name = tensor("op_1883"), val = tensor([1, -1, 8, 128])]; tensor k_41_cast_fp16 = reshape(shape = var_1883, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; tensor module_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520152128)))]; tensor linear_95_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; tensor var_1887 = const()[name = tensor("op_1887"), val = tensor([1, -1, 8, 128])]; tensor v_21_cast_fp16 = reshape(shape = var_1887, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522249344)))]; tensor var_1899_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1899_cast_fp16")]; tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522251456)))]; tensor var_1901_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1901_cast_fp16")]; tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; tensor var_1903_to_fp16 = const()[name = tensor("op_1903_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522253568)))]; tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_1901_cast_fp16)[name = tensor("transpose_217")]; tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_1903_to_fp16)[name = tensor("x_227_cast_fp16")]; tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; tensor const_114_to_fp16 = const()[name = tensor("const_114_to_fp16"), val = tensor(0x0p+0)]; tensor x_229_cast_fp16 = pad(constant_val = const_114_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; tensor var_1911 = const()[name = tensor("op_1911"), val = tensor([1, 8, -1, 188])]; tensor x_231_cast_fp16 = reshape(shape = var_1911, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; tensor var_1915_begin_0 = const()[name = tensor("op_1915_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1915_end_0 = const()[name = tensor("op_1915_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1915_end_mask_0 = const()[name = tensor("op_1915_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1915_cast_fp16 = slice_by_index(begin = var_1915_begin_0, end = var_1915_end_0, end_mask = var_1915_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_1915_cast_fp16")]; tensor var_1916 = const()[name = tensor("op_1916"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_41_cast_fp16 = reshape(shape = var_1916, x = var_1915_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; tensor transpose_92_perm_0 = const()[name = tensor("transpose_92_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_93_perm_0 = const()[name = tensor("transpose_93_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_93 = transpose(perm = transpose_93_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_215")]; tensor transpose_92 = transpose(perm = transpose_92_perm_0, x = var_1899_cast_fp16)[name = tensor("transpose_216")]; tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_92, y = transpose_93)[name = tensor("matrix_ac_21_cast_fp16")]; tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; tensor var_1925_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_1925_cast_fp16")]; tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_41_cast_fp16 = mul(x = var_1925_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_3)[name = tensor("scores_43_cast_fp16")]; tensor var_1931_cast_fp16 = softmax(axis = var_30, x = scores_43_cast_fp16)[name = tensor("op_1931_cast_fp16")]; tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_1931_cast_fp16, cond = mask_3)[name = tensor("input_553_cast_fp16")]; tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_218")]; tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; tensor var_1935_perm_0 = const()[name = tensor("op_1935_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 1024])]; tensor var_1935_cast_fp16 = transpose(perm = var_1935_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_214")]; tensor input_555_cast_fp16 = reshape(shape = var_1936, x = var_1935_cast_fp16)[name = tensor("input_555_cast_fp16")]; tensor module_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523021632)))]; tensor linear_97_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525118848)))]; tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525120960)))]; tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525123072)))]; tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_213")]; tensor input_563_cast_fp16 = conv(dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_328)[name = tensor("input_565_cast_fp16")]; tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; tensor const_117_to_fp16 = const()[name = tensor("const_117_to_fp16"), val = tensor(0x0p+0)]; tensor input_567_cast_fp16 = pad(constant_val = const_117_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(1024)]; tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; tensor const_268_to_fp16 = const()[name = tensor("const_268_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529317440)))]; tensor const_269_to_fp16 = const()[name = tensor("const_269_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529335936)))]; tensor input_571_cast_fp16 = conv(bias = const_269_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_268_to_fp16, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529338048)))]; tensor x_241_cast_fp16 = conv(dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_212")]; tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531435264)))]; tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531437376)))]; tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; tensor module_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531439488)))]; tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; tensor module_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539828160)))]; tensor linear_99_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; tensor var_1996_to_fp16 = const()[name = tensor("op_1996_to_fp16"), val = tensor(0x1p-1)]; tensor var_1997_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_1996_to_fp16)[name = tensor("op_1997_cast_fp16")]; tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_1997_cast_fp16)[name = tensor("input_589_cast_fp16")]; tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548216832)))]; tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548218944)))]; tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548221056)))]; tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548223168)))]; tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; tensor module_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548225280)))]; tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; tensor module_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556613952)))]; tensor linear_101_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; tensor var_2025_to_fp16 = const()[name = tensor("op_2025_to_fp16"), val = tensor(0x1p-1)]; tensor var_2026_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2025_to_fp16)[name = tensor("op_2026_cast_fp16")]; tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2026_cast_fp16)[name = tensor("input_603_cast_fp16")]; tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565002624)))]; tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565004736)))]; tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; tensor module_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565006848)))]; tensor linear_102_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; tensor var_2042 = const()[name = tensor("op_2042"), val = tensor([1, -1, 8, 128])]; tensor q_67_cast_fp16 = reshape(shape = var_2042, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; tensor module_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567104064)))]; tensor linear_103_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; tensor var_2046 = const()[name = tensor("op_2046"), val = tensor([1, -1, 8, 128])]; tensor k_45_cast_fp16 = reshape(shape = var_2046, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; tensor module_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569201280)))]; tensor linear_104_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; tensor var_2050 = const()[name = tensor("op_2050"), val = tensor([1, -1, 8, 128])]; tensor v_23_cast_fp16 = reshape(shape = var_2050, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571298496)))]; tensor var_2062_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2062_cast_fp16")]; tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571300608)))]; tensor var_2064_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2064_cast_fp16")]; tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; tensor var_2066_to_fp16 = const()[name = tensor("op_2066_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571302720)))]; tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2064_cast_fp16)[name = tensor("transpose_210")]; tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2066_to_fp16)[name = tensor("x_249_cast_fp16")]; tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; tensor const_124_to_fp16 = const()[name = tensor("const_124_to_fp16"), val = tensor(0x0p+0)]; tensor x_251_cast_fp16 = pad(constant_val = const_124_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; tensor var_2074 = const()[name = tensor("op_2074"), val = tensor([1, 8, -1, 188])]; tensor x_253_cast_fp16 = reshape(shape = var_2074, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; tensor var_2078_begin_0 = const()[name = tensor("op_2078_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2078_end_0 = const()[name = tensor("op_2078_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2078_end_mask_0 = const()[name = tensor("op_2078_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2078_cast_fp16 = slice_by_index(begin = var_2078_begin_0, end = var_2078_end_0, end_mask = var_2078_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2078_cast_fp16")]; tensor var_2079 = const()[name = tensor("op_2079"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2079, x = var_2078_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; tensor transpose_94_perm_0 = const()[name = tensor("transpose_94_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_95_perm_0 = const()[name = tensor("transpose_95_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_95 = transpose(perm = transpose_95_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_208")]; tensor transpose_94 = transpose(perm = transpose_94_perm_0, x = var_2062_cast_fp16)[name = tensor("transpose_209")]; tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_94, y = transpose_95)[name = tensor("matrix_ac_23_cast_fp16")]; tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; tensor var_2088_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2088_cast_fp16")]; tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_45_cast_fp16 = mul(x = var_2088_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_3)[name = tensor("scores_47_cast_fp16")]; tensor var_2094_cast_fp16 = softmax(axis = var_30, x = scores_47_cast_fp16)[name = tensor("op_2094_cast_fp16")]; tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2094_cast_fp16, cond = mask_3)[name = tensor("input_605_cast_fp16")]; tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_211")]; tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; tensor var_2098_perm_0 = const()[name = tensor("op_2098_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2099 = const()[name = tensor("op_2099"), val = tensor([1, -1, 1024])]; tensor var_2098_cast_fp16 = transpose(perm = var_2098_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_207")]; tensor input_607_cast_fp16 = reshape(shape = var_2099, x = var_2098_cast_fp16)[name = tensor("input_607_cast_fp16")]; tensor module_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572070784)))]; tensor linear_106_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574168000)))]; tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574170112)))]; tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574172224)))]; tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_206")]; tensor input_615_cast_fp16 = conv(dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_328)[name = tensor("input_617_cast_fp16")]; tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; tensor const_127_to_fp16 = const()[name = tensor("const_127_to_fp16"), val = tensor(0x0p+0)]; tensor input_619_cast_fp16 = pad(constant_val = const_127_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(1024)]; tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; tensor const_270_to_fp16 = const()[name = tensor("const_270_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578366592)))]; tensor const_271_to_fp16 = const()[name = tensor("const_271_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578385088)))]; tensor input_623_cast_fp16 = conv(bias = const_271_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_270_to_fp16, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578387200)))]; tensor x_263_cast_fp16 = conv(dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_205")]; tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580484416)))]; tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580486528)))]; tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; tensor module_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580488640)))]; tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; tensor module_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588877312)))]; tensor linear_108_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; tensor var_2159_to_fp16 = const()[name = tensor("op_2159_to_fp16"), val = tensor(0x1p-1)]; tensor var_2160_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2159_to_fp16)[name = tensor("op_2160_cast_fp16")]; tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2160_cast_fp16)[name = tensor("input_641_cast_fp16")]; tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597265984)))]; tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597268096)))]; tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597270208)))]; tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597272320)))]; tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; tensor module_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597274432)))]; tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; tensor module_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(605663104)))]; tensor linear_110_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; tensor var_2188_to_fp16 = const()[name = tensor("op_2188_to_fp16"), val = tensor(0x1p-1)]; tensor var_2189_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2188_to_fp16)[name = tensor("op_2189_cast_fp16")]; tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2189_cast_fp16)[name = tensor("input_655_cast_fp16")]; tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(614051776)))]; tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(614053888)))]; tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; tensor module_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(614056000)))]; tensor linear_111_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; tensor var_2205 = const()[name = tensor("op_2205"), val = tensor([1, -1, 8, 128])]; tensor q_73_cast_fp16 = reshape(shape = var_2205, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; tensor module_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(616153216)))]; tensor linear_112_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; tensor var_2209 = const()[name = tensor("op_2209"), val = tensor([1, -1, 8, 128])]; tensor k_49_cast_fp16 = reshape(shape = var_2209, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; tensor module_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(618250432)))]; tensor linear_113_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; tensor var_2213 = const()[name = tensor("op_2213"), val = tensor([1, -1, 8, 128])]; tensor v_25_cast_fp16 = reshape(shape = var_2213, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(620347648)))]; tensor var_2225_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2225_cast_fp16")]; tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(620349760)))]; tensor var_2227_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2227_cast_fp16")]; tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; tensor var_2229_to_fp16 = const()[name = tensor("op_2229_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(620351872)))]; tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2227_cast_fp16)[name = tensor("transpose_203")]; tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2229_to_fp16)[name = tensor("x_271_cast_fp16")]; tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; tensor const_134_to_fp16 = const()[name = tensor("const_134_to_fp16"), val = tensor(0x0p+0)]; tensor x_273_cast_fp16 = pad(constant_val = const_134_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; tensor var_2237 = const()[name = tensor("op_2237"), val = tensor([1, 8, -1, 188])]; tensor x_275_cast_fp16 = reshape(shape = var_2237, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; tensor var_2241_begin_0 = const()[name = tensor("op_2241_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2241_end_0 = const()[name = tensor("op_2241_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2241_end_mask_0 = const()[name = tensor("op_2241_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2241_cast_fp16 = slice_by_index(begin = var_2241_begin_0, end = var_2241_end_0, end_mask = var_2241_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2241_cast_fp16")]; tensor var_2242 = const()[name = tensor("op_2242"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2242, x = var_2241_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; tensor transpose_96_perm_0 = const()[name = tensor("transpose_96_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_97_perm_0 = const()[name = tensor("transpose_97_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_97 = transpose(perm = transpose_97_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_201")]; tensor transpose_96 = transpose(perm = transpose_96_perm_0, x = var_2225_cast_fp16)[name = tensor("transpose_202")]; tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_96, y = transpose_97)[name = tensor("matrix_ac_25_cast_fp16")]; tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; tensor var_2251_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2251_cast_fp16")]; tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_49_cast_fp16 = mul(x = var_2251_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_3)[name = tensor("scores_51_cast_fp16")]; tensor var_2257_cast_fp16 = softmax(axis = var_30, x = scores_51_cast_fp16)[name = tensor("op_2257_cast_fp16")]; tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2257_cast_fp16, cond = mask_3)[name = tensor("input_657_cast_fp16")]; tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_204")]; tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; tensor var_2261_perm_0 = const()[name = tensor("op_2261_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2262 = const()[name = tensor("op_2262"), val = tensor([1, -1, 1024])]; tensor var_2261_cast_fp16 = transpose(perm = var_2261_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_200")]; tensor input_659_cast_fp16 = reshape(shape = var_2262, x = var_2261_cast_fp16)[name = tensor("input_659_cast_fp16")]; tensor module_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(621119936)))]; tensor linear_115_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(623217152)))]; tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(623219264)))]; tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(623221376)))]; tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_199")]; tensor input_667_cast_fp16 = conv(dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_328)[name = tensor("input_669_cast_fp16")]; tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; tensor const_137_to_fp16 = const()[name = tensor("const_137_to_fp16"), val = tensor(0x0p+0)]; tensor input_671_cast_fp16 = pad(constant_val = const_137_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(1024)]; tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; tensor const_272_to_fp16 = const()[name = tensor("const_272_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(627415744)))]; tensor const_273_to_fp16 = const()[name = tensor("const_273_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(627434240)))]; tensor input_675_cast_fp16 = conv(bias = const_273_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_272_to_fp16, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(627436352)))]; tensor x_285_cast_fp16 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_198")]; tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(629533568)))]; tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(629535680)))]; tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; tensor module_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(629537792)))]; tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; tensor module_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(637926464)))]; tensor linear_117_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; tensor var_2322_to_fp16 = const()[name = tensor("op_2322_to_fp16"), val = tensor(0x1p-1)]; tensor var_2323_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2322_to_fp16)[name = tensor("op_2323_cast_fp16")]; tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2323_cast_fp16)[name = tensor("input_693_cast_fp16")]; tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(646315136)))]; tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(646317248)))]; tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(646319360)))]; tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(646321472)))]; tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; tensor module_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(646323584)))]; tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; tensor module_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(654712256)))]; tensor linear_119_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; tensor var_2351_to_fp16 = const()[name = tensor("op_2351_to_fp16"), val = tensor(0x1p-1)]; tensor var_2352_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2351_to_fp16)[name = tensor("op_2352_cast_fp16")]; tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2352_cast_fp16)[name = tensor("input_707_cast_fp16")]; tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(663100928)))]; tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(663103040)))]; tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; tensor module_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(663105152)))]; tensor linear_120_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; tensor var_2368 = const()[name = tensor("op_2368"), val = tensor([1, -1, 8, 128])]; tensor q_79_cast_fp16 = reshape(shape = var_2368, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; tensor module_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(665202368)))]; tensor linear_121_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; tensor var_2372 = const()[name = tensor("op_2372"), val = tensor([1, -1, 8, 128])]; tensor k_53_cast_fp16 = reshape(shape = var_2372, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; tensor module_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(667299584)))]; tensor linear_122_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; tensor var_2376 = const()[name = tensor("op_2376"), val = tensor([1, -1, 8, 128])]; tensor v_27_cast_fp16 = reshape(shape = var_2376, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(669396800)))]; tensor var_2388_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2388_cast_fp16")]; tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(669398912)))]; tensor var_2390_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2390_cast_fp16")]; tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; tensor var_2392_to_fp16 = const()[name = tensor("op_2392_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(669401024)))]; tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2390_cast_fp16)[name = tensor("transpose_196")]; tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2392_to_fp16)[name = tensor("x_293_cast_fp16")]; tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; tensor const_144_to_fp16 = const()[name = tensor("const_144_to_fp16"), val = tensor(0x0p+0)]; tensor x_295_cast_fp16 = pad(constant_val = const_144_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; tensor var_2400 = const()[name = tensor("op_2400"), val = tensor([1, 8, -1, 188])]; tensor x_297_cast_fp16 = reshape(shape = var_2400, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; tensor var_2404_begin_0 = const()[name = tensor("op_2404_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2404_end_0 = const()[name = tensor("op_2404_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2404_end_mask_0 = const()[name = tensor("op_2404_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2404_cast_fp16 = slice_by_index(begin = var_2404_begin_0, end = var_2404_end_0, end_mask = var_2404_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2404_cast_fp16")]; tensor var_2405 = const()[name = tensor("op_2405"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2405, x = var_2404_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; tensor transpose_98_perm_0 = const()[name = tensor("transpose_98_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_99_perm_0 = const()[name = tensor("transpose_99_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_99 = transpose(perm = transpose_99_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_194")]; tensor transpose_98 = transpose(perm = transpose_98_perm_0, x = var_2388_cast_fp16)[name = tensor("transpose_195")]; tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_98, y = transpose_99)[name = tensor("matrix_ac_27_cast_fp16")]; tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; tensor var_2414_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2414_cast_fp16")]; tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_53_cast_fp16 = mul(x = var_2414_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_3)[name = tensor("scores_55_cast_fp16")]; tensor var_2420_cast_fp16 = softmax(axis = var_30, x = scores_55_cast_fp16)[name = tensor("op_2420_cast_fp16")]; tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2420_cast_fp16, cond = mask_3)[name = tensor("input_709_cast_fp16")]; tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_197")]; tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; tensor var_2424_perm_0 = const()[name = tensor("op_2424_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2425 = const()[name = tensor("op_2425"), val = tensor([1, -1, 1024])]; tensor var_2424_cast_fp16 = transpose(perm = var_2424_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_193")]; tensor input_711_cast_fp16 = reshape(shape = var_2425, x = var_2424_cast_fp16)[name = tensor("input_711_cast_fp16")]; tensor module_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(670169088)))]; tensor linear_124_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(672266304)))]; tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(672268416)))]; tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(672270528)))]; tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_192")]; tensor input_719_cast_fp16 = conv(dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_328)[name = tensor("input_721_cast_fp16")]; tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; tensor const_147_to_fp16 = const()[name = tensor("const_147_to_fp16"), val = tensor(0x0p+0)]; tensor input_723_cast_fp16 = pad(constant_val = const_147_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(1024)]; tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; tensor const_274_to_fp16 = const()[name = tensor("const_274_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(676464896)))]; tensor const_275_to_fp16 = const()[name = tensor("const_275_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(676483392)))]; tensor input_727_cast_fp16 = conv(bias = const_275_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_274_to_fp16, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(676485504)))]; tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_191")]; tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(678582720)))]; tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(678584832)))]; tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; tensor module_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(678586944)))]; tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; tensor module_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(686975616)))]; tensor linear_126_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; tensor var_2485_to_fp16 = const()[name = tensor("op_2485_to_fp16"), val = tensor(0x1p-1)]; tensor var_2486_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2485_to_fp16)[name = tensor("op_2486_cast_fp16")]; tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2486_cast_fp16)[name = tensor("input_745_cast_fp16")]; tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(695364288)))]; tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(695366400)))]; tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(695368512)))]; tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(695370624)))]; tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; tensor module_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(695372736)))]; tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; tensor module_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(703761408)))]; tensor linear_128_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; tensor var_2514_to_fp16 = const()[name = tensor("op_2514_to_fp16"), val = tensor(0x1p-1)]; tensor var_2515_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2514_to_fp16)[name = tensor("op_2515_cast_fp16")]; tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2515_cast_fp16)[name = tensor("input_759_cast_fp16")]; tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(712150080)))]; tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(712152192)))]; tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; tensor module_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(712154304)))]; tensor linear_129_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; tensor var_2531 = const()[name = tensor("op_2531"), val = tensor([1, -1, 8, 128])]; tensor q_85_cast_fp16 = reshape(shape = var_2531, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; tensor module_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(714251520)))]; tensor linear_130_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; tensor var_2535 = const()[name = tensor("op_2535"), val = tensor([1, -1, 8, 128])]; tensor k_57_cast_fp16 = reshape(shape = var_2535, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; tensor module_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(716348736)))]; tensor linear_131_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; tensor var_2539 = const()[name = tensor("op_2539"), val = tensor([1, -1, 8, 128])]; tensor v_29_cast_fp16 = reshape(shape = var_2539, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(718445952)))]; tensor var_2551_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2551_cast_fp16")]; tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(718448064)))]; tensor var_2553_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2553_cast_fp16")]; tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; tensor var_2555_to_fp16 = const()[name = tensor("op_2555_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(718450176)))]; tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2553_cast_fp16)[name = tensor("transpose_189")]; tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_2555_to_fp16)[name = tensor("x_315_cast_fp16")]; tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; tensor const_154_to_fp16 = const()[name = tensor("const_154_to_fp16"), val = tensor(0x0p+0)]; tensor x_317_cast_fp16 = pad(constant_val = const_154_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; tensor var_2563 = const()[name = tensor("op_2563"), val = tensor([1, 8, -1, 188])]; tensor x_319_cast_fp16 = reshape(shape = var_2563, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; tensor var_2567_begin_0 = const()[name = tensor("op_2567_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2567_end_0 = const()[name = tensor("op_2567_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2567_end_mask_0 = const()[name = tensor("op_2567_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2567_cast_fp16 = slice_by_index(begin = var_2567_begin_0, end = var_2567_end_0, end_mask = var_2567_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2567_cast_fp16")]; tensor var_2568 = const()[name = tensor("op_2568"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2568, x = var_2567_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; tensor transpose_100_perm_0 = const()[name = tensor("transpose_100_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_101_perm_0 = const()[name = tensor("transpose_101_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_101 = transpose(perm = transpose_101_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_187")]; tensor transpose_100 = transpose(perm = transpose_100_perm_0, x = var_2551_cast_fp16)[name = tensor("transpose_188")]; tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_100, y = transpose_101)[name = tensor("matrix_ac_29_cast_fp16")]; tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; tensor var_2577_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2577_cast_fp16")]; tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_57_cast_fp16 = mul(x = var_2577_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_3)[name = tensor("scores_59_cast_fp16")]; tensor var_2583_cast_fp16 = softmax(axis = var_30, x = scores_59_cast_fp16)[name = tensor("op_2583_cast_fp16")]; tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2583_cast_fp16, cond = mask_3)[name = tensor("input_761_cast_fp16")]; tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_190")]; tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; tensor var_2587_perm_0 = const()[name = tensor("op_2587_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2588 = const()[name = tensor("op_2588"), val = tensor([1, -1, 1024])]; tensor var_2587_cast_fp16 = transpose(perm = var_2587_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_186")]; tensor input_763_cast_fp16 = reshape(shape = var_2588, x = var_2587_cast_fp16)[name = tensor("input_763_cast_fp16")]; tensor module_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(719218240)))]; tensor linear_133_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(721315456)))]; tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(721317568)))]; tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(721319680)))]; tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_185")]; tensor input_771_cast_fp16 = conv(dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_328)[name = tensor("input_773_cast_fp16")]; tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; tensor const_157_to_fp16 = const()[name = tensor("const_157_to_fp16"), val = tensor(0x0p+0)]; tensor input_775_cast_fp16 = pad(constant_val = const_157_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(1024)]; tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; tensor const_276_to_fp16 = const()[name = tensor("const_276_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(725514048)))]; tensor const_277_to_fp16 = const()[name = tensor("const_277_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(725532544)))]; tensor input_779_cast_fp16 = conv(bias = const_277_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_276_to_fp16, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(725534656)))]; tensor x_329_cast_fp16 = conv(dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_184")]; tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(727631872)))]; tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(727633984)))]; tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; tensor module_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(727636096)))]; tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; tensor module_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(736024768)))]; tensor linear_135_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; tensor var_2648_to_fp16 = const()[name = tensor("op_2648_to_fp16"), val = tensor(0x1p-1)]; tensor var_2649_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2648_to_fp16)[name = tensor("op_2649_cast_fp16")]; tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2649_cast_fp16)[name = tensor("input_797_cast_fp16")]; tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(744413440)))]; tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(744415552)))]; tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(744417664)))]; tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(744419776)))]; tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; tensor module_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(744421888)))]; tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; tensor module_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(752810560)))]; tensor linear_137_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; tensor var_2677_to_fp16 = const()[name = tensor("op_2677_to_fp16"), val = tensor(0x1p-1)]; tensor var_2678_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2677_to_fp16)[name = tensor("op_2678_cast_fp16")]; tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2678_cast_fp16)[name = tensor("input_811_cast_fp16")]; tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(761199232)))]; tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(761201344)))]; tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; tensor module_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(761203456)))]; tensor linear_138_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; tensor var_2694 = const()[name = tensor("op_2694"), val = tensor([1, -1, 8, 128])]; tensor q_91_cast_fp16 = reshape(shape = var_2694, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; tensor module_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(763300672)))]; tensor linear_139_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; tensor var_2698 = const()[name = tensor("op_2698"), val = tensor([1, -1, 8, 128])]; tensor k_61_cast_fp16 = reshape(shape = var_2698, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; tensor module_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(765397888)))]; tensor linear_140_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; tensor var_2702 = const()[name = tensor("op_2702"), val = tensor([1, -1, 8, 128])]; tensor v_31_cast_fp16 = reshape(shape = var_2702, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(767495104)))]; tensor var_2714_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2714_cast_fp16")]; tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(767497216)))]; tensor var_2716_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2716_cast_fp16")]; tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; tensor var_2718_to_fp16 = const()[name = tensor("op_2718_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(767499328)))]; tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_2716_cast_fp16)[name = tensor("transpose_182")]; tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_2718_to_fp16)[name = tensor("x_337_cast_fp16")]; tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; tensor const_164_to_fp16 = const()[name = tensor("const_164_to_fp16"), val = tensor(0x0p+0)]; tensor x_339_cast_fp16 = pad(constant_val = const_164_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; tensor var_2726 = const()[name = tensor("op_2726"), val = tensor([1, 8, -1, 188])]; tensor x_341_cast_fp16 = reshape(shape = var_2726, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; tensor var_2730_begin_0 = const()[name = tensor("op_2730_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2730_end_0 = const()[name = tensor("op_2730_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2730_end_mask_0 = const()[name = tensor("op_2730_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2730_cast_fp16 = slice_by_index(begin = var_2730_begin_0, end = var_2730_end_0, end_mask = var_2730_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_2730_cast_fp16")]; tensor var_2731 = const()[name = tensor("op_2731"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_61_cast_fp16 = reshape(shape = var_2731, x = var_2730_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; tensor transpose_102_perm_0 = const()[name = tensor("transpose_102_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_103_perm_0 = const()[name = tensor("transpose_103_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_103 = transpose(perm = transpose_103_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_180")]; tensor transpose_102 = transpose(perm = transpose_102_perm_0, x = var_2714_cast_fp16)[name = tensor("transpose_181")]; tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_102, y = transpose_103)[name = tensor("matrix_ac_31_cast_fp16")]; tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; tensor var_2740_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_2740_cast_fp16")]; tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_61_cast_fp16 = mul(x = var_2740_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_3)[name = tensor("scores_63_cast_fp16")]; tensor var_2746_cast_fp16 = softmax(axis = var_30, x = scores_63_cast_fp16)[name = tensor("op_2746_cast_fp16")]; tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_2746_cast_fp16, cond = mask_3)[name = tensor("input_813_cast_fp16")]; tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_183")]; tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; tensor var_2750_perm_0 = const()[name = tensor("op_2750_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2751 = const()[name = tensor("op_2751"), val = tensor([1, -1, 1024])]; tensor var_2750_cast_fp16 = transpose(perm = var_2750_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_179")]; tensor input_815_cast_fp16 = reshape(shape = var_2751, x = var_2750_cast_fp16)[name = tensor("input_815_cast_fp16")]; tensor module_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(768267392)))]; tensor linear_142_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(770364608)))]; tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(770366720)))]; tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(770368832)))]; tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_178")]; tensor input_823_cast_fp16 = conv(dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_328)[name = tensor("input_825_cast_fp16")]; tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; tensor const_167_to_fp16 = const()[name = tensor("const_167_to_fp16"), val = tensor(0x0p+0)]; tensor input_827_cast_fp16 = pad(constant_val = const_167_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(1024)]; tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; tensor const_278_to_fp16 = const()[name = tensor("const_278_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(774563200)))]; tensor const_279_to_fp16 = const()[name = tensor("const_279_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(774581696)))]; tensor input_831_cast_fp16 = conv(bias = const_279_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_278_to_fp16, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(774583808)))]; tensor x_351_cast_fp16 = conv(dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_177")]; tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(776681024)))]; tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(776683136)))]; tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; tensor module_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(776685248)))]; tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; tensor module_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(785073920)))]; tensor linear_144_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; tensor var_2811_to_fp16 = const()[name = tensor("op_2811_to_fp16"), val = tensor(0x1p-1)]; tensor var_2812_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_2811_to_fp16)[name = tensor("op_2812_cast_fp16")]; tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_2812_cast_fp16)[name = tensor("input_849_cast_fp16")]; tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(793462592)))]; tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(793464704)))]; tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(793466816)))]; tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(793468928)))]; tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; tensor module_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(793471040)))]; tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; tensor module_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(801859712)))]; tensor linear_146_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; tensor var_2840_to_fp16 = const()[name = tensor("op_2840_to_fp16"), val = tensor(0x1p-1)]; tensor var_2841_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_2840_to_fp16)[name = tensor("op_2841_cast_fp16")]; tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_2841_cast_fp16)[name = tensor("input_863_cast_fp16")]; tensor query_33_axes_0 = const()[name = tensor("query_33_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(810248384)))]; tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(810250496)))]; tensor query_33_cast_fp16 = layer_norm(axes = query_33_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_33_cast_fp16")]; tensor module_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(810252608)))]; tensor linear_147_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16, x = query_33_cast_fp16)[name = tensor("linear_147_cast_fp16")]; tensor var_2857 = const()[name = tensor("op_2857"), val = tensor([1, -1, 8, 128])]; tensor q_97_cast_fp16 = reshape(shape = var_2857, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; tensor module_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(812349824)))]; tensor linear_148_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16, x = query_33_cast_fp16)[name = tensor("linear_148_cast_fp16")]; tensor var_2861 = const()[name = tensor("op_2861"), val = tensor([1, -1, 8, 128])]; tensor k_65_cast_fp16 = reshape(shape = var_2861, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; tensor module_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(814447040)))]; tensor linear_149_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16, x = query_33_cast_fp16)[name = tensor("linear_149_cast_fp16")]; tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 8, 128])]; tensor v_33_cast_fp16 = reshape(shape = var_2865, x = linear_149_cast_fp16)[name = tensor("v_33_cast_fp16")]; tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(816544256)))]; tensor var_2877_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2877_cast_fp16")]; tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(816546368)))]; tensor var_2879_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2879_cast_fp16")]; tensor q_with_bias_v_33_perm_0 = const()[name = tensor("q_with_bias_v_33_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; tensor var_2881_to_fp16 = const()[name = tensor("op_2881_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(816548480)))]; tensor q_with_bias_v_33_cast_fp16 = transpose(perm = q_with_bias_v_33_perm_0, x = var_2879_cast_fp16)[name = tensor("transpose_175")]; tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_33_cast_fp16, y = var_2881_to_fp16)[name = tensor("x_359_cast_fp16")]; tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; tensor const_174_to_fp16 = const()[name = tensor("const_174_to_fp16"), val = tensor(0x0p+0)]; tensor x_361_cast_fp16 = pad(constant_val = const_174_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; tensor var_2889 = const()[name = tensor("op_2889"), val = tensor([1, 8, -1, 188])]; tensor x_363_cast_fp16 = reshape(shape = var_2889, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; tensor var_2893_begin_0 = const()[name = tensor("op_2893_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2893_end_0 = const()[name = tensor("op_2893_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2893_end_mask_0 = const()[name = tensor("op_2893_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2893_cast_fp16 = slice_by_index(begin = var_2893_begin_0, end = var_2893_end_0, end_mask = var_2893_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_2893_cast_fp16")]; tensor var_2894 = const()[name = tensor("op_2894"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_65_cast_fp16 = reshape(shape = var_2894, x = var_2893_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; tensor matrix_ac_33_transpose_x_0 = const()[name = tensor("matrix_ac_33_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_33_transpose_y_0 = const()[name = tensor("matrix_ac_33_transpose_y_0"), val = tensor(false)]; tensor transpose_104_perm_0 = const()[name = tensor("transpose_104_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_105_perm_0 = const()[name = tensor("transpose_105_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_105 = transpose(perm = transpose_105_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_173")]; tensor transpose_104 = transpose(perm = transpose_104_perm_0, x = var_2877_cast_fp16)[name = tensor("transpose_174")]; tensor matrix_ac_33_cast_fp16 = matmul(transpose_x = matrix_ac_33_transpose_x_0, transpose_y = matrix_ac_33_transpose_y_0, x = transpose_104, y = transpose_105)[name = tensor("matrix_ac_33_cast_fp16")]; tensor matrix_bd_67_begin_0 = const()[name = tensor("matrix_bd_67_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_67_end_0 = const()[name = tensor("matrix_bd_67_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_67_end_mask_0 = const()[name = tensor("matrix_bd_67_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_67_cast_fp16 = slice_by_index(begin = matrix_bd_67_begin_0, end = matrix_bd_67_end_0, end_mask = matrix_bd_67_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_67_cast_fp16")]; tensor var_2903_cast_fp16 = add(x = matrix_ac_33_cast_fp16, y = matrix_bd_67_cast_fp16)[name = tensor("op_2903_cast_fp16")]; tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_65_cast_fp16 = mul(x = var_2903_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; tensor scores_67_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_3)[name = tensor("scores_67_cast_fp16")]; tensor var_2909_cast_fp16 = softmax(axis = var_30, x = scores_67_cast_fp16)[name = tensor("op_2909_cast_fp16")]; tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_2909_cast_fp16, cond = mask_3)[name = tensor("input_865_cast_fp16")]; tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; tensor value_35_cast_fp16 = transpose(perm = value_35_perm_0, x = v_33_cast_fp16)[name = tensor("transpose_176")]; tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_35_cast_fp16)[name = tensor("x_365_cast_fp16")]; tensor var_2913_perm_0 = const()[name = tensor("op_2913_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2914 = const()[name = tensor("op_2914"), val = tensor([1, -1, 1024])]; tensor var_2913_cast_fp16 = transpose(perm = var_2913_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_172")]; tensor input_867_cast_fp16 = reshape(shape = var_2914, x = var_2913_cast_fp16)[name = tensor("input_867_cast_fp16")]; tensor module_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(817316544)))]; tensor linear_151_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(819413760)))]; tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(819415872)))]; tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(819417984)))]; tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_171")]; tensor input_875_cast_fp16 = conv(dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_328)[name = tensor("input_877_cast_fp16")]; tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; tensor const_177_to_fp16 = const()[name = tensor("const_177_to_fp16"), val = tensor(0x0p+0)]; tensor input_879_cast_fp16 = pad(constant_val = const_177_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(1024)]; tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; tensor const_280_to_fp16 = const()[name = tensor("const_280_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(823612352)))]; tensor const_281_to_fp16 = const()[name = tensor("const_281_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(823630848)))]; tensor input_883_cast_fp16 = conv(bias = const_281_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_280_to_fp16, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(823632960)))]; tensor x_373_cast_fp16 = conv(dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_170")]; tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(825730176)))]; tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(825732288)))]; tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; tensor module_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(825734400)))]; tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; tensor module_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(834123072)))]; tensor linear_153_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; tensor var_2974_to_fp16 = const()[name = tensor("op_2974_to_fp16"), val = tensor(0x1p-1)]; tensor var_2975_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_2974_to_fp16)[name = tensor("op_2975_cast_fp16")]; tensor input_901_cast_fp16 = add(x = input_889_cast_fp16, y = var_2975_cast_fp16)[name = tensor("input_901_cast_fp16")]; tensor input_903_axes_0 = const()[name = tensor("input_903_axes_0"), val = tensor([-1])]; tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(842511744)))]; tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(842513856)))]; tensor input_903_cast_fp16 = layer_norm(axes = input_903_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("input_903_cast_fp16")]; tensor input_905_axes_0 = const()[name = tensor("input_905_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(842515968)))]; tensor module_layers_17_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(842518080)))]; tensor input_905_cast_fp16 = layer_norm(axes = input_905_axes_0, beta = module_layers_17_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward1_weight_to_fp16, x = input_903_cast_fp16)[name = tensor("input_905_cast_fp16")]; tensor module_layers_17_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_17_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(842520192)))]; tensor linear_154_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_17_feed_forward1_linear1_weight_to_fp16, x = input_905_cast_fp16)[name = tensor("linear_154_cast_fp16")]; tensor input_909_cast_fp16 = silu(x = linear_154_cast_fp16)[name = tensor("input_909_cast_fp16")]; tensor module_layers_17_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_17_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(850908864)))]; tensor linear_155_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_feed_forward1_linear2_weight_to_fp16, x = input_909_cast_fp16)[name = tensor("linear_155_cast_fp16")]; tensor var_3003_to_fp16 = const()[name = tensor("op_3003_to_fp16"), val = tensor(0x1p-1)]; tensor var_3004_cast_fp16 = mul(x = linear_155_cast_fp16, y = var_3003_to_fp16)[name = tensor("op_3004_cast_fp16")]; tensor input_915_cast_fp16 = add(x = input_903_cast_fp16, y = var_3004_cast_fp16)[name = tensor("input_915_cast_fp16")]; tensor query_35_axes_0 = const()[name = tensor("query_35_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(859297536)))]; tensor module_layers_17_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(859299648)))]; tensor query_35_cast_fp16 = layer_norm(axes = query_35_axes_0, beta = module_layers_17_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_self_att_weight_to_fp16, x = input_915_cast_fp16)[name = tensor("query_35_cast_fp16")]; tensor module_layers_17_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(859301760)))]; tensor linear_156_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_q_weight_to_fp16, x = query_35_cast_fp16)[name = tensor("linear_156_cast_fp16")]; tensor var_3020 = const()[name = tensor("op_3020"), val = tensor([1, -1, 8, 128])]; tensor q_103_cast_fp16 = reshape(shape = var_3020, x = linear_156_cast_fp16)[name = tensor("q_103_cast_fp16")]; tensor module_layers_17_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(861398976)))]; tensor linear_157_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_k_weight_to_fp16, x = query_35_cast_fp16)[name = tensor("linear_157_cast_fp16")]; tensor var_3024 = const()[name = tensor("op_3024"), val = tensor([1, -1, 8, 128])]; tensor k_69_cast_fp16 = reshape(shape = var_3024, x = linear_157_cast_fp16)[name = tensor("k_69_cast_fp16")]; tensor module_layers_17_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(863496192)))]; tensor linear_158_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_v_weight_to_fp16, x = query_35_cast_fp16)[name = tensor("linear_158_cast_fp16")]; tensor var_3028 = const()[name = tensor("op_3028"), val = tensor([1, -1, 8, 128])]; tensor v_35_cast_fp16 = reshape(shape = var_3028, x = linear_158_cast_fp16)[name = tensor("v_35_cast_fp16")]; tensor value_37_perm_0 = const()[name = tensor("value_37_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_17_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(865593408)))]; tensor var_3040_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3040_cast_fp16")]; tensor module_layers_17_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(865595520)))]; tensor var_3042_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3042_cast_fp16")]; tensor q_with_bias_v_35_perm_0 = const()[name = tensor("q_with_bias_v_35_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_381_transpose_x_0 = const()[name = tensor("x_381_transpose_x_0"), val = tensor(false)]; tensor x_381_transpose_y_0 = const()[name = tensor("x_381_transpose_y_0"), val = tensor(false)]; tensor var_3044_to_fp16 = const()[name = tensor("op_3044_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(865597632)))]; tensor q_with_bias_v_35_cast_fp16 = transpose(perm = q_with_bias_v_35_perm_0, x = var_3042_cast_fp16)[name = tensor("transpose_168")]; tensor x_381_cast_fp16 = matmul(transpose_x = x_381_transpose_x_0, transpose_y = x_381_transpose_y_0, x = q_with_bias_v_35_cast_fp16, y = var_3044_to_fp16)[name = tensor("x_381_cast_fp16")]; tensor x_383_pad_0 = const()[name = tensor("x_383_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_383_mode_0 = const()[name = tensor("x_383_mode_0"), val = tensor("constant")]; tensor const_184_to_fp16 = const()[name = tensor("const_184_to_fp16"), val = tensor(0x0p+0)]; tensor x_383_cast_fp16 = pad(constant_val = const_184_to_fp16, mode = x_383_mode_0, pad = x_383_pad_0, x = x_381_cast_fp16)[name = tensor("x_383_cast_fp16")]; tensor var_3052 = const()[name = tensor("op_3052"), val = tensor([1, 8, -1, 188])]; tensor x_385_cast_fp16 = reshape(shape = var_3052, x = x_383_cast_fp16)[name = tensor("x_385_cast_fp16")]; tensor var_3056_begin_0 = const()[name = tensor("op_3056_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3056_end_0 = const()[name = tensor("op_3056_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3056_end_mask_0 = const()[name = tensor("op_3056_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3056_cast_fp16 = slice_by_index(begin = var_3056_begin_0, end = var_3056_end_0, end_mask = var_3056_end_mask_0, x = x_385_cast_fp16)[name = tensor("op_3056_cast_fp16")]; tensor var_3057 = const()[name = tensor("op_3057"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_69_cast_fp16 = reshape(shape = var_3057, x = var_3056_cast_fp16)[name = tensor("matrix_bd_69_cast_fp16")]; tensor matrix_ac_35_transpose_x_0 = const()[name = tensor("matrix_ac_35_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_35_transpose_y_0 = const()[name = tensor("matrix_ac_35_transpose_y_0"), val = tensor(false)]; tensor transpose_106_perm_0 = const()[name = tensor("transpose_106_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_107_perm_0 = const()[name = tensor("transpose_107_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_107 = transpose(perm = transpose_107_perm_0, x = k_69_cast_fp16)[name = tensor("transpose_166")]; tensor transpose_106 = transpose(perm = transpose_106_perm_0, x = var_3040_cast_fp16)[name = tensor("transpose_167")]; tensor matrix_ac_35_cast_fp16 = matmul(transpose_x = matrix_ac_35_transpose_x_0, transpose_y = matrix_ac_35_transpose_y_0, x = transpose_106, y = transpose_107)[name = tensor("matrix_ac_35_cast_fp16")]; tensor matrix_bd_71_begin_0 = const()[name = tensor("matrix_bd_71_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_71_end_0 = const()[name = tensor("matrix_bd_71_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_71_end_mask_0 = const()[name = tensor("matrix_bd_71_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_71_cast_fp16 = slice_by_index(begin = matrix_bd_71_begin_0, end = matrix_bd_71_end_0, end_mask = matrix_bd_71_end_mask_0, x = matrix_bd_69_cast_fp16)[name = tensor("matrix_bd_71_cast_fp16")]; tensor var_3066_cast_fp16 = add(x = matrix_ac_35_cast_fp16, y = matrix_bd_71_cast_fp16)[name = tensor("op_3066_cast_fp16")]; tensor _inversed_scores_69_y_0_to_fp16 = const()[name = tensor("_inversed_scores_69_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_69_cast_fp16 = mul(x = var_3066_cast_fp16, y = _inversed_scores_69_y_0_to_fp16)[name = tensor("_inversed_scores_69_cast_fp16")]; tensor scores_71_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_69_cast_fp16, cond = mask_3)[name = tensor("scores_71_cast_fp16")]; tensor var_3072_cast_fp16 = softmax(axis = var_30, x = scores_71_cast_fp16)[name = tensor("op_3072_cast_fp16")]; tensor input_917_cast_fp16 = select(a = var_11_to_fp16, b = var_3072_cast_fp16, cond = mask_3)[name = tensor("input_917_cast_fp16")]; tensor x_387_transpose_x_0 = const()[name = tensor("x_387_transpose_x_0"), val = tensor(false)]; tensor x_387_transpose_y_0 = const()[name = tensor("x_387_transpose_y_0"), val = tensor(false)]; tensor value_37_cast_fp16 = transpose(perm = value_37_perm_0, x = v_35_cast_fp16)[name = tensor("transpose_169")]; tensor x_387_cast_fp16 = matmul(transpose_x = x_387_transpose_x_0, transpose_y = x_387_transpose_y_0, x = input_917_cast_fp16, y = value_37_cast_fp16)[name = tensor("x_387_cast_fp16")]; tensor var_3076_perm_0 = const()[name = tensor("op_3076_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3077 = const()[name = tensor("op_3077"), val = tensor([1, -1, 1024])]; tensor var_3076_cast_fp16 = transpose(perm = var_3076_perm_0, x = x_387_cast_fp16)[name = tensor("transpose_165")]; tensor input_919_cast_fp16 = reshape(shape = var_3077, x = var_3076_cast_fp16)[name = tensor("input_919_cast_fp16")]; tensor module_layers_17_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_17_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(866365696)))]; tensor linear_160_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_out_weight_to_fp16, x = input_919_cast_fp16)[name = tensor("linear_160_cast_fp16")]; tensor input_923_cast_fp16 = add(x = input_915_cast_fp16, y = linear_160_cast_fp16)[name = tensor("input_923_cast_fp16")]; tensor x_391_axes_0 = const()[name = tensor("x_391_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(868462912)))]; tensor module_layers_17_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(868465024)))]; tensor x_391_cast_fp16 = layer_norm(axes = x_391_axes_0, beta = module_layers_17_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_conv_weight_to_fp16, x = input_923_cast_fp16)[name = tensor("x_391_cast_fp16")]; tensor input_925_perm_0 = const()[name = tensor("input_925_perm_0"), val = tensor([0, 2, 1])]; tensor input_927_pad_type_0 = const()[name = tensor("input_927_pad_type_0"), val = tensor("valid")]; tensor input_927_strides_0 = const()[name = tensor("input_927_strides_0"), val = tensor([1])]; tensor input_927_pad_0 = const()[name = tensor("input_927_pad_0"), val = tensor([0, 0])]; tensor input_927_dilations_0 = const()[name = tensor("input_927_dilations_0"), val = tensor([1])]; tensor input_927_groups_0 = const()[name = tensor("input_927_groups_0"), val = tensor(1)]; tensor module_layers_17_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_17_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(868467136)))]; tensor input_925_cast_fp16 = transpose(perm = input_925_perm_0, x = x_391_cast_fp16)[name = tensor("transpose_164")]; tensor input_927_cast_fp16 = conv(dilations = input_927_dilations_0, groups = input_927_groups_0, pad = input_927_pad_0, pad_type = input_927_pad_type_0, strides = input_927_strides_0, weight = module_layers_17_conv_pointwise_conv1_weight_to_fp16, x = input_925_cast_fp16)[name = tensor("input_927_cast_fp16")]; tensor x_393_split_num_splits_0 = const()[name = tensor("x_393_split_num_splits_0"), val = tensor(2)]; tensor x_393_split_axis_0 = const()[name = tensor("x_393_split_axis_0"), val = tensor(1)]; tensor x_393_split_cast_fp16_0, tensor x_393_split_cast_fp16_1 = split(axis = x_393_split_axis_0, num_splits = x_393_split_num_splits_0, x = input_927_cast_fp16)[name = tensor("x_393_split_cast_fp16")]; tensor x_393_split_1_sigmoid_cast_fp16 = sigmoid(x = x_393_split_cast_fp16_1)[name = tensor("x_393_split_1_sigmoid_cast_fp16")]; tensor x_393_cast_fp16 = mul(x = x_393_split_cast_fp16_0, y = x_393_split_1_sigmoid_cast_fp16)[name = tensor("x_393_cast_fp16")]; tensor input_929_cast_fp16 = select(a = var_11_to_fp16, b = x_393_cast_fp16, cond = var_328)[name = tensor("input_929_cast_fp16")]; tensor input_931_pad_0 = const()[name = tensor("input_931_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_931_mode_0 = const()[name = tensor("input_931_mode_0"), val = tensor("constant")]; tensor const_187_to_fp16 = const()[name = tensor("const_187_to_fp16"), val = tensor(0x0p+0)]; tensor input_931_cast_fp16 = pad(constant_val = const_187_to_fp16, mode = input_931_mode_0, pad = input_931_pad_0, x = input_929_cast_fp16)[name = tensor("input_931_cast_fp16")]; tensor input_933_pad_type_0 = const()[name = tensor("input_933_pad_type_0"), val = tensor("valid")]; tensor input_933_groups_0 = const()[name = tensor("input_933_groups_0"), val = tensor(1024)]; tensor input_933_strides_0 = const()[name = tensor("input_933_strides_0"), val = tensor([1])]; tensor input_933_pad_0 = const()[name = tensor("input_933_pad_0"), val = tensor([0, 0])]; tensor input_933_dilations_0 = const()[name = tensor("input_933_dilations_0"), val = tensor([1])]; tensor const_282_to_fp16 = const()[name = tensor("const_282_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(872661504)))]; tensor const_283_to_fp16 = const()[name = tensor("const_283_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(872680000)))]; tensor input_935_cast_fp16 = conv(bias = const_283_to_fp16, dilations = input_933_dilations_0, groups = input_933_groups_0, pad = input_933_pad_0, pad_type = input_933_pad_type_0, strides = input_933_strides_0, weight = const_282_to_fp16, x = input_931_cast_fp16)[name = tensor("input_935_cast_fp16")]; tensor input_937_cast_fp16 = silu(x = input_935_cast_fp16)[name = tensor("input_937_cast_fp16")]; tensor x_395_pad_type_0 = const()[name = tensor("x_395_pad_type_0"), val = tensor("valid")]; tensor x_395_strides_0 = const()[name = tensor("x_395_strides_0"), val = tensor([1])]; tensor x_395_pad_0 = const()[name = tensor("x_395_pad_0"), val = tensor([0, 0])]; tensor x_395_dilations_0 = const()[name = tensor("x_395_dilations_0"), val = tensor([1])]; tensor x_395_groups_0 = const()[name = tensor("x_395_groups_0"), val = tensor(1)]; tensor module_layers_17_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_17_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(872682112)))]; tensor x_395_cast_fp16 = conv(dilations = x_395_dilations_0, groups = x_395_groups_0, pad = x_395_pad_0, pad_type = x_395_pad_type_0, strides = x_395_strides_0, weight = module_layers_17_conv_pointwise_conv2_weight_to_fp16, x = input_937_cast_fp16)[name = tensor("x_395_cast_fp16")]; tensor input_939_perm_0 = const()[name = tensor("input_939_perm_0"), val = tensor([0, 2, 1])]; tensor input_939_cast_fp16 = transpose(perm = input_939_perm_0, x = x_395_cast_fp16)[name = tensor("transpose_163")]; tensor input_941_cast_fp16 = add(x = input_923_cast_fp16, y = input_939_cast_fp16)[name = tensor("input_941_cast_fp16")]; tensor input_943_axes_0 = const()[name = tensor("input_943_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(874779328)))]; tensor module_layers_17_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(874781440)))]; tensor input_943_cast_fp16 = layer_norm(axes = input_943_axes_0, beta = module_layers_17_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward2_weight_to_fp16, x = input_941_cast_fp16)[name = tensor("input_943_cast_fp16")]; tensor module_layers_17_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_17_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(874783552)))]; tensor linear_161_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_17_feed_forward2_linear1_weight_to_fp16, x = input_943_cast_fp16)[name = tensor("linear_161_cast_fp16")]; tensor input_947_cast_fp16 = silu(x = linear_161_cast_fp16)[name = tensor("input_947_cast_fp16")]; tensor module_layers_17_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_17_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(883172224)))]; tensor linear_162_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_feed_forward2_linear2_weight_to_fp16, x = input_947_cast_fp16)[name = tensor("linear_162_cast_fp16")]; tensor var_3137_to_fp16 = const()[name = tensor("op_3137_to_fp16"), val = tensor(0x1p-1)]; tensor var_3138_cast_fp16 = mul(x = linear_162_cast_fp16, y = var_3137_to_fp16)[name = tensor("op_3138_cast_fp16")]; tensor input_953_cast_fp16 = add(x = input_941_cast_fp16, y = var_3138_cast_fp16)[name = tensor("input_953_cast_fp16")]; tensor input_955_axes_0 = const()[name = tensor("input_955_axes_0"), val = tensor([-1])]; tensor module_layers_17_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(891560896)))]; tensor module_layers_17_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(891563008)))]; tensor input_955_cast_fp16 = layer_norm(axes = input_955_axes_0, beta = module_layers_17_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_out_weight_to_fp16, x = input_953_cast_fp16)[name = tensor("input_955_cast_fp16")]; tensor input_957_axes_0 = const()[name = tensor("input_957_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(891565120)))]; tensor module_layers_18_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(891567232)))]; tensor input_957_cast_fp16 = layer_norm(axes = input_957_axes_0, beta = module_layers_18_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward1_weight_to_fp16, x = input_955_cast_fp16)[name = tensor("input_957_cast_fp16")]; tensor module_layers_18_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_18_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(891569344)))]; tensor linear_163_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_18_feed_forward1_linear1_weight_to_fp16, x = input_957_cast_fp16)[name = tensor("linear_163_cast_fp16")]; tensor input_961_cast_fp16 = silu(x = linear_163_cast_fp16)[name = tensor("input_961_cast_fp16")]; tensor module_layers_18_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_18_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(899958016)))]; tensor linear_164_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_feed_forward1_linear2_weight_to_fp16, x = input_961_cast_fp16)[name = tensor("linear_164_cast_fp16")]; tensor var_3166_to_fp16 = const()[name = tensor("op_3166_to_fp16"), val = tensor(0x1p-1)]; tensor var_3167_cast_fp16 = mul(x = linear_164_cast_fp16, y = var_3166_to_fp16)[name = tensor("op_3167_cast_fp16")]; tensor input_967_cast_fp16 = add(x = input_955_cast_fp16, y = var_3167_cast_fp16)[name = tensor("input_967_cast_fp16")]; tensor query_37_axes_0 = const()[name = tensor("query_37_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(908346688)))]; tensor module_layers_18_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(908348800)))]; tensor query_37_cast_fp16 = layer_norm(axes = query_37_axes_0, beta = module_layers_18_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_self_att_weight_to_fp16, x = input_967_cast_fp16)[name = tensor("query_37_cast_fp16")]; tensor module_layers_18_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(908350912)))]; tensor linear_165_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_q_weight_to_fp16, x = query_37_cast_fp16)[name = tensor("linear_165_cast_fp16")]; tensor var_3183 = const()[name = tensor("op_3183"), val = tensor([1, -1, 8, 128])]; tensor q_109_cast_fp16 = reshape(shape = var_3183, x = linear_165_cast_fp16)[name = tensor("q_109_cast_fp16")]; tensor module_layers_18_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(910448128)))]; tensor linear_166_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_k_weight_to_fp16, x = query_37_cast_fp16)[name = tensor("linear_166_cast_fp16")]; tensor var_3187 = const()[name = tensor("op_3187"), val = tensor([1, -1, 8, 128])]; tensor k_73_cast_fp16 = reshape(shape = var_3187, x = linear_166_cast_fp16)[name = tensor("k_73_cast_fp16")]; tensor module_layers_18_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(912545344)))]; tensor linear_167_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_v_weight_to_fp16, x = query_37_cast_fp16)[name = tensor("linear_167_cast_fp16")]; tensor var_3191 = const()[name = tensor("op_3191"), val = tensor([1, -1, 8, 128])]; tensor v_37_cast_fp16 = reshape(shape = var_3191, x = linear_167_cast_fp16)[name = tensor("v_37_cast_fp16")]; tensor value_39_perm_0 = const()[name = tensor("value_39_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_18_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(914642560)))]; tensor var_3203_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3203_cast_fp16")]; tensor module_layers_18_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(914644672)))]; tensor var_3205_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3205_cast_fp16")]; tensor q_with_bias_v_37_perm_0 = const()[name = tensor("q_with_bias_v_37_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; tensor var_3207_to_fp16 = const()[name = tensor("op_3207_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(914646784)))]; tensor q_with_bias_v_37_cast_fp16 = transpose(perm = q_with_bias_v_37_perm_0, x = var_3205_cast_fp16)[name = tensor("transpose_161")]; tensor x_403_cast_fp16 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = q_with_bias_v_37_cast_fp16, y = var_3207_to_fp16)[name = tensor("x_403_cast_fp16")]; tensor x_405_pad_0 = const()[name = tensor("x_405_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_405_mode_0 = const()[name = tensor("x_405_mode_0"), val = tensor("constant")]; tensor const_194_to_fp16 = const()[name = tensor("const_194_to_fp16"), val = tensor(0x0p+0)]; tensor x_405_cast_fp16 = pad(constant_val = const_194_to_fp16, mode = x_405_mode_0, pad = x_405_pad_0, x = x_403_cast_fp16)[name = tensor("x_405_cast_fp16")]; tensor var_3215 = const()[name = tensor("op_3215"), val = tensor([1, 8, -1, 188])]; tensor x_407_cast_fp16 = reshape(shape = var_3215, x = x_405_cast_fp16)[name = tensor("x_407_cast_fp16")]; tensor var_3219_begin_0 = const()[name = tensor("op_3219_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3219_end_0 = const()[name = tensor("op_3219_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3219_end_mask_0 = const()[name = tensor("op_3219_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3219_cast_fp16 = slice_by_index(begin = var_3219_begin_0, end = var_3219_end_0, end_mask = var_3219_end_mask_0, x = x_407_cast_fp16)[name = tensor("op_3219_cast_fp16")]; tensor var_3220 = const()[name = tensor("op_3220"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_73_cast_fp16 = reshape(shape = var_3220, x = var_3219_cast_fp16)[name = tensor("matrix_bd_73_cast_fp16")]; tensor matrix_ac_37_transpose_x_0 = const()[name = tensor("matrix_ac_37_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_37_transpose_y_0 = const()[name = tensor("matrix_ac_37_transpose_y_0"), val = tensor(false)]; tensor transpose_108_perm_0 = const()[name = tensor("transpose_108_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_109_perm_0 = const()[name = tensor("transpose_109_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_109 = transpose(perm = transpose_109_perm_0, x = k_73_cast_fp16)[name = tensor("transpose_159")]; tensor transpose_108 = transpose(perm = transpose_108_perm_0, x = var_3203_cast_fp16)[name = tensor("transpose_160")]; tensor matrix_ac_37_cast_fp16 = matmul(transpose_x = matrix_ac_37_transpose_x_0, transpose_y = matrix_ac_37_transpose_y_0, x = transpose_108, y = transpose_109)[name = tensor("matrix_ac_37_cast_fp16")]; tensor matrix_bd_75_begin_0 = const()[name = tensor("matrix_bd_75_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_75_end_0 = const()[name = tensor("matrix_bd_75_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_75_end_mask_0 = const()[name = tensor("matrix_bd_75_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_75_cast_fp16 = slice_by_index(begin = matrix_bd_75_begin_0, end = matrix_bd_75_end_0, end_mask = matrix_bd_75_end_mask_0, x = matrix_bd_73_cast_fp16)[name = tensor("matrix_bd_75_cast_fp16")]; tensor var_3229_cast_fp16 = add(x = matrix_ac_37_cast_fp16, y = matrix_bd_75_cast_fp16)[name = tensor("op_3229_cast_fp16")]; tensor _inversed_scores_73_y_0_to_fp16 = const()[name = tensor("_inversed_scores_73_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_73_cast_fp16 = mul(x = var_3229_cast_fp16, y = _inversed_scores_73_y_0_to_fp16)[name = tensor("_inversed_scores_73_cast_fp16")]; tensor scores_75_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_73_cast_fp16, cond = mask_3)[name = tensor("scores_75_cast_fp16")]; tensor var_3235_cast_fp16 = softmax(axis = var_30, x = scores_75_cast_fp16)[name = tensor("op_3235_cast_fp16")]; tensor input_969_cast_fp16 = select(a = var_11_to_fp16, b = var_3235_cast_fp16, cond = mask_3)[name = tensor("input_969_cast_fp16")]; tensor x_409_transpose_x_0 = const()[name = tensor("x_409_transpose_x_0"), val = tensor(false)]; tensor x_409_transpose_y_0 = const()[name = tensor("x_409_transpose_y_0"), val = tensor(false)]; tensor value_39_cast_fp16 = transpose(perm = value_39_perm_0, x = v_37_cast_fp16)[name = tensor("transpose_162")]; tensor x_409_cast_fp16 = matmul(transpose_x = x_409_transpose_x_0, transpose_y = x_409_transpose_y_0, x = input_969_cast_fp16, y = value_39_cast_fp16)[name = tensor("x_409_cast_fp16")]; tensor var_3239_perm_0 = const()[name = tensor("op_3239_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3240 = const()[name = tensor("op_3240"), val = tensor([1, -1, 1024])]; tensor var_3239_cast_fp16 = transpose(perm = var_3239_perm_0, x = x_409_cast_fp16)[name = tensor("transpose_158")]; tensor input_971_cast_fp16 = reshape(shape = var_3240, x = var_3239_cast_fp16)[name = tensor("input_971_cast_fp16")]; tensor module_layers_18_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_18_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(915414848)))]; tensor linear_169_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_out_weight_to_fp16, x = input_971_cast_fp16)[name = tensor("linear_169_cast_fp16")]; tensor input_975_cast_fp16 = add(x = input_967_cast_fp16, y = linear_169_cast_fp16)[name = tensor("input_975_cast_fp16")]; tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(917512064)))]; tensor module_layers_18_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(917514176)))]; tensor x_413_cast_fp16 = layer_norm(axes = x_413_axes_0, beta = module_layers_18_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_conv_weight_to_fp16, x = input_975_cast_fp16)[name = tensor("x_413_cast_fp16")]; tensor input_977_perm_0 = const()[name = tensor("input_977_perm_0"), val = tensor([0, 2, 1])]; tensor input_979_pad_type_0 = const()[name = tensor("input_979_pad_type_0"), val = tensor("valid")]; tensor input_979_strides_0 = const()[name = tensor("input_979_strides_0"), val = tensor([1])]; tensor input_979_pad_0 = const()[name = tensor("input_979_pad_0"), val = tensor([0, 0])]; tensor input_979_dilations_0 = const()[name = tensor("input_979_dilations_0"), val = tensor([1])]; tensor input_979_groups_0 = const()[name = tensor("input_979_groups_0"), val = tensor(1)]; tensor module_layers_18_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_18_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(917516288)))]; tensor input_977_cast_fp16 = transpose(perm = input_977_perm_0, x = x_413_cast_fp16)[name = tensor("transpose_157")]; tensor input_979_cast_fp16 = conv(dilations = input_979_dilations_0, groups = input_979_groups_0, pad = input_979_pad_0, pad_type = input_979_pad_type_0, strides = input_979_strides_0, weight = module_layers_18_conv_pointwise_conv1_weight_to_fp16, x = input_977_cast_fp16)[name = tensor("input_979_cast_fp16")]; tensor x_415_split_num_splits_0 = const()[name = tensor("x_415_split_num_splits_0"), val = tensor(2)]; tensor x_415_split_axis_0 = const()[name = tensor("x_415_split_axis_0"), val = tensor(1)]; tensor x_415_split_cast_fp16_0, tensor x_415_split_cast_fp16_1 = split(axis = x_415_split_axis_0, num_splits = x_415_split_num_splits_0, x = input_979_cast_fp16)[name = tensor("x_415_split_cast_fp16")]; tensor x_415_split_1_sigmoid_cast_fp16 = sigmoid(x = x_415_split_cast_fp16_1)[name = tensor("x_415_split_1_sigmoid_cast_fp16")]; tensor x_415_cast_fp16 = mul(x = x_415_split_cast_fp16_0, y = x_415_split_1_sigmoid_cast_fp16)[name = tensor("x_415_cast_fp16")]; tensor input_981_cast_fp16 = select(a = var_11_to_fp16, b = x_415_cast_fp16, cond = var_328)[name = tensor("input_981_cast_fp16")]; tensor input_983_pad_0 = const()[name = tensor("input_983_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_983_mode_0 = const()[name = tensor("input_983_mode_0"), val = tensor("constant")]; tensor const_197_to_fp16 = const()[name = tensor("const_197_to_fp16"), val = tensor(0x0p+0)]; tensor input_983_cast_fp16 = pad(constant_val = const_197_to_fp16, mode = input_983_mode_0, pad = input_983_pad_0, x = input_981_cast_fp16)[name = tensor("input_983_cast_fp16")]; tensor input_985_pad_type_0 = const()[name = tensor("input_985_pad_type_0"), val = tensor("valid")]; tensor input_985_groups_0 = const()[name = tensor("input_985_groups_0"), val = tensor(1024)]; tensor input_985_strides_0 = const()[name = tensor("input_985_strides_0"), val = tensor([1])]; tensor input_985_pad_0 = const()[name = tensor("input_985_pad_0"), val = tensor([0, 0])]; tensor input_985_dilations_0 = const()[name = tensor("input_985_dilations_0"), val = tensor([1])]; tensor const_284_to_fp16 = const()[name = tensor("const_284_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(921710656)))]; tensor const_285_to_fp16 = const()[name = tensor("const_285_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(921729152)))]; tensor input_987_cast_fp16 = conv(bias = const_285_to_fp16, dilations = input_985_dilations_0, groups = input_985_groups_0, pad = input_985_pad_0, pad_type = input_985_pad_type_0, strides = input_985_strides_0, weight = const_284_to_fp16, x = input_983_cast_fp16)[name = tensor("input_987_cast_fp16")]; tensor input_989_cast_fp16 = silu(x = input_987_cast_fp16)[name = tensor("input_989_cast_fp16")]; tensor x_417_pad_type_0 = const()[name = tensor("x_417_pad_type_0"), val = tensor("valid")]; tensor x_417_strides_0 = const()[name = tensor("x_417_strides_0"), val = tensor([1])]; tensor x_417_pad_0 = const()[name = tensor("x_417_pad_0"), val = tensor([0, 0])]; tensor x_417_dilations_0 = const()[name = tensor("x_417_dilations_0"), val = tensor([1])]; tensor x_417_groups_0 = const()[name = tensor("x_417_groups_0"), val = tensor(1)]; tensor module_layers_18_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_18_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(921731264)))]; tensor x_417_cast_fp16 = conv(dilations = x_417_dilations_0, groups = x_417_groups_0, pad = x_417_pad_0, pad_type = x_417_pad_type_0, strides = x_417_strides_0, weight = module_layers_18_conv_pointwise_conv2_weight_to_fp16, x = input_989_cast_fp16)[name = tensor("x_417_cast_fp16")]; tensor input_991_perm_0 = const()[name = tensor("input_991_perm_0"), val = tensor([0, 2, 1])]; tensor input_991_cast_fp16 = transpose(perm = input_991_perm_0, x = x_417_cast_fp16)[name = tensor("transpose_156")]; tensor input_993_cast_fp16 = add(x = input_975_cast_fp16, y = input_991_cast_fp16)[name = tensor("input_993_cast_fp16")]; tensor input_995_axes_0 = const()[name = tensor("input_995_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(923828480)))]; tensor module_layers_18_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(923830592)))]; tensor input_995_cast_fp16 = layer_norm(axes = input_995_axes_0, beta = module_layers_18_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward2_weight_to_fp16, x = input_993_cast_fp16)[name = tensor("input_995_cast_fp16")]; tensor module_layers_18_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_18_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(923832704)))]; tensor linear_170_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_18_feed_forward2_linear1_weight_to_fp16, x = input_995_cast_fp16)[name = tensor("linear_170_cast_fp16")]; tensor input_999_cast_fp16 = silu(x = linear_170_cast_fp16)[name = tensor("input_999_cast_fp16")]; tensor module_layers_18_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_18_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(932221376)))]; tensor linear_171_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_feed_forward2_linear2_weight_to_fp16, x = input_999_cast_fp16)[name = tensor("linear_171_cast_fp16")]; tensor var_3300_to_fp16 = const()[name = tensor("op_3300_to_fp16"), val = tensor(0x1p-1)]; tensor var_3301_cast_fp16 = mul(x = linear_171_cast_fp16, y = var_3300_to_fp16)[name = tensor("op_3301_cast_fp16")]; tensor input_1005_cast_fp16 = add(x = input_993_cast_fp16, y = var_3301_cast_fp16)[name = tensor("input_1005_cast_fp16")]; tensor input_1007_axes_0 = const()[name = tensor("input_1007_axes_0"), val = tensor([-1])]; tensor module_layers_18_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(940610048)))]; tensor module_layers_18_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(940612160)))]; tensor input_1007_cast_fp16 = layer_norm(axes = input_1007_axes_0, beta = module_layers_18_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_out_weight_to_fp16, x = input_1005_cast_fp16)[name = tensor("input_1007_cast_fp16")]; tensor input_1009_axes_0 = const()[name = tensor("input_1009_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(940614272)))]; tensor module_layers_19_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(940616384)))]; tensor input_1009_cast_fp16 = layer_norm(axes = input_1009_axes_0, beta = module_layers_19_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward1_weight_to_fp16, x = input_1007_cast_fp16)[name = tensor("input_1009_cast_fp16")]; tensor module_layers_19_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_19_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(940618496)))]; tensor linear_172_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_19_feed_forward1_linear1_weight_to_fp16, x = input_1009_cast_fp16)[name = tensor("linear_172_cast_fp16")]; tensor input_1013_cast_fp16 = silu(x = linear_172_cast_fp16)[name = tensor("input_1013_cast_fp16")]; tensor module_layers_19_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_19_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(949007168)))]; tensor linear_173_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_feed_forward1_linear2_weight_to_fp16, x = input_1013_cast_fp16)[name = tensor("linear_173_cast_fp16")]; tensor var_3329_to_fp16 = const()[name = tensor("op_3329_to_fp16"), val = tensor(0x1p-1)]; tensor var_3330_cast_fp16 = mul(x = linear_173_cast_fp16, y = var_3329_to_fp16)[name = tensor("op_3330_cast_fp16")]; tensor input_1019_cast_fp16 = add(x = input_1007_cast_fp16, y = var_3330_cast_fp16)[name = tensor("input_1019_cast_fp16")]; tensor query_39_axes_0 = const()[name = tensor("query_39_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(957395840)))]; tensor module_layers_19_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(957397952)))]; tensor query_39_cast_fp16 = layer_norm(axes = query_39_axes_0, beta = module_layers_19_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_self_att_weight_to_fp16, x = input_1019_cast_fp16)[name = tensor("query_39_cast_fp16")]; tensor module_layers_19_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(957400064)))]; tensor linear_174_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_q_weight_to_fp16, x = query_39_cast_fp16)[name = tensor("linear_174_cast_fp16")]; tensor var_3346 = const()[name = tensor("op_3346"), val = tensor([1, -1, 8, 128])]; tensor q_115_cast_fp16 = reshape(shape = var_3346, x = linear_174_cast_fp16)[name = tensor("q_115_cast_fp16")]; tensor module_layers_19_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(959497280)))]; tensor linear_175_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_k_weight_to_fp16, x = query_39_cast_fp16)[name = tensor("linear_175_cast_fp16")]; tensor var_3350 = const()[name = tensor("op_3350"), val = tensor([1, -1, 8, 128])]; tensor k_77_cast_fp16 = reshape(shape = var_3350, x = linear_175_cast_fp16)[name = tensor("k_77_cast_fp16")]; tensor module_layers_19_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(961594496)))]; tensor linear_176_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_v_weight_to_fp16, x = query_39_cast_fp16)[name = tensor("linear_176_cast_fp16")]; tensor var_3354 = const()[name = tensor("op_3354"), val = tensor([1, -1, 8, 128])]; tensor v_39_cast_fp16 = reshape(shape = var_3354, x = linear_176_cast_fp16)[name = tensor("v_39_cast_fp16")]; tensor value_41_perm_0 = const()[name = tensor("value_41_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_19_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(963691712)))]; tensor var_3366_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3366_cast_fp16")]; tensor module_layers_19_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(963693824)))]; tensor var_3368_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3368_cast_fp16")]; tensor q_with_bias_v_39_perm_0 = const()[name = tensor("q_with_bias_v_39_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_425_transpose_x_0 = const()[name = tensor("x_425_transpose_x_0"), val = tensor(false)]; tensor x_425_transpose_y_0 = const()[name = tensor("x_425_transpose_y_0"), val = tensor(false)]; tensor var_3370_to_fp16 = const()[name = tensor("op_3370_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(963695936)))]; tensor q_with_bias_v_39_cast_fp16 = transpose(perm = q_with_bias_v_39_perm_0, x = var_3368_cast_fp16)[name = tensor("transpose_154")]; tensor x_425_cast_fp16 = matmul(transpose_x = x_425_transpose_x_0, transpose_y = x_425_transpose_y_0, x = q_with_bias_v_39_cast_fp16, y = var_3370_to_fp16)[name = tensor("x_425_cast_fp16")]; tensor x_427_pad_0 = const()[name = tensor("x_427_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_427_mode_0 = const()[name = tensor("x_427_mode_0"), val = tensor("constant")]; tensor const_204_to_fp16 = const()[name = tensor("const_204_to_fp16"), val = tensor(0x0p+0)]; tensor x_427_cast_fp16 = pad(constant_val = const_204_to_fp16, mode = x_427_mode_0, pad = x_427_pad_0, x = x_425_cast_fp16)[name = tensor("x_427_cast_fp16")]; tensor var_3378 = const()[name = tensor("op_3378"), val = tensor([1, 8, -1, 188])]; tensor x_429_cast_fp16 = reshape(shape = var_3378, x = x_427_cast_fp16)[name = tensor("x_429_cast_fp16")]; tensor var_3382_begin_0 = const()[name = tensor("op_3382_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3382_end_0 = const()[name = tensor("op_3382_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3382_end_mask_0 = const()[name = tensor("op_3382_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3382_cast_fp16 = slice_by_index(begin = var_3382_begin_0, end = var_3382_end_0, end_mask = var_3382_end_mask_0, x = x_429_cast_fp16)[name = tensor("op_3382_cast_fp16")]; tensor var_3383 = const()[name = tensor("op_3383"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_77_cast_fp16 = reshape(shape = var_3383, x = var_3382_cast_fp16)[name = tensor("matrix_bd_77_cast_fp16")]; tensor matrix_ac_39_transpose_x_0 = const()[name = tensor("matrix_ac_39_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_39_transpose_y_0 = const()[name = tensor("matrix_ac_39_transpose_y_0"), val = tensor(false)]; tensor transpose_110_perm_0 = const()[name = tensor("transpose_110_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_111_perm_0 = const()[name = tensor("transpose_111_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_111 = transpose(perm = transpose_111_perm_0, x = k_77_cast_fp16)[name = tensor("transpose_152")]; tensor transpose_110 = transpose(perm = transpose_110_perm_0, x = var_3366_cast_fp16)[name = tensor("transpose_153")]; tensor matrix_ac_39_cast_fp16 = matmul(transpose_x = matrix_ac_39_transpose_x_0, transpose_y = matrix_ac_39_transpose_y_0, x = transpose_110, y = transpose_111)[name = tensor("matrix_ac_39_cast_fp16")]; tensor matrix_bd_79_begin_0 = const()[name = tensor("matrix_bd_79_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_79_end_0 = const()[name = tensor("matrix_bd_79_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_79_end_mask_0 = const()[name = tensor("matrix_bd_79_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_79_cast_fp16 = slice_by_index(begin = matrix_bd_79_begin_0, end = matrix_bd_79_end_0, end_mask = matrix_bd_79_end_mask_0, x = matrix_bd_77_cast_fp16)[name = tensor("matrix_bd_79_cast_fp16")]; tensor var_3392_cast_fp16 = add(x = matrix_ac_39_cast_fp16, y = matrix_bd_79_cast_fp16)[name = tensor("op_3392_cast_fp16")]; tensor _inversed_scores_77_y_0_to_fp16 = const()[name = tensor("_inversed_scores_77_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_77_cast_fp16 = mul(x = var_3392_cast_fp16, y = _inversed_scores_77_y_0_to_fp16)[name = tensor("_inversed_scores_77_cast_fp16")]; tensor scores_79_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_77_cast_fp16, cond = mask_3)[name = tensor("scores_79_cast_fp16")]; tensor var_3398_cast_fp16 = softmax(axis = var_30, x = scores_79_cast_fp16)[name = tensor("op_3398_cast_fp16")]; tensor input_1021_cast_fp16 = select(a = var_11_to_fp16, b = var_3398_cast_fp16, cond = mask_3)[name = tensor("input_1021_cast_fp16")]; tensor x_431_transpose_x_0 = const()[name = tensor("x_431_transpose_x_0"), val = tensor(false)]; tensor x_431_transpose_y_0 = const()[name = tensor("x_431_transpose_y_0"), val = tensor(false)]; tensor value_41_cast_fp16 = transpose(perm = value_41_perm_0, x = v_39_cast_fp16)[name = tensor("transpose_155")]; tensor x_431_cast_fp16 = matmul(transpose_x = x_431_transpose_x_0, transpose_y = x_431_transpose_y_0, x = input_1021_cast_fp16, y = value_41_cast_fp16)[name = tensor("x_431_cast_fp16")]; tensor var_3402_perm_0 = const()[name = tensor("op_3402_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3403 = const()[name = tensor("op_3403"), val = tensor([1, -1, 1024])]; tensor var_3402_cast_fp16 = transpose(perm = var_3402_perm_0, x = x_431_cast_fp16)[name = tensor("transpose_151")]; tensor input_1023_cast_fp16 = reshape(shape = var_3403, x = var_3402_cast_fp16)[name = tensor("input_1023_cast_fp16")]; tensor module_layers_19_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_19_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(964464000)))]; tensor linear_178_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_out_weight_to_fp16, x = input_1023_cast_fp16)[name = tensor("linear_178_cast_fp16")]; tensor input_1027_cast_fp16 = add(x = input_1019_cast_fp16, y = linear_178_cast_fp16)[name = tensor("input_1027_cast_fp16")]; tensor x_435_axes_0 = const()[name = tensor("x_435_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(966561216)))]; tensor module_layers_19_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(966563328)))]; tensor x_435_cast_fp16 = layer_norm(axes = x_435_axes_0, beta = module_layers_19_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_conv_weight_to_fp16, x = input_1027_cast_fp16)[name = tensor("x_435_cast_fp16")]; tensor input_1029_perm_0 = const()[name = tensor("input_1029_perm_0"), val = tensor([0, 2, 1])]; tensor input_1031_pad_type_0 = const()[name = tensor("input_1031_pad_type_0"), val = tensor("valid")]; tensor input_1031_strides_0 = const()[name = tensor("input_1031_strides_0"), val = tensor([1])]; tensor input_1031_pad_0 = const()[name = tensor("input_1031_pad_0"), val = tensor([0, 0])]; tensor input_1031_dilations_0 = const()[name = tensor("input_1031_dilations_0"), val = tensor([1])]; tensor input_1031_groups_0 = const()[name = tensor("input_1031_groups_0"), val = tensor(1)]; tensor module_layers_19_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_19_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(966565440)))]; tensor input_1029_cast_fp16 = transpose(perm = input_1029_perm_0, x = x_435_cast_fp16)[name = tensor("transpose_150")]; tensor input_1031_cast_fp16 = conv(dilations = input_1031_dilations_0, groups = input_1031_groups_0, pad = input_1031_pad_0, pad_type = input_1031_pad_type_0, strides = input_1031_strides_0, weight = module_layers_19_conv_pointwise_conv1_weight_to_fp16, x = input_1029_cast_fp16)[name = tensor("input_1031_cast_fp16")]; tensor x_437_split_num_splits_0 = const()[name = tensor("x_437_split_num_splits_0"), val = tensor(2)]; tensor x_437_split_axis_0 = const()[name = tensor("x_437_split_axis_0"), val = tensor(1)]; tensor x_437_split_cast_fp16_0, tensor x_437_split_cast_fp16_1 = split(axis = x_437_split_axis_0, num_splits = x_437_split_num_splits_0, x = input_1031_cast_fp16)[name = tensor("x_437_split_cast_fp16")]; tensor x_437_split_1_sigmoid_cast_fp16 = sigmoid(x = x_437_split_cast_fp16_1)[name = tensor("x_437_split_1_sigmoid_cast_fp16")]; tensor x_437_cast_fp16 = mul(x = x_437_split_cast_fp16_0, y = x_437_split_1_sigmoid_cast_fp16)[name = tensor("x_437_cast_fp16")]; tensor input_1033_cast_fp16 = select(a = var_11_to_fp16, b = x_437_cast_fp16, cond = var_328)[name = tensor("input_1033_cast_fp16")]; tensor input_1035_pad_0 = const()[name = tensor("input_1035_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1035_mode_0 = const()[name = tensor("input_1035_mode_0"), val = tensor("constant")]; tensor const_207_to_fp16 = const()[name = tensor("const_207_to_fp16"), val = tensor(0x0p+0)]; tensor input_1035_cast_fp16 = pad(constant_val = const_207_to_fp16, mode = input_1035_mode_0, pad = input_1035_pad_0, x = input_1033_cast_fp16)[name = tensor("input_1035_cast_fp16")]; tensor input_1037_pad_type_0 = const()[name = tensor("input_1037_pad_type_0"), val = tensor("valid")]; tensor input_1037_groups_0 = const()[name = tensor("input_1037_groups_0"), val = tensor(1024)]; tensor input_1037_strides_0 = const()[name = tensor("input_1037_strides_0"), val = tensor([1])]; tensor input_1037_pad_0 = const()[name = tensor("input_1037_pad_0"), val = tensor([0, 0])]; tensor input_1037_dilations_0 = const()[name = tensor("input_1037_dilations_0"), val = tensor([1])]; tensor const_286_to_fp16 = const()[name = tensor("const_286_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(970759808)))]; tensor const_287_to_fp16 = const()[name = tensor("const_287_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(970778304)))]; tensor input_1039_cast_fp16 = conv(bias = const_287_to_fp16, dilations = input_1037_dilations_0, groups = input_1037_groups_0, pad = input_1037_pad_0, pad_type = input_1037_pad_type_0, strides = input_1037_strides_0, weight = const_286_to_fp16, x = input_1035_cast_fp16)[name = tensor("input_1039_cast_fp16")]; tensor input_1041_cast_fp16 = silu(x = input_1039_cast_fp16)[name = tensor("input_1041_cast_fp16")]; tensor x_439_pad_type_0 = const()[name = tensor("x_439_pad_type_0"), val = tensor("valid")]; tensor x_439_strides_0 = const()[name = tensor("x_439_strides_0"), val = tensor([1])]; tensor x_439_pad_0 = const()[name = tensor("x_439_pad_0"), val = tensor([0, 0])]; tensor x_439_dilations_0 = const()[name = tensor("x_439_dilations_0"), val = tensor([1])]; tensor x_439_groups_0 = const()[name = tensor("x_439_groups_0"), val = tensor(1)]; tensor module_layers_19_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_19_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(970780416)))]; tensor x_439_cast_fp16 = conv(dilations = x_439_dilations_0, groups = x_439_groups_0, pad = x_439_pad_0, pad_type = x_439_pad_type_0, strides = x_439_strides_0, weight = module_layers_19_conv_pointwise_conv2_weight_to_fp16, x = input_1041_cast_fp16)[name = tensor("x_439_cast_fp16")]; tensor input_1043_perm_0 = const()[name = tensor("input_1043_perm_0"), val = tensor([0, 2, 1])]; tensor input_1043_cast_fp16 = transpose(perm = input_1043_perm_0, x = x_439_cast_fp16)[name = tensor("transpose_149")]; tensor input_1045_cast_fp16 = add(x = input_1027_cast_fp16, y = input_1043_cast_fp16)[name = tensor("input_1045_cast_fp16")]; tensor input_1047_axes_0 = const()[name = tensor("input_1047_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(972877632)))]; tensor module_layers_19_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(972879744)))]; tensor input_1047_cast_fp16 = layer_norm(axes = input_1047_axes_0, beta = module_layers_19_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward2_weight_to_fp16, x = input_1045_cast_fp16)[name = tensor("input_1047_cast_fp16")]; tensor module_layers_19_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_19_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(972881856)))]; tensor linear_179_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_19_feed_forward2_linear1_weight_to_fp16, x = input_1047_cast_fp16)[name = tensor("linear_179_cast_fp16")]; tensor input_1051_cast_fp16 = silu(x = linear_179_cast_fp16)[name = tensor("input_1051_cast_fp16")]; tensor module_layers_19_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_19_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(981270528)))]; tensor linear_180_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_feed_forward2_linear2_weight_to_fp16, x = input_1051_cast_fp16)[name = tensor("linear_180_cast_fp16")]; tensor var_3463_to_fp16 = const()[name = tensor("op_3463_to_fp16"), val = tensor(0x1p-1)]; tensor var_3464_cast_fp16 = mul(x = linear_180_cast_fp16, y = var_3463_to_fp16)[name = tensor("op_3464_cast_fp16")]; tensor input_1057_cast_fp16 = add(x = input_1045_cast_fp16, y = var_3464_cast_fp16)[name = tensor("input_1057_cast_fp16")]; tensor input_1059_axes_0 = const()[name = tensor("input_1059_axes_0"), val = tensor([-1])]; tensor module_layers_19_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(989659200)))]; tensor module_layers_19_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(989661312)))]; tensor input_1059_cast_fp16 = layer_norm(axes = input_1059_axes_0, beta = module_layers_19_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_out_weight_to_fp16, x = input_1057_cast_fp16)[name = tensor("input_1059_cast_fp16")]; tensor input_1061_axes_0 = const()[name = tensor("input_1061_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(989663424)))]; tensor module_layers_20_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(989665536)))]; tensor input_1061_cast_fp16 = layer_norm(axes = input_1061_axes_0, beta = module_layers_20_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward1_weight_to_fp16, x = input_1059_cast_fp16)[name = tensor("input_1061_cast_fp16")]; tensor module_layers_20_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_20_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(989667648)))]; tensor linear_181_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_20_feed_forward1_linear1_weight_to_fp16, x = input_1061_cast_fp16)[name = tensor("linear_181_cast_fp16")]; tensor input_1065_cast_fp16 = silu(x = linear_181_cast_fp16)[name = tensor("input_1065_cast_fp16")]; tensor module_layers_20_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_20_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(998056320)))]; tensor linear_182_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_feed_forward1_linear2_weight_to_fp16, x = input_1065_cast_fp16)[name = tensor("linear_182_cast_fp16")]; tensor var_3492_to_fp16 = const()[name = tensor("op_3492_to_fp16"), val = tensor(0x1p-1)]; tensor var_3493_cast_fp16 = mul(x = linear_182_cast_fp16, y = var_3492_to_fp16)[name = tensor("op_3493_cast_fp16")]; tensor input_1071_cast_fp16 = add(x = input_1059_cast_fp16, y = var_3493_cast_fp16)[name = tensor("input_1071_cast_fp16")]; tensor query_41_axes_0 = const()[name = tensor("query_41_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1006444992)))]; tensor module_layers_20_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1006447104)))]; tensor query_41_cast_fp16 = layer_norm(axes = query_41_axes_0, beta = module_layers_20_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_self_att_weight_to_fp16, x = input_1071_cast_fp16)[name = tensor("query_41_cast_fp16")]; tensor module_layers_20_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1006449216)))]; tensor linear_183_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_q_weight_to_fp16, x = query_41_cast_fp16)[name = tensor("linear_183_cast_fp16")]; tensor var_3509 = const()[name = tensor("op_3509"), val = tensor([1, -1, 8, 128])]; tensor q_121_cast_fp16 = reshape(shape = var_3509, x = linear_183_cast_fp16)[name = tensor("q_121_cast_fp16")]; tensor module_layers_20_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1008546432)))]; tensor linear_184_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_k_weight_to_fp16, x = query_41_cast_fp16)[name = tensor("linear_184_cast_fp16")]; tensor var_3513 = const()[name = tensor("op_3513"), val = tensor([1, -1, 8, 128])]; tensor k_81_cast_fp16 = reshape(shape = var_3513, x = linear_184_cast_fp16)[name = tensor("k_81_cast_fp16")]; tensor module_layers_20_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1010643648)))]; tensor linear_185_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_v_weight_to_fp16, x = query_41_cast_fp16)[name = tensor("linear_185_cast_fp16")]; tensor var_3517 = const()[name = tensor("op_3517"), val = tensor([1, -1, 8, 128])]; tensor v_41_cast_fp16 = reshape(shape = var_3517, x = linear_185_cast_fp16)[name = tensor("v_41_cast_fp16")]; tensor value_43_perm_0 = const()[name = tensor("value_43_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_20_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1012740864)))]; tensor var_3529_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3529_cast_fp16")]; tensor module_layers_20_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1012742976)))]; tensor var_3531_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3531_cast_fp16")]; tensor q_with_bias_v_41_perm_0 = const()[name = tensor("q_with_bias_v_41_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_447_transpose_x_0 = const()[name = tensor("x_447_transpose_x_0"), val = tensor(false)]; tensor x_447_transpose_y_0 = const()[name = tensor("x_447_transpose_y_0"), val = tensor(false)]; tensor var_3533_to_fp16 = const()[name = tensor("op_3533_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1012745088)))]; tensor q_with_bias_v_41_cast_fp16 = transpose(perm = q_with_bias_v_41_perm_0, x = var_3531_cast_fp16)[name = tensor("transpose_147")]; tensor x_447_cast_fp16 = matmul(transpose_x = x_447_transpose_x_0, transpose_y = x_447_transpose_y_0, x = q_with_bias_v_41_cast_fp16, y = var_3533_to_fp16)[name = tensor("x_447_cast_fp16")]; tensor x_449_pad_0 = const()[name = tensor("x_449_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_449_mode_0 = const()[name = tensor("x_449_mode_0"), val = tensor("constant")]; tensor const_214_to_fp16 = const()[name = tensor("const_214_to_fp16"), val = tensor(0x0p+0)]; tensor x_449_cast_fp16 = pad(constant_val = const_214_to_fp16, mode = x_449_mode_0, pad = x_449_pad_0, x = x_447_cast_fp16)[name = tensor("x_449_cast_fp16")]; tensor var_3541 = const()[name = tensor("op_3541"), val = tensor([1, 8, -1, 188])]; tensor x_451_cast_fp16 = reshape(shape = var_3541, x = x_449_cast_fp16)[name = tensor("x_451_cast_fp16")]; tensor var_3545_begin_0 = const()[name = tensor("op_3545_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3545_end_0 = const()[name = tensor("op_3545_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3545_end_mask_0 = const()[name = tensor("op_3545_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3545_cast_fp16 = slice_by_index(begin = var_3545_begin_0, end = var_3545_end_0, end_mask = var_3545_end_mask_0, x = x_451_cast_fp16)[name = tensor("op_3545_cast_fp16")]; tensor var_3546 = const()[name = tensor("op_3546"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_81_cast_fp16 = reshape(shape = var_3546, x = var_3545_cast_fp16)[name = tensor("matrix_bd_81_cast_fp16")]; tensor matrix_ac_41_transpose_x_0 = const()[name = tensor("matrix_ac_41_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_41_transpose_y_0 = const()[name = tensor("matrix_ac_41_transpose_y_0"), val = tensor(false)]; tensor transpose_112_perm_0 = const()[name = tensor("transpose_112_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_113_perm_0 = const()[name = tensor("transpose_113_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_113 = transpose(perm = transpose_113_perm_0, x = k_81_cast_fp16)[name = tensor("transpose_145")]; tensor transpose_112 = transpose(perm = transpose_112_perm_0, x = var_3529_cast_fp16)[name = tensor("transpose_146")]; tensor matrix_ac_41_cast_fp16 = matmul(transpose_x = matrix_ac_41_transpose_x_0, transpose_y = matrix_ac_41_transpose_y_0, x = transpose_112, y = transpose_113)[name = tensor("matrix_ac_41_cast_fp16")]; tensor matrix_bd_83_begin_0 = const()[name = tensor("matrix_bd_83_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_83_end_0 = const()[name = tensor("matrix_bd_83_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_83_end_mask_0 = const()[name = tensor("matrix_bd_83_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_83_cast_fp16 = slice_by_index(begin = matrix_bd_83_begin_0, end = matrix_bd_83_end_0, end_mask = matrix_bd_83_end_mask_0, x = matrix_bd_81_cast_fp16)[name = tensor("matrix_bd_83_cast_fp16")]; tensor var_3555_cast_fp16 = add(x = matrix_ac_41_cast_fp16, y = matrix_bd_83_cast_fp16)[name = tensor("op_3555_cast_fp16")]; tensor _inversed_scores_81_y_0_to_fp16 = const()[name = tensor("_inversed_scores_81_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_81_cast_fp16 = mul(x = var_3555_cast_fp16, y = _inversed_scores_81_y_0_to_fp16)[name = tensor("_inversed_scores_81_cast_fp16")]; tensor scores_83_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_81_cast_fp16, cond = mask_3)[name = tensor("scores_83_cast_fp16")]; tensor var_3561_cast_fp16 = softmax(axis = var_30, x = scores_83_cast_fp16)[name = tensor("op_3561_cast_fp16")]; tensor input_1073_cast_fp16 = select(a = var_11_to_fp16, b = var_3561_cast_fp16, cond = mask_3)[name = tensor("input_1073_cast_fp16")]; tensor x_453_transpose_x_0 = const()[name = tensor("x_453_transpose_x_0"), val = tensor(false)]; tensor x_453_transpose_y_0 = const()[name = tensor("x_453_transpose_y_0"), val = tensor(false)]; tensor value_43_cast_fp16 = transpose(perm = value_43_perm_0, x = v_41_cast_fp16)[name = tensor("transpose_148")]; tensor x_453_cast_fp16 = matmul(transpose_x = x_453_transpose_x_0, transpose_y = x_453_transpose_y_0, x = input_1073_cast_fp16, y = value_43_cast_fp16)[name = tensor("x_453_cast_fp16")]; tensor var_3565_perm_0 = const()[name = tensor("op_3565_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3566 = const()[name = tensor("op_3566"), val = tensor([1, -1, 1024])]; tensor var_3565_cast_fp16 = transpose(perm = var_3565_perm_0, x = x_453_cast_fp16)[name = tensor("transpose_144")]; tensor input_1075_cast_fp16 = reshape(shape = var_3566, x = var_3565_cast_fp16)[name = tensor("input_1075_cast_fp16")]; tensor module_layers_20_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_20_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1013513152)))]; tensor linear_187_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_out_weight_to_fp16, x = input_1075_cast_fp16)[name = tensor("linear_187_cast_fp16")]; tensor input_1079_cast_fp16 = add(x = input_1071_cast_fp16, y = linear_187_cast_fp16)[name = tensor("input_1079_cast_fp16")]; tensor x_457_axes_0 = const()[name = tensor("x_457_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1015610368)))]; tensor module_layers_20_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1015612480)))]; tensor x_457_cast_fp16 = layer_norm(axes = x_457_axes_0, beta = module_layers_20_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_conv_weight_to_fp16, x = input_1079_cast_fp16)[name = tensor("x_457_cast_fp16")]; tensor input_1081_perm_0 = const()[name = tensor("input_1081_perm_0"), val = tensor([0, 2, 1])]; tensor input_1083_pad_type_0 = const()[name = tensor("input_1083_pad_type_0"), val = tensor("valid")]; tensor input_1083_strides_0 = const()[name = tensor("input_1083_strides_0"), val = tensor([1])]; tensor input_1083_pad_0 = const()[name = tensor("input_1083_pad_0"), val = tensor([0, 0])]; tensor input_1083_dilations_0 = const()[name = tensor("input_1083_dilations_0"), val = tensor([1])]; tensor input_1083_groups_0 = const()[name = tensor("input_1083_groups_0"), val = tensor(1)]; tensor module_layers_20_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_20_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1015614592)))]; tensor input_1081_cast_fp16 = transpose(perm = input_1081_perm_0, x = x_457_cast_fp16)[name = tensor("transpose_143")]; tensor input_1083_cast_fp16 = conv(dilations = input_1083_dilations_0, groups = input_1083_groups_0, pad = input_1083_pad_0, pad_type = input_1083_pad_type_0, strides = input_1083_strides_0, weight = module_layers_20_conv_pointwise_conv1_weight_to_fp16, x = input_1081_cast_fp16)[name = tensor("input_1083_cast_fp16")]; tensor x_459_split_num_splits_0 = const()[name = tensor("x_459_split_num_splits_0"), val = tensor(2)]; tensor x_459_split_axis_0 = const()[name = tensor("x_459_split_axis_0"), val = tensor(1)]; tensor x_459_split_cast_fp16_0, tensor x_459_split_cast_fp16_1 = split(axis = x_459_split_axis_0, num_splits = x_459_split_num_splits_0, x = input_1083_cast_fp16)[name = tensor("x_459_split_cast_fp16")]; tensor x_459_split_1_sigmoid_cast_fp16 = sigmoid(x = x_459_split_cast_fp16_1)[name = tensor("x_459_split_1_sigmoid_cast_fp16")]; tensor x_459_cast_fp16 = mul(x = x_459_split_cast_fp16_0, y = x_459_split_1_sigmoid_cast_fp16)[name = tensor("x_459_cast_fp16")]; tensor input_1085_cast_fp16 = select(a = var_11_to_fp16, b = x_459_cast_fp16, cond = var_328)[name = tensor("input_1085_cast_fp16")]; tensor input_1087_pad_0 = const()[name = tensor("input_1087_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1087_mode_0 = const()[name = tensor("input_1087_mode_0"), val = tensor("constant")]; tensor const_217_to_fp16 = const()[name = tensor("const_217_to_fp16"), val = tensor(0x0p+0)]; tensor input_1087_cast_fp16 = pad(constant_val = const_217_to_fp16, mode = input_1087_mode_0, pad = input_1087_pad_0, x = input_1085_cast_fp16)[name = tensor("input_1087_cast_fp16")]; tensor input_1089_pad_type_0 = const()[name = tensor("input_1089_pad_type_0"), val = tensor("valid")]; tensor input_1089_groups_0 = const()[name = tensor("input_1089_groups_0"), val = tensor(1024)]; tensor input_1089_strides_0 = const()[name = tensor("input_1089_strides_0"), val = tensor([1])]; tensor input_1089_pad_0 = const()[name = tensor("input_1089_pad_0"), val = tensor([0, 0])]; tensor input_1089_dilations_0 = const()[name = tensor("input_1089_dilations_0"), val = tensor([1])]; tensor const_288_to_fp16 = const()[name = tensor("const_288_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1019808960)))]; tensor const_289_to_fp16 = const()[name = tensor("const_289_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1019827456)))]; tensor input_1091_cast_fp16 = conv(bias = const_289_to_fp16, dilations = input_1089_dilations_0, groups = input_1089_groups_0, pad = input_1089_pad_0, pad_type = input_1089_pad_type_0, strides = input_1089_strides_0, weight = const_288_to_fp16, x = input_1087_cast_fp16)[name = tensor("input_1091_cast_fp16")]; tensor input_1093_cast_fp16 = silu(x = input_1091_cast_fp16)[name = tensor("input_1093_cast_fp16")]; tensor x_461_pad_type_0 = const()[name = tensor("x_461_pad_type_0"), val = tensor("valid")]; tensor x_461_strides_0 = const()[name = tensor("x_461_strides_0"), val = tensor([1])]; tensor x_461_pad_0 = const()[name = tensor("x_461_pad_0"), val = tensor([0, 0])]; tensor x_461_dilations_0 = const()[name = tensor("x_461_dilations_0"), val = tensor([1])]; tensor x_461_groups_0 = const()[name = tensor("x_461_groups_0"), val = tensor(1)]; tensor module_layers_20_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_20_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1019829568)))]; tensor x_461_cast_fp16 = conv(dilations = x_461_dilations_0, groups = x_461_groups_0, pad = x_461_pad_0, pad_type = x_461_pad_type_0, strides = x_461_strides_0, weight = module_layers_20_conv_pointwise_conv2_weight_to_fp16, x = input_1093_cast_fp16)[name = tensor("x_461_cast_fp16")]; tensor input_1095_perm_0 = const()[name = tensor("input_1095_perm_0"), val = tensor([0, 2, 1])]; tensor input_1095_cast_fp16 = transpose(perm = input_1095_perm_0, x = x_461_cast_fp16)[name = tensor("transpose_142")]; tensor input_1097_cast_fp16 = add(x = input_1079_cast_fp16, y = input_1095_cast_fp16)[name = tensor("input_1097_cast_fp16")]; tensor input_1099_axes_0 = const()[name = tensor("input_1099_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1021926784)))]; tensor module_layers_20_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1021928896)))]; tensor input_1099_cast_fp16 = layer_norm(axes = input_1099_axes_0, beta = module_layers_20_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward2_weight_to_fp16, x = input_1097_cast_fp16)[name = tensor("input_1099_cast_fp16")]; tensor module_layers_20_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_20_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1021931008)))]; tensor linear_188_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_20_feed_forward2_linear1_weight_to_fp16, x = input_1099_cast_fp16)[name = tensor("linear_188_cast_fp16")]; tensor input_1103_cast_fp16 = silu(x = linear_188_cast_fp16)[name = tensor("input_1103_cast_fp16")]; tensor module_layers_20_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_20_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1030319680)))]; tensor linear_189_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_feed_forward2_linear2_weight_to_fp16, x = input_1103_cast_fp16)[name = tensor("linear_189_cast_fp16")]; tensor var_3626_to_fp16 = const()[name = tensor("op_3626_to_fp16"), val = tensor(0x1p-1)]; tensor var_3627_cast_fp16 = mul(x = linear_189_cast_fp16, y = var_3626_to_fp16)[name = tensor("op_3627_cast_fp16")]; tensor input_1109_cast_fp16 = add(x = input_1097_cast_fp16, y = var_3627_cast_fp16)[name = tensor("input_1109_cast_fp16")]; tensor input_1111_axes_0 = const()[name = tensor("input_1111_axes_0"), val = tensor([-1])]; tensor module_layers_20_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1038708352)))]; tensor module_layers_20_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1038710464)))]; tensor input_1111_cast_fp16 = layer_norm(axes = input_1111_axes_0, beta = module_layers_20_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_out_weight_to_fp16, x = input_1109_cast_fp16)[name = tensor("input_1111_cast_fp16")]; tensor input_1113_axes_0 = const()[name = tensor("input_1113_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1038712576)))]; tensor module_layers_21_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1038714688)))]; tensor input_1113_cast_fp16 = layer_norm(axes = input_1113_axes_0, beta = module_layers_21_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward1_weight_to_fp16, x = input_1111_cast_fp16)[name = tensor("input_1113_cast_fp16")]; tensor module_layers_21_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_21_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1038716800)))]; tensor linear_190_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_21_feed_forward1_linear1_weight_to_fp16, x = input_1113_cast_fp16)[name = tensor("linear_190_cast_fp16")]; tensor input_1117_cast_fp16 = silu(x = linear_190_cast_fp16)[name = tensor("input_1117_cast_fp16")]; tensor module_layers_21_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_21_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1047105472)))]; tensor linear_191_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_feed_forward1_linear2_weight_to_fp16, x = input_1117_cast_fp16)[name = tensor("linear_191_cast_fp16")]; tensor var_3655_to_fp16 = const()[name = tensor("op_3655_to_fp16"), val = tensor(0x1p-1)]; tensor var_3656_cast_fp16 = mul(x = linear_191_cast_fp16, y = var_3655_to_fp16)[name = tensor("op_3656_cast_fp16")]; tensor input_1123_cast_fp16 = add(x = input_1111_cast_fp16, y = var_3656_cast_fp16)[name = tensor("input_1123_cast_fp16")]; tensor query_43_axes_0 = const()[name = tensor("query_43_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1055494144)))]; tensor module_layers_21_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1055496256)))]; tensor query_43_cast_fp16 = layer_norm(axes = query_43_axes_0, beta = module_layers_21_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_self_att_weight_to_fp16, x = input_1123_cast_fp16)[name = tensor("query_43_cast_fp16")]; tensor module_layers_21_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1055498368)))]; tensor linear_192_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_q_weight_to_fp16, x = query_43_cast_fp16)[name = tensor("linear_192_cast_fp16")]; tensor var_3672 = const()[name = tensor("op_3672"), val = tensor([1, -1, 8, 128])]; tensor q_127_cast_fp16 = reshape(shape = var_3672, x = linear_192_cast_fp16)[name = tensor("q_127_cast_fp16")]; tensor module_layers_21_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1057595584)))]; tensor linear_193_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_k_weight_to_fp16, x = query_43_cast_fp16)[name = tensor("linear_193_cast_fp16")]; tensor var_3676 = const()[name = tensor("op_3676"), val = tensor([1, -1, 8, 128])]; tensor k_85_cast_fp16 = reshape(shape = var_3676, x = linear_193_cast_fp16)[name = tensor("k_85_cast_fp16")]; tensor module_layers_21_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1059692800)))]; tensor linear_194_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_v_weight_to_fp16, x = query_43_cast_fp16)[name = tensor("linear_194_cast_fp16")]; tensor var_3680 = const()[name = tensor("op_3680"), val = tensor([1, -1, 8, 128])]; tensor v_43_cast_fp16 = reshape(shape = var_3680, x = linear_194_cast_fp16)[name = tensor("v_43_cast_fp16")]; tensor value_45_perm_0 = const()[name = tensor("value_45_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_21_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1061790016)))]; tensor var_3692_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3692_cast_fp16")]; tensor module_layers_21_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1061792128)))]; tensor var_3694_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3694_cast_fp16")]; tensor q_with_bias_v_43_perm_0 = const()[name = tensor("q_with_bias_v_43_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_469_transpose_x_0 = const()[name = tensor("x_469_transpose_x_0"), val = tensor(false)]; tensor x_469_transpose_y_0 = const()[name = tensor("x_469_transpose_y_0"), val = tensor(false)]; tensor var_3696_to_fp16 = const()[name = tensor("op_3696_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1061794240)))]; tensor q_with_bias_v_43_cast_fp16 = transpose(perm = q_with_bias_v_43_perm_0, x = var_3694_cast_fp16)[name = tensor("transpose_140")]; tensor x_469_cast_fp16 = matmul(transpose_x = x_469_transpose_x_0, transpose_y = x_469_transpose_y_0, x = q_with_bias_v_43_cast_fp16, y = var_3696_to_fp16)[name = tensor("x_469_cast_fp16")]; tensor x_471_pad_0 = const()[name = tensor("x_471_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_471_mode_0 = const()[name = tensor("x_471_mode_0"), val = tensor("constant")]; tensor const_224_to_fp16 = const()[name = tensor("const_224_to_fp16"), val = tensor(0x0p+0)]; tensor x_471_cast_fp16 = pad(constant_val = const_224_to_fp16, mode = x_471_mode_0, pad = x_471_pad_0, x = x_469_cast_fp16)[name = tensor("x_471_cast_fp16")]; tensor var_3704 = const()[name = tensor("op_3704"), val = tensor([1, 8, -1, 188])]; tensor x_473_cast_fp16 = reshape(shape = var_3704, x = x_471_cast_fp16)[name = tensor("x_473_cast_fp16")]; tensor var_3708_begin_0 = const()[name = tensor("op_3708_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3708_end_0 = const()[name = tensor("op_3708_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3708_end_mask_0 = const()[name = tensor("op_3708_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3708_cast_fp16 = slice_by_index(begin = var_3708_begin_0, end = var_3708_end_0, end_mask = var_3708_end_mask_0, x = x_473_cast_fp16)[name = tensor("op_3708_cast_fp16")]; tensor var_3709 = const()[name = tensor("op_3709"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_85_cast_fp16 = reshape(shape = var_3709, x = var_3708_cast_fp16)[name = tensor("matrix_bd_85_cast_fp16")]; tensor matrix_ac_43_transpose_x_0 = const()[name = tensor("matrix_ac_43_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_43_transpose_y_0 = const()[name = tensor("matrix_ac_43_transpose_y_0"), val = tensor(false)]; tensor transpose_114_perm_0 = const()[name = tensor("transpose_114_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_115_perm_0 = const()[name = tensor("transpose_115_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_115 = transpose(perm = transpose_115_perm_0, x = k_85_cast_fp16)[name = tensor("transpose_138")]; tensor transpose_114 = transpose(perm = transpose_114_perm_0, x = var_3692_cast_fp16)[name = tensor("transpose_139")]; tensor matrix_ac_43_cast_fp16 = matmul(transpose_x = matrix_ac_43_transpose_x_0, transpose_y = matrix_ac_43_transpose_y_0, x = transpose_114, y = transpose_115)[name = tensor("matrix_ac_43_cast_fp16")]; tensor matrix_bd_87_begin_0 = const()[name = tensor("matrix_bd_87_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_87_end_0 = const()[name = tensor("matrix_bd_87_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_87_end_mask_0 = const()[name = tensor("matrix_bd_87_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_87_cast_fp16 = slice_by_index(begin = matrix_bd_87_begin_0, end = matrix_bd_87_end_0, end_mask = matrix_bd_87_end_mask_0, x = matrix_bd_85_cast_fp16)[name = tensor("matrix_bd_87_cast_fp16")]; tensor var_3718_cast_fp16 = add(x = matrix_ac_43_cast_fp16, y = matrix_bd_87_cast_fp16)[name = tensor("op_3718_cast_fp16")]; tensor _inversed_scores_85_y_0_to_fp16 = const()[name = tensor("_inversed_scores_85_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_85_cast_fp16 = mul(x = var_3718_cast_fp16, y = _inversed_scores_85_y_0_to_fp16)[name = tensor("_inversed_scores_85_cast_fp16")]; tensor scores_87_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_85_cast_fp16, cond = mask_3)[name = tensor("scores_87_cast_fp16")]; tensor var_3724_cast_fp16 = softmax(axis = var_30, x = scores_87_cast_fp16)[name = tensor("op_3724_cast_fp16")]; tensor input_1125_cast_fp16 = select(a = var_11_to_fp16, b = var_3724_cast_fp16, cond = mask_3)[name = tensor("input_1125_cast_fp16")]; tensor x_475_transpose_x_0 = const()[name = tensor("x_475_transpose_x_0"), val = tensor(false)]; tensor x_475_transpose_y_0 = const()[name = tensor("x_475_transpose_y_0"), val = tensor(false)]; tensor value_45_cast_fp16 = transpose(perm = value_45_perm_0, x = v_43_cast_fp16)[name = tensor("transpose_141")]; tensor x_475_cast_fp16 = matmul(transpose_x = x_475_transpose_x_0, transpose_y = x_475_transpose_y_0, x = input_1125_cast_fp16, y = value_45_cast_fp16)[name = tensor("x_475_cast_fp16")]; tensor var_3728_perm_0 = const()[name = tensor("op_3728_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3729 = const()[name = tensor("op_3729"), val = tensor([1, -1, 1024])]; tensor var_3728_cast_fp16 = transpose(perm = var_3728_perm_0, x = x_475_cast_fp16)[name = tensor("transpose_137")]; tensor input_1127_cast_fp16 = reshape(shape = var_3729, x = var_3728_cast_fp16)[name = tensor("input_1127_cast_fp16")]; tensor module_layers_21_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_21_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1062562304)))]; tensor linear_196_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_out_weight_to_fp16, x = input_1127_cast_fp16)[name = tensor("linear_196_cast_fp16")]; tensor input_1131_cast_fp16 = add(x = input_1123_cast_fp16, y = linear_196_cast_fp16)[name = tensor("input_1131_cast_fp16")]; tensor x_479_axes_0 = const()[name = tensor("x_479_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1064659520)))]; tensor module_layers_21_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1064661632)))]; tensor x_479_cast_fp16 = layer_norm(axes = x_479_axes_0, beta = module_layers_21_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_conv_weight_to_fp16, x = input_1131_cast_fp16)[name = tensor("x_479_cast_fp16")]; tensor input_1133_perm_0 = const()[name = tensor("input_1133_perm_0"), val = tensor([0, 2, 1])]; tensor input_1135_pad_type_0 = const()[name = tensor("input_1135_pad_type_0"), val = tensor("valid")]; tensor input_1135_strides_0 = const()[name = tensor("input_1135_strides_0"), val = tensor([1])]; tensor input_1135_pad_0 = const()[name = tensor("input_1135_pad_0"), val = tensor([0, 0])]; tensor input_1135_dilations_0 = const()[name = tensor("input_1135_dilations_0"), val = tensor([1])]; tensor input_1135_groups_0 = const()[name = tensor("input_1135_groups_0"), val = tensor(1)]; tensor module_layers_21_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_21_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1064663744)))]; tensor input_1133_cast_fp16 = transpose(perm = input_1133_perm_0, x = x_479_cast_fp16)[name = tensor("transpose_136")]; tensor input_1135_cast_fp16 = conv(dilations = input_1135_dilations_0, groups = input_1135_groups_0, pad = input_1135_pad_0, pad_type = input_1135_pad_type_0, strides = input_1135_strides_0, weight = module_layers_21_conv_pointwise_conv1_weight_to_fp16, x = input_1133_cast_fp16)[name = tensor("input_1135_cast_fp16")]; tensor x_481_split_num_splits_0 = const()[name = tensor("x_481_split_num_splits_0"), val = tensor(2)]; tensor x_481_split_axis_0 = const()[name = tensor("x_481_split_axis_0"), val = tensor(1)]; tensor x_481_split_cast_fp16_0, tensor x_481_split_cast_fp16_1 = split(axis = x_481_split_axis_0, num_splits = x_481_split_num_splits_0, x = input_1135_cast_fp16)[name = tensor("x_481_split_cast_fp16")]; tensor x_481_split_1_sigmoid_cast_fp16 = sigmoid(x = x_481_split_cast_fp16_1)[name = tensor("x_481_split_1_sigmoid_cast_fp16")]; tensor x_481_cast_fp16 = mul(x = x_481_split_cast_fp16_0, y = x_481_split_1_sigmoid_cast_fp16)[name = tensor("x_481_cast_fp16")]; tensor input_1137_cast_fp16 = select(a = var_11_to_fp16, b = x_481_cast_fp16, cond = var_328)[name = tensor("input_1137_cast_fp16")]; tensor input_1139_pad_0 = const()[name = tensor("input_1139_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1139_mode_0 = const()[name = tensor("input_1139_mode_0"), val = tensor("constant")]; tensor const_227_to_fp16 = const()[name = tensor("const_227_to_fp16"), val = tensor(0x0p+0)]; tensor input_1139_cast_fp16 = pad(constant_val = const_227_to_fp16, mode = input_1139_mode_0, pad = input_1139_pad_0, x = input_1137_cast_fp16)[name = tensor("input_1139_cast_fp16")]; tensor input_1141_pad_type_0 = const()[name = tensor("input_1141_pad_type_0"), val = tensor("valid")]; tensor input_1141_groups_0 = const()[name = tensor("input_1141_groups_0"), val = tensor(1024)]; tensor input_1141_strides_0 = const()[name = tensor("input_1141_strides_0"), val = tensor([1])]; tensor input_1141_pad_0 = const()[name = tensor("input_1141_pad_0"), val = tensor([0, 0])]; tensor input_1141_dilations_0 = const()[name = tensor("input_1141_dilations_0"), val = tensor([1])]; tensor const_290_to_fp16 = const()[name = tensor("const_290_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1068858112)))]; tensor const_291_to_fp16 = const()[name = tensor("const_291_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1068876608)))]; tensor input_1143_cast_fp16 = conv(bias = const_291_to_fp16, dilations = input_1141_dilations_0, groups = input_1141_groups_0, pad = input_1141_pad_0, pad_type = input_1141_pad_type_0, strides = input_1141_strides_0, weight = const_290_to_fp16, x = input_1139_cast_fp16)[name = tensor("input_1143_cast_fp16")]; tensor input_1145_cast_fp16 = silu(x = input_1143_cast_fp16)[name = tensor("input_1145_cast_fp16")]; tensor x_483_pad_type_0 = const()[name = tensor("x_483_pad_type_0"), val = tensor("valid")]; tensor x_483_strides_0 = const()[name = tensor("x_483_strides_0"), val = tensor([1])]; tensor x_483_pad_0 = const()[name = tensor("x_483_pad_0"), val = tensor([0, 0])]; tensor x_483_dilations_0 = const()[name = tensor("x_483_dilations_0"), val = tensor([1])]; tensor x_483_groups_0 = const()[name = tensor("x_483_groups_0"), val = tensor(1)]; tensor module_layers_21_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_21_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1068878720)))]; tensor x_483_cast_fp16 = conv(dilations = x_483_dilations_0, groups = x_483_groups_0, pad = x_483_pad_0, pad_type = x_483_pad_type_0, strides = x_483_strides_0, weight = module_layers_21_conv_pointwise_conv2_weight_to_fp16, x = input_1145_cast_fp16)[name = tensor("x_483_cast_fp16")]; tensor input_1147_perm_0 = const()[name = tensor("input_1147_perm_0"), val = tensor([0, 2, 1])]; tensor input_1147_cast_fp16 = transpose(perm = input_1147_perm_0, x = x_483_cast_fp16)[name = tensor("transpose_135")]; tensor input_1149_cast_fp16 = add(x = input_1131_cast_fp16, y = input_1147_cast_fp16)[name = tensor("input_1149_cast_fp16")]; tensor input_1151_axes_0 = const()[name = tensor("input_1151_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1070975936)))]; tensor module_layers_21_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1070978048)))]; tensor input_1151_cast_fp16 = layer_norm(axes = input_1151_axes_0, beta = module_layers_21_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward2_weight_to_fp16, x = input_1149_cast_fp16)[name = tensor("input_1151_cast_fp16")]; tensor module_layers_21_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_21_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1070980160)))]; tensor linear_197_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_21_feed_forward2_linear1_weight_to_fp16, x = input_1151_cast_fp16)[name = tensor("linear_197_cast_fp16")]; tensor input_1155_cast_fp16 = silu(x = linear_197_cast_fp16)[name = tensor("input_1155_cast_fp16")]; tensor module_layers_21_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_21_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1079368832)))]; tensor linear_198_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_feed_forward2_linear2_weight_to_fp16, x = input_1155_cast_fp16)[name = tensor("linear_198_cast_fp16")]; tensor var_3789_to_fp16 = const()[name = tensor("op_3789_to_fp16"), val = tensor(0x1p-1)]; tensor var_3790_cast_fp16 = mul(x = linear_198_cast_fp16, y = var_3789_to_fp16)[name = tensor("op_3790_cast_fp16")]; tensor input_1161_cast_fp16 = add(x = input_1149_cast_fp16, y = var_3790_cast_fp16)[name = tensor("input_1161_cast_fp16")]; tensor input_1163_axes_0 = const()[name = tensor("input_1163_axes_0"), val = tensor([-1])]; tensor module_layers_21_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1087757504)))]; tensor module_layers_21_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1087759616)))]; tensor input_1163_cast_fp16 = layer_norm(axes = input_1163_axes_0, beta = module_layers_21_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_out_weight_to_fp16, x = input_1161_cast_fp16)[name = tensor("input_1163_cast_fp16")]; tensor input_1165_axes_0 = const()[name = tensor("input_1165_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1087761728)))]; tensor module_layers_22_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1087763840)))]; tensor input_1165_cast_fp16 = layer_norm(axes = input_1165_axes_0, beta = module_layers_22_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward1_weight_to_fp16, x = input_1163_cast_fp16)[name = tensor("input_1165_cast_fp16")]; tensor module_layers_22_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_22_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1087765952)))]; tensor linear_199_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_22_feed_forward1_linear1_weight_to_fp16, x = input_1165_cast_fp16)[name = tensor("linear_199_cast_fp16")]; tensor input_1169_cast_fp16 = silu(x = linear_199_cast_fp16)[name = tensor("input_1169_cast_fp16")]; tensor module_layers_22_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_22_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1096154624)))]; tensor linear_200_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_feed_forward1_linear2_weight_to_fp16, x = input_1169_cast_fp16)[name = tensor("linear_200_cast_fp16")]; tensor var_3818_to_fp16 = const()[name = tensor("op_3818_to_fp16"), val = tensor(0x1p-1)]; tensor var_3819_cast_fp16 = mul(x = linear_200_cast_fp16, y = var_3818_to_fp16)[name = tensor("op_3819_cast_fp16")]; tensor input_1175_cast_fp16 = add(x = input_1163_cast_fp16, y = var_3819_cast_fp16)[name = tensor("input_1175_cast_fp16")]; tensor query_45_axes_0 = const()[name = tensor("query_45_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1104543296)))]; tensor module_layers_22_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1104545408)))]; tensor query_45_cast_fp16 = layer_norm(axes = query_45_axes_0, beta = module_layers_22_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_self_att_weight_to_fp16, x = input_1175_cast_fp16)[name = tensor("query_45_cast_fp16")]; tensor module_layers_22_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1104547520)))]; tensor linear_201_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_q_weight_to_fp16, x = query_45_cast_fp16)[name = tensor("linear_201_cast_fp16")]; tensor var_3835 = const()[name = tensor("op_3835"), val = tensor([1, -1, 8, 128])]; tensor q_133_cast_fp16 = reshape(shape = var_3835, x = linear_201_cast_fp16)[name = tensor("q_133_cast_fp16")]; tensor module_layers_22_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1106644736)))]; tensor linear_202_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_k_weight_to_fp16, x = query_45_cast_fp16)[name = tensor("linear_202_cast_fp16")]; tensor var_3839 = const()[name = tensor("op_3839"), val = tensor([1, -1, 8, 128])]; tensor k_89_cast_fp16 = reshape(shape = var_3839, x = linear_202_cast_fp16)[name = tensor("k_89_cast_fp16")]; tensor module_layers_22_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1108741952)))]; tensor linear_203_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_v_weight_to_fp16, x = query_45_cast_fp16)[name = tensor("linear_203_cast_fp16")]; tensor var_3843 = const()[name = tensor("op_3843"), val = tensor([1, -1, 8, 128])]; tensor v_45_cast_fp16 = reshape(shape = var_3843, x = linear_203_cast_fp16)[name = tensor("v_45_cast_fp16")]; tensor value_47_perm_0 = const()[name = tensor("value_47_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_22_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1110839168)))]; tensor var_3855_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3855_cast_fp16")]; tensor module_layers_22_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1110841280)))]; tensor var_3857_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3857_cast_fp16")]; tensor q_with_bias_v_45_perm_0 = const()[name = tensor("q_with_bias_v_45_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_491_transpose_x_0 = const()[name = tensor("x_491_transpose_x_0"), val = tensor(false)]; tensor x_491_transpose_y_0 = const()[name = tensor("x_491_transpose_y_0"), val = tensor(false)]; tensor var_3859_to_fp16 = const()[name = tensor("op_3859_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1110843392)))]; tensor q_with_bias_v_45_cast_fp16 = transpose(perm = q_with_bias_v_45_perm_0, x = var_3857_cast_fp16)[name = tensor("transpose_133")]; tensor x_491_cast_fp16 = matmul(transpose_x = x_491_transpose_x_0, transpose_y = x_491_transpose_y_0, x = q_with_bias_v_45_cast_fp16, y = var_3859_to_fp16)[name = tensor("x_491_cast_fp16")]; tensor x_493_pad_0 = const()[name = tensor("x_493_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_493_mode_0 = const()[name = tensor("x_493_mode_0"), val = tensor("constant")]; tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(0x0p+0)]; tensor x_493_cast_fp16 = pad(constant_val = const_234_to_fp16, mode = x_493_mode_0, pad = x_493_pad_0, x = x_491_cast_fp16)[name = tensor("x_493_cast_fp16")]; tensor var_3867 = const()[name = tensor("op_3867"), val = tensor([1, 8, -1, 188])]; tensor x_495_cast_fp16 = reshape(shape = var_3867, x = x_493_cast_fp16)[name = tensor("x_495_cast_fp16")]; tensor var_3871_begin_0 = const()[name = tensor("op_3871_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3871_end_0 = const()[name = tensor("op_3871_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3871_end_mask_0 = const()[name = tensor("op_3871_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3871_cast_fp16 = slice_by_index(begin = var_3871_begin_0, end = var_3871_end_0, end_mask = var_3871_end_mask_0, x = x_495_cast_fp16)[name = tensor("op_3871_cast_fp16")]; tensor var_3872 = const()[name = tensor("op_3872"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_89_cast_fp16 = reshape(shape = var_3872, x = var_3871_cast_fp16)[name = tensor("matrix_bd_89_cast_fp16")]; tensor matrix_ac_45_transpose_x_0 = const()[name = tensor("matrix_ac_45_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_45_transpose_y_0 = const()[name = tensor("matrix_ac_45_transpose_y_0"), val = tensor(false)]; tensor transpose_116_perm_0 = const()[name = tensor("transpose_116_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_117_perm_0 = const()[name = tensor("transpose_117_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_117 = transpose(perm = transpose_117_perm_0, x = k_89_cast_fp16)[name = tensor("transpose_131")]; tensor transpose_116 = transpose(perm = transpose_116_perm_0, x = var_3855_cast_fp16)[name = tensor("transpose_132")]; tensor matrix_ac_45_cast_fp16 = matmul(transpose_x = matrix_ac_45_transpose_x_0, transpose_y = matrix_ac_45_transpose_y_0, x = transpose_116, y = transpose_117)[name = tensor("matrix_ac_45_cast_fp16")]; tensor matrix_bd_91_begin_0 = const()[name = tensor("matrix_bd_91_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_91_end_0 = const()[name = tensor("matrix_bd_91_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_91_end_mask_0 = const()[name = tensor("matrix_bd_91_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_91_cast_fp16 = slice_by_index(begin = matrix_bd_91_begin_0, end = matrix_bd_91_end_0, end_mask = matrix_bd_91_end_mask_0, x = matrix_bd_89_cast_fp16)[name = tensor("matrix_bd_91_cast_fp16")]; tensor var_3881_cast_fp16 = add(x = matrix_ac_45_cast_fp16, y = matrix_bd_91_cast_fp16)[name = tensor("op_3881_cast_fp16")]; tensor _inversed_scores_89_y_0_to_fp16 = const()[name = tensor("_inversed_scores_89_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_89_cast_fp16 = mul(x = var_3881_cast_fp16, y = _inversed_scores_89_y_0_to_fp16)[name = tensor("_inversed_scores_89_cast_fp16")]; tensor scores_91_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_89_cast_fp16, cond = mask_3)[name = tensor("scores_91_cast_fp16")]; tensor var_3887_cast_fp16 = softmax(axis = var_30, x = scores_91_cast_fp16)[name = tensor("op_3887_cast_fp16")]; tensor input_1177_cast_fp16 = select(a = var_11_to_fp16, b = var_3887_cast_fp16, cond = mask_3)[name = tensor("input_1177_cast_fp16")]; tensor x_497_transpose_x_0 = const()[name = tensor("x_497_transpose_x_0"), val = tensor(false)]; tensor x_497_transpose_y_0 = const()[name = tensor("x_497_transpose_y_0"), val = tensor(false)]; tensor value_47_cast_fp16 = transpose(perm = value_47_perm_0, x = v_45_cast_fp16)[name = tensor("transpose_134")]; tensor x_497_cast_fp16 = matmul(transpose_x = x_497_transpose_x_0, transpose_y = x_497_transpose_y_0, x = input_1177_cast_fp16, y = value_47_cast_fp16)[name = tensor("x_497_cast_fp16")]; tensor var_3891_perm_0 = const()[name = tensor("op_3891_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3892 = const()[name = tensor("op_3892"), val = tensor([1, -1, 1024])]; tensor var_3891_cast_fp16 = transpose(perm = var_3891_perm_0, x = x_497_cast_fp16)[name = tensor("transpose_130")]; tensor input_1179_cast_fp16 = reshape(shape = var_3892, x = var_3891_cast_fp16)[name = tensor("input_1179_cast_fp16")]; tensor module_layers_22_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_22_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1111611456)))]; tensor linear_205_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_out_weight_to_fp16, x = input_1179_cast_fp16)[name = tensor("linear_205_cast_fp16")]; tensor input_1183_cast_fp16 = add(x = input_1175_cast_fp16, y = linear_205_cast_fp16)[name = tensor("input_1183_cast_fp16")]; tensor x_501_axes_0 = const()[name = tensor("x_501_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1113708672)))]; tensor module_layers_22_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1113710784)))]; tensor x_501_cast_fp16 = layer_norm(axes = x_501_axes_0, beta = module_layers_22_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_conv_weight_to_fp16, x = input_1183_cast_fp16)[name = tensor("x_501_cast_fp16")]; tensor input_1185_perm_0 = const()[name = tensor("input_1185_perm_0"), val = tensor([0, 2, 1])]; tensor input_1187_pad_type_0 = const()[name = tensor("input_1187_pad_type_0"), val = tensor("valid")]; tensor input_1187_strides_0 = const()[name = tensor("input_1187_strides_0"), val = tensor([1])]; tensor input_1187_pad_0 = const()[name = tensor("input_1187_pad_0"), val = tensor([0, 0])]; tensor input_1187_dilations_0 = const()[name = tensor("input_1187_dilations_0"), val = tensor([1])]; tensor input_1187_groups_0 = const()[name = tensor("input_1187_groups_0"), val = tensor(1)]; tensor module_layers_22_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_22_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1113712896)))]; tensor input_1185_cast_fp16 = transpose(perm = input_1185_perm_0, x = x_501_cast_fp16)[name = tensor("transpose_129")]; tensor input_1187_cast_fp16 = conv(dilations = input_1187_dilations_0, groups = input_1187_groups_0, pad = input_1187_pad_0, pad_type = input_1187_pad_type_0, strides = input_1187_strides_0, weight = module_layers_22_conv_pointwise_conv1_weight_to_fp16, x = input_1185_cast_fp16)[name = tensor("input_1187_cast_fp16")]; tensor x_503_split_num_splits_0 = const()[name = tensor("x_503_split_num_splits_0"), val = tensor(2)]; tensor x_503_split_axis_0 = const()[name = tensor("x_503_split_axis_0"), val = tensor(1)]; tensor x_503_split_cast_fp16_0, tensor x_503_split_cast_fp16_1 = split(axis = x_503_split_axis_0, num_splits = x_503_split_num_splits_0, x = input_1187_cast_fp16)[name = tensor("x_503_split_cast_fp16")]; tensor x_503_split_1_sigmoid_cast_fp16 = sigmoid(x = x_503_split_cast_fp16_1)[name = tensor("x_503_split_1_sigmoid_cast_fp16")]; tensor x_503_cast_fp16 = mul(x = x_503_split_cast_fp16_0, y = x_503_split_1_sigmoid_cast_fp16)[name = tensor("x_503_cast_fp16")]; tensor input_1189_cast_fp16 = select(a = var_11_to_fp16, b = x_503_cast_fp16, cond = var_328)[name = tensor("input_1189_cast_fp16")]; tensor input_1191_pad_0 = const()[name = tensor("input_1191_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1191_mode_0 = const()[name = tensor("input_1191_mode_0"), val = tensor("constant")]; tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(0x0p+0)]; tensor input_1191_cast_fp16 = pad(constant_val = const_237_to_fp16, mode = input_1191_mode_0, pad = input_1191_pad_0, x = input_1189_cast_fp16)[name = tensor("input_1191_cast_fp16")]; tensor input_1193_pad_type_0 = const()[name = tensor("input_1193_pad_type_0"), val = tensor("valid")]; tensor input_1193_groups_0 = const()[name = tensor("input_1193_groups_0"), val = tensor(1024)]; tensor input_1193_strides_0 = const()[name = tensor("input_1193_strides_0"), val = tensor([1])]; tensor input_1193_pad_0 = const()[name = tensor("input_1193_pad_0"), val = tensor([0, 0])]; tensor input_1193_dilations_0 = const()[name = tensor("input_1193_dilations_0"), val = tensor([1])]; tensor const_292_to_fp16 = const()[name = tensor("const_292_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1117907264)))]; tensor const_293_to_fp16 = const()[name = tensor("const_293_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1117925760)))]; tensor input_1195_cast_fp16 = conv(bias = const_293_to_fp16, dilations = input_1193_dilations_0, groups = input_1193_groups_0, pad = input_1193_pad_0, pad_type = input_1193_pad_type_0, strides = input_1193_strides_0, weight = const_292_to_fp16, x = input_1191_cast_fp16)[name = tensor("input_1195_cast_fp16")]; tensor input_1197_cast_fp16 = silu(x = input_1195_cast_fp16)[name = tensor("input_1197_cast_fp16")]; tensor x_505_pad_type_0 = const()[name = tensor("x_505_pad_type_0"), val = tensor("valid")]; tensor x_505_strides_0 = const()[name = tensor("x_505_strides_0"), val = tensor([1])]; tensor x_505_pad_0 = const()[name = tensor("x_505_pad_0"), val = tensor([0, 0])]; tensor x_505_dilations_0 = const()[name = tensor("x_505_dilations_0"), val = tensor([1])]; tensor x_505_groups_0 = const()[name = tensor("x_505_groups_0"), val = tensor(1)]; tensor module_layers_22_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_22_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1117927872)))]; tensor x_505_cast_fp16 = conv(dilations = x_505_dilations_0, groups = x_505_groups_0, pad = x_505_pad_0, pad_type = x_505_pad_type_0, strides = x_505_strides_0, weight = module_layers_22_conv_pointwise_conv2_weight_to_fp16, x = input_1197_cast_fp16)[name = tensor("x_505_cast_fp16")]; tensor input_1199_perm_0 = const()[name = tensor("input_1199_perm_0"), val = tensor([0, 2, 1])]; tensor input_1199_cast_fp16 = transpose(perm = input_1199_perm_0, x = x_505_cast_fp16)[name = tensor("transpose_128")]; tensor input_1201_cast_fp16 = add(x = input_1183_cast_fp16, y = input_1199_cast_fp16)[name = tensor("input_1201_cast_fp16")]; tensor input_1203_axes_0 = const()[name = tensor("input_1203_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1120025088)))]; tensor module_layers_22_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1120027200)))]; tensor input_1203_cast_fp16 = layer_norm(axes = input_1203_axes_0, beta = module_layers_22_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward2_weight_to_fp16, x = input_1201_cast_fp16)[name = tensor("input_1203_cast_fp16")]; tensor module_layers_22_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_22_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1120029312)))]; tensor linear_206_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_22_feed_forward2_linear1_weight_to_fp16, x = input_1203_cast_fp16)[name = tensor("linear_206_cast_fp16")]; tensor input_1207_cast_fp16 = silu(x = linear_206_cast_fp16)[name = tensor("input_1207_cast_fp16")]; tensor module_layers_22_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_22_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1128417984)))]; tensor linear_207_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_feed_forward2_linear2_weight_to_fp16, x = input_1207_cast_fp16)[name = tensor("linear_207_cast_fp16")]; tensor var_3952_to_fp16 = const()[name = tensor("op_3952_to_fp16"), val = tensor(0x1p-1)]; tensor var_3953_cast_fp16 = mul(x = linear_207_cast_fp16, y = var_3952_to_fp16)[name = tensor("op_3953_cast_fp16")]; tensor input_1213_cast_fp16 = add(x = input_1201_cast_fp16, y = var_3953_cast_fp16)[name = tensor("input_1213_cast_fp16")]; tensor input_1215_axes_0 = const()[name = tensor("input_1215_axes_0"), val = tensor([-1])]; tensor module_layers_22_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136806656)))]; tensor module_layers_22_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136808768)))]; tensor input_1215_cast_fp16 = layer_norm(axes = input_1215_axes_0, beta = module_layers_22_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_out_weight_to_fp16, x = input_1213_cast_fp16)[name = tensor("input_1215_cast_fp16")]; tensor input_1217_axes_0 = const()[name = tensor("input_1217_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136810880)))]; tensor module_layers_23_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136812992)))]; tensor input_1217_cast_fp16 = layer_norm(axes = input_1217_axes_0, beta = module_layers_23_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward1_weight_to_fp16, x = input_1215_cast_fp16)[name = tensor("input_1217_cast_fp16")]; tensor module_layers_23_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_23_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136815104)))]; tensor linear_208_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_23_feed_forward1_linear1_weight_to_fp16, x = input_1217_cast_fp16)[name = tensor("linear_208_cast_fp16")]; tensor input_1221_cast_fp16 = silu(x = linear_208_cast_fp16)[name = tensor("input_1221_cast_fp16")]; tensor module_layers_23_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_23_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1145203776)))]; tensor linear_209_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_feed_forward1_linear2_weight_to_fp16, x = input_1221_cast_fp16)[name = tensor("linear_209_cast_fp16")]; tensor var_3981_to_fp16 = const()[name = tensor("op_3981_to_fp16"), val = tensor(0x1p-1)]; tensor var_3982_cast_fp16 = mul(x = linear_209_cast_fp16, y = var_3981_to_fp16)[name = tensor("op_3982_cast_fp16")]; tensor input_1227_cast_fp16 = add(x = input_1215_cast_fp16, y = var_3982_cast_fp16)[name = tensor("input_1227_cast_fp16")]; tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1153592448)))]; tensor module_layers_23_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1153594560)))]; tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_23_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_self_att_weight_to_fp16, x = input_1227_cast_fp16)[name = tensor("query_cast_fp16")]; tensor module_layers_23_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1153596672)))]; tensor linear_210_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_q_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_210_cast_fp16")]; tensor var_3998 = const()[name = tensor("op_3998"), val = tensor([1, -1, 8, 128])]; tensor q_139_cast_fp16 = reshape(shape = var_3998, x = linear_210_cast_fp16)[name = tensor("q_139_cast_fp16")]; tensor module_layers_23_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1155693888)))]; tensor linear_211_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_k_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_211_cast_fp16")]; tensor var_4002 = const()[name = tensor("op_4002"), val = tensor([1, -1, 8, 128])]; tensor k_93_cast_fp16 = reshape(shape = var_4002, x = linear_211_cast_fp16)[name = tensor("k_93_cast_fp16")]; tensor module_layers_23_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1157791104)))]; tensor linear_212_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_v_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_212_cast_fp16")]; tensor var_4006 = const()[name = tensor("op_4006"), val = tensor([1, -1, 8, 128])]; tensor v_cast_fp16 = reshape(shape = var_4006, x = linear_212_cast_fp16)[name = tensor("v_cast_fp16")]; tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; tensor module_layers_23_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1159888320)))]; tensor var_4018_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_u_to_fp16)[name = tensor("op_4018_cast_fp16")]; tensor module_layers_23_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1159890432)))]; tensor var_4020_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_v_to_fp16)[name = tensor("op_4020_cast_fp16")]; tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_513_transpose_x_0 = const()[name = tensor("x_513_transpose_x_0"), val = tensor(false)]; tensor x_513_transpose_y_0 = const()[name = tensor("x_513_transpose_y_0"), val = tensor(false)]; tensor var_4022_to_fp16 = const()[name = tensor("op_4022_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1159892544)))]; tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_4020_cast_fp16)[name = tensor("transpose_126")]; tensor x_513_cast_fp16 = matmul(transpose_x = x_513_transpose_x_0, transpose_y = x_513_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_4022_to_fp16)[name = tensor("x_513_cast_fp16")]; tensor x_515_pad_0 = const()[name = tensor("x_515_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_515_mode_0 = const()[name = tensor("x_515_mode_0"), val = tensor("constant")]; tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(0x0p+0)]; tensor x_515_cast_fp16 = pad(constant_val = const_244_to_fp16, mode = x_515_mode_0, pad = x_515_pad_0, x = x_513_cast_fp16)[name = tensor("x_515_cast_fp16")]; tensor var_4030 = const()[name = tensor("op_4030"), val = tensor([1, 8, -1, 188])]; tensor x_517_cast_fp16 = reshape(shape = var_4030, x = x_515_cast_fp16)[name = tensor("x_517_cast_fp16")]; tensor var_4034_begin_0 = const()[name = tensor("op_4034_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_4034_end_0 = const()[name = tensor("op_4034_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_4034_end_mask_0 = const()[name = tensor("op_4034_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_4034_cast_fp16 = slice_by_index(begin = var_4034_begin_0, end = var_4034_end_0, end_mask = var_4034_end_mask_0, x = x_517_cast_fp16)[name = tensor("op_4034_cast_fp16")]; tensor var_4035 = const()[name = tensor("op_4035"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_93_cast_fp16 = reshape(shape = var_4035, x = var_4034_cast_fp16)[name = tensor("matrix_bd_93_cast_fp16")]; tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; tensor transpose_118_perm_0 = const()[name = tensor("transpose_118_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_119_perm_0 = const()[name = tensor("transpose_119_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_119 = transpose(perm = transpose_119_perm_0, x = k_93_cast_fp16)[name = tensor("transpose_124")]; tensor transpose_118 = transpose(perm = transpose_118_perm_0, x = var_4018_cast_fp16)[name = tensor("transpose_125")]; tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_118, y = transpose_119)[name = tensor("matrix_ac_cast_fp16")]; tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_93_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; tensor var_4044_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_4044_cast_fp16")]; tensor _inversed_scores_93_y_0_to_fp16 = const()[name = tensor("_inversed_scores_93_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; tensor _inversed_scores_93_cast_fp16 = mul(x = var_4044_cast_fp16, y = _inversed_scores_93_y_0_to_fp16)[name = tensor("_inversed_scores_93_cast_fp16")]; tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_93_cast_fp16, cond = mask_3)[name = tensor("scores_cast_fp16")]; tensor var_4050_cast_fp16 = softmax(axis = var_30, x = scores_cast_fp16)[name = tensor("op_4050_cast_fp16")]; tensor input_1229_cast_fp16 = select(a = var_11_to_fp16, b = var_4050_cast_fp16, cond = mask_3)[name = tensor("input_1229_cast_fp16")]; tensor x_519_transpose_x_0 = const()[name = tensor("x_519_transpose_x_0"), val = tensor(false)]; tensor x_519_transpose_y_0 = const()[name = tensor("x_519_transpose_y_0"), val = tensor(false)]; tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_127")]; tensor x_519_cast_fp16 = matmul(transpose_x = x_519_transpose_x_0, transpose_y = x_519_transpose_y_0, x = input_1229_cast_fp16, y = value_cast_fp16)[name = tensor("x_519_cast_fp16")]; tensor var_4054_perm_0 = const()[name = tensor("op_4054_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4055 = const()[name = tensor("op_4055"), val = tensor([1, -1, 1024])]; tensor var_4054_cast_fp16 = transpose(perm = var_4054_perm_0, x = x_519_cast_fp16)[name = tensor("transpose_123")]; tensor input_1231_cast_fp16 = reshape(shape = var_4055, x = var_4054_cast_fp16)[name = tensor("input_1231_cast_fp16")]; tensor module_layers_23_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_23_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1160660608)))]; tensor linear_214_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_out_weight_to_fp16, x = input_1231_cast_fp16)[name = tensor("linear_214_cast_fp16")]; tensor input_1235_cast_fp16 = add(x = input_1227_cast_fp16, y = linear_214_cast_fp16)[name = tensor("input_1235_cast_fp16")]; tensor x_523_axes_0 = const()[name = tensor("x_523_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1162757824)))]; tensor module_layers_23_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1162759936)))]; tensor x_523_cast_fp16 = layer_norm(axes = x_523_axes_0, beta = module_layers_23_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_conv_weight_to_fp16, x = input_1235_cast_fp16)[name = tensor("x_523_cast_fp16")]; tensor input_1237_perm_0 = const()[name = tensor("input_1237_perm_0"), val = tensor([0, 2, 1])]; tensor input_1239_pad_type_0 = const()[name = tensor("input_1239_pad_type_0"), val = tensor("valid")]; tensor input_1239_strides_0 = const()[name = tensor("input_1239_strides_0"), val = tensor([1])]; tensor input_1239_pad_0 = const()[name = tensor("input_1239_pad_0"), val = tensor([0, 0])]; tensor input_1239_dilations_0 = const()[name = tensor("input_1239_dilations_0"), val = tensor([1])]; tensor input_1239_groups_0 = const()[name = tensor("input_1239_groups_0"), val = tensor(1)]; tensor module_layers_23_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_23_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1162762048)))]; tensor input_1237_cast_fp16 = transpose(perm = input_1237_perm_0, x = x_523_cast_fp16)[name = tensor("transpose_122")]; tensor input_1239_cast_fp16 = conv(dilations = input_1239_dilations_0, groups = input_1239_groups_0, pad = input_1239_pad_0, pad_type = input_1239_pad_type_0, strides = input_1239_strides_0, weight = module_layers_23_conv_pointwise_conv1_weight_to_fp16, x = input_1237_cast_fp16)[name = tensor("input_1239_cast_fp16")]; tensor x_525_split_num_splits_0 = const()[name = tensor("x_525_split_num_splits_0"), val = tensor(2)]; tensor x_525_split_axis_0 = const()[name = tensor("x_525_split_axis_0"), val = tensor(1)]; tensor x_525_split_cast_fp16_0, tensor x_525_split_cast_fp16_1 = split(axis = x_525_split_axis_0, num_splits = x_525_split_num_splits_0, x = input_1239_cast_fp16)[name = tensor("x_525_split_cast_fp16")]; tensor x_525_split_1_sigmoid_cast_fp16 = sigmoid(x = x_525_split_cast_fp16_1)[name = tensor("x_525_split_1_sigmoid_cast_fp16")]; tensor x_525_cast_fp16 = mul(x = x_525_split_cast_fp16_0, y = x_525_split_1_sigmoid_cast_fp16)[name = tensor("x_525_cast_fp16")]; tensor input_1241_cast_fp16 = select(a = var_11_to_fp16, b = x_525_cast_fp16, cond = var_328)[name = tensor("input_1241_cast_fp16")]; tensor input_1243_pad_0 = const()[name = tensor("input_1243_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1243_mode_0 = const()[name = tensor("input_1243_mode_0"), val = tensor("constant")]; tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(0x0p+0)]; tensor input_1243_cast_fp16 = pad(constant_val = const_247_to_fp16, mode = input_1243_mode_0, pad = input_1243_pad_0, x = input_1241_cast_fp16)[name = tensor("input_1243_cast_fp16")]; tensor input_1245_pad_type_0 = const()[name = tensor("input_1245_pad_type_0"), val = tensor("valid")]; tensor input_1245_groups_0 = const()[name = tensor("input_1245_groups_0"), val = tensor(1024)]; tensor input_1245_strides_0 = const()[name = tensor("input_1245_strides_0"), val = tensor([1])]; tensor input_1245_pad_0 = const()[name = tensor("input_1245_pad_0"), val = tensor([0, 0])]; tensor input_1245_dilations_0 = const()[name = tensor("input_1245_dilations_0"), val = tensor([1])]; tensor const_294_to_fp16 = const()[name = tensor("const_294_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1166956416)))]; tensor const_295_to_fp16 = const()[name = tensor("const_295_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1166974912)))]; tensor input_1247_cast_fp16 = conv(bias = const_295_to_fp16, dilations = input_1245_dilations_0, groups = input_1245_groups_0, pad = input_1245_pad_0, pad_type = input_1245_pad_type_0, strides = input_1245_strides_0, weight = const_294_to_fp16, x = input_1243_cast_fp16)[name = tensor("input_1247_cast_fp16")]; tensor input_1249_cast_fp16 = silu(x = input_1247_cast_fp16)[name = tensor("input_1249_cast_fp16")]; tensor x_527_pad_type_0 = const()[name = tensor("x_527_pad_type_0"), val = tensor("valid")]; tensor x_527_strides_0 = const()[name = tensor("x_527_strides_0"), val = tensor([1])]; tensor x_527_pad_0 = const()[name = tensor("x_527_pad_0"), val = tensor([0, 0])]; tensor x_527_dilations_0 = const()[name = tensor("x_527_dilations_0"), val = tensor([1])]; tensor x_527_groups_0 = const()[name = tensor("x_527_groups_0"), val = tensor(1)]; tensor module_layers_23_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_23_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1166977024)))]; tensor x_527_cast_fp16 = conv(dilations = x_527_dilations_0, groups = x_527_groups_0, pad = x_527_pad_0, pad_type = x_527_pad_type_0, strides = x_527_strides_0, weight = module_layers_23_conv_pointwise_conv2_weight_to_fp16, x = input_1249_cast_fp16)[name = tensor("x_527_cast_fp16")]; tensor input_1251_perm_0 = const()[name = tensor("input_1251_perm_0"), val = tensor([0, 2, 1])]; tensor input_1251_cast_fp16 = transpose(perm = input_1251_perm_0, x = x_527_cast_fp16)[name = tensor("transpose_121")]; tensor input_1253_cast_fp16 = add(x = input_1235_cast_fp16, y = input_1251_cast_fp16)[name = tensor("input_1253_cast_fp16")]; tensor input_1255_axes_0 = const()[name = tensor("input_1255_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1169074240)))]; tensor module_layers_23_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1169076352)))]; tensor input_1255_cast_fp16 = layer_norm(axes = input_1255_axes_0, beta = module_layers_23_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward2_weight_to_fp16, x = input_1253_cast_fp16)[name = tensor("input_1255_cast_fp16")]; tensor module_layers_23_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_23_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1169078464)))]; tensor linear_215_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_23_feed_forward2_linear1_weight_to_fp16, x = input_1255_cast_fp16)[name = tensor("linear_215_cast_fp16")]; tensor input_1259_cast_fp16 = silu(x = linear_215_cast_fp16)[name = tensor("input_1259_cast_fp16")]; tensor module_layers_23_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_23_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1177467136)))]; tensor linear_216_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_feed_forward2_linear2_weight_to_fp16, x = input_1259_cast_fp16)[name = tensor("linear_216_cast_fp16")]; tensor var_4115_to_fp16 = const()[name = tensor("op_4115_to_fp16"), val = tensor(0x1p-1)]; tensor var_4116_cast_fp16 = mul(x = linear_216_cast_fp16, y = var_4115_to_fp16)[name = tensor("op_4116_cast_fp16")]; tensor input_cast_fp16 = add(x = input_1253_cast_fp16, y = var_4116_cast_fp16)[name = tensor("input_cast_fp16")]; tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; tensor module_layers_23_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1185855808)))]; tensor module_layers_23_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1185857920)))]; tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_23_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_120")]; tensor encoder = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_227")]; } -> (encoder, encoder_length); }