program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}})] { func main(tensor mel_length, tensor melspectrogram_features) { tensor proj_bias = const()[name = tensor("proj_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor proj_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("proj_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1055104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1053952)))]; tensor encoder_pre_encode_conv_0_bias = const()[name = tensor("encoder_pre_encode_conv_0_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1059328)))]; tensor encoder_pre_encode_conv_0_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_pre_encode_conv_0_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1060416))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1063104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1062784)))]; tensor encoder_pre_encode_conv_2_bias = const()[name = tensor("encoder_pre_encode_conv_2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1064192)))]; tensor encoder_pre_encode_conv_2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_pre_encode_conv_2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1065280))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1067968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1067648)))]; tensor encoder_pre_encode_conv_3_bias = const()[name = tensor("encoder_pre_encode_conv_3_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1069056)))]; tensor encoder_pre_encode_conv_3_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_pre_encode_conv_3_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1070144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1135744)))]; tensor encoder_pre_encode_conv_5_bias = const()[name = tensor("encoder_pre_encode_conv_5_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1137152)))]; tensor encoder_pre_encode_conv_5_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_pre_encode_conv_5_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1138240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1140928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1140608)))]; tensor encoder_pre_encode_conv_6_bias = const()[name = tensor("encoder_pre_encode_conv_6_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1142016)))]; tensor encoder_pre_encode_conv_6_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_pre_encode_conv_6_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1143104))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1209024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1208704)))]; tensor encoder_pre_encode_out_bias = const()[name = tensor("encoder_pre_encode_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1210112)))]; tensor encoder_pre_encode_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_pre_encode_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1214272))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3836864))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3835776)))]; tensor encoder_layers_0_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_0_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3841024)))]; tensor encoder_layers_0_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_0_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3845184)))]; tensor encoder_layers_0_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3849344)))]; tensor encoder_layers_0_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3865792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8064320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8060160)))]; tensor encoder_layers_0_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8080768)))]; tensor encoder_layers_0_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8084928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12280384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12279296)))]; tensor encoder_layers_0_norm_self_att_bias = const()[name = tensor("encoder_layers_0_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12284544)))]; tensor encoder_layers_0_norm_self_att_weight = const()[name = tensor("encoder_layers_0_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12288704)))]; tensor encoder_layers_0_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12292864)))]; tensor encoder_layers_0_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12297024)))]; tensor encoder_layers_0_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12301184)))]; tensor encoder_layers_0_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12305344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13355072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13353984)))]; tensor encoder_layers_0_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13359232)))]; tensor encoder_layers_0_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13363392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14413120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14412032)))]; tensor encoder_layers_0_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14417280)))]; tensor encoder_layers_0_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14421440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15471168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15470080)))]; tensor encoder_layers_0_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15475328)))]; tensor encoder_layers_0_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15479488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16529216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16528128)))]; tensor encoder_layers_0_norm_conv_bias = const()[name = tensor("encoder_layers_0_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16533376)))]; tensor encoder_layers_0_norm_conv_weight = const()[name = tensor("encoder_layers_0_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16537536)))]; tensor encoder_layers_0_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16541696)))]; tensor encoder_layers_0_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16549952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18649280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18647168)))]; tensor encoder_layers_0_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18657536)))]; tensor encoder_layers_0_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18661696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19711424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19710336)))]; tensor encoder_layers_0_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_0_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19715584)))]; tensor encoder_layers_0_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_0_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19719744)))]; tensor encoder_layers_0_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19723904)))]; tensor encoder_layers_0_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19740352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23938880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23934720)))]; tensor encoder_layers_0_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23955328)))]; tensor encoder_layers_0_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23959488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28154944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28153856)))]; tensor encoder_layers_0_norm_out_bias = const()[name = tensor("encoder_layers_0_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28159104)))]; tensor encoder_layers_0_norm_out_weight = const()[name = tensor("encoder_layers_0_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28163264)))]; tensor encoder_layers_1_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_1_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28167424)))]; tensor encoder_layers_1_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_1_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28171584)))]; tensor encoder_layers_1_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28175744)))]; tensor encoder_layers_1_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28192192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32390720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32386560)))]; tensor encoder_layers_1_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32407168)))]; tensor encoder_layers_1_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32411328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36606784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36605696)))]; tensor encoder_layers_1_norm_self_att_bias = const()[name = tensor("encoder_layers_1_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36610944)))]; tensor encoder_layers_1_norm_self_att_weight = const()[name = tensor("encoder_layers_1_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36615104)))]; tensor encoder_layers_1_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36619264)))]; tensor encoder_layers_1_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36623424)))]; tensor encoder_layers_1_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36627584)))]; tensor encoder_layers_1_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36631744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37681472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37680384)))]; tensor encoder_layers_1_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37685632)))]; tensor encoder_layers_1_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37689792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38739520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38738432)))]; tensor encoder_layers_1_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38743680)))]; tensor encoder_layers_1_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38747840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39797568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39796480)))]; tensor encoder_layers_1_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39801728)))]; tensor encoder_layers_1_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39805888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40855616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40854528)))]; tensor encoder_layers_1_norm_conv_bias = const()[name = tensor("encoder_layers_1_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40859776)))]; tensor encoder_layers_1_norm_conv_weight = const()[name = tensor("encoder_layers_1_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40863936)))]; tensor encoder_layers_1_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40868096)))]; tensor encoder_layers_1_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40876352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42975680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42973568)))]; tensor encoder_layers_1_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42983936)))]; tensor encoder_layers_1_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42988096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44037824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44036736)))]; tensor encoder_layers_1_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_1_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44041984)))]; tensor encoder_layers_1_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_1_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44046144)))]; tensor encoder_layers_1_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44050304)))]; tensor encoder_layers_1_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44066752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48265280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48261120)))]; tensor encoder_layers_1_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48281728)))]; tensor encoder_layers_1_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48285888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52481344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52480256)))]; tensor encoder_layers_1_norm_out_bias = const()[name = tensor("encoder_layers_1_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52485504)))]; tensor encoder_layers_1_norm_out_weight = const()[name = tensor("encoder_layers_1_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52489664)))]; tensor encoder_layers_2_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_2_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52493824)))]; tensor encoder_layers_2_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_2_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52497984)))]; tensor encoder_layers_2_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52502144)))]; tensor encoder_layers_2_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52518592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56717120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56712960)))]; tensor encoder_layers_2_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56733568)))]; tensor encoder_layers_2_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56737728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60933184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60932096)))]; tensor encoder_layers_2_norm_self_att_bias = const()[name = tensor("encoder_layers_2_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60937344)))]; tensor encoder_layers_2_norm_self_att_weight = const()[name = tensor("encoder_layers_2_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60941504)))]; tensor encoder_layers_2_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60945664)))]; tensor encoder_layers_2_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60949824)))]; tensor encoder_layers_2_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60953984)))]; tensor encoder_layers_2_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60958144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62007872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62006784)))]; tensor encoder_layers_2_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62012032)))]; tensor encoder_layers_2_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62016192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63065920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63064832)))]; tensor encoder_layers_2_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63070080)))]; tensor encoder_layers_2_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63074240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64123968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64122880)))]; tensor encoder_layers_2_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64128128)))]; tensor encoder_layers_2_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64132288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65182016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65180928)))]; tensor encoder_layers_2_norm_conv_bias = const()[name = tensor("encoder_layers_2_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65186176)))]; tensor encoder_layers_2_norm_conv_weight = const()[name = tensor("encoder_layers_2_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65190336)))]; tensor encoder_layers_2_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65194496)))]; tensor encoder_layers_2_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65202752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67302080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67299968)))]; tensor encoder_layers_2_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67310336)))]; tensor encoder_layers_2_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67314496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68364224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68363136)))]; tensor encoder_layers_2_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_2_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68368384)))]; tensor encoder_layers_2_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_2_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68372544)))]; tensor encoder_layers_2_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68376704)))]; tensor encoder_layers_2_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68393152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72591680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72587520)))]; tensor encoder_layers_2_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72608128)))]; tensor encoder_layers_2_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72612288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76807744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76806656)))]; tensor encoder_layers_2_norm_out_bias = const()[name = tensor("encoder_layers_2_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76811904)))]; tensor encoder_layers_2_norm_out_weight = const()[name = tensor("encoder_layers_2_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76816064)))]; tensor encoder_layers_3_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_3_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76820224)))]; tensor encoder_layers_3_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_3_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76824384)))]; tensor encoder_layers_3_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76828544)))]; tensor encoder_layers_3_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76844992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81043520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81039360)))]; tensor encoder_layers_3_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81059968)))]; tensor encoder_layers_3_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81064128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85259584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85258496)))]; tensor encoder_layers_3_norm_self_att_bias = const()[name = tensor("encoder_layers_3_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85263744)))]; tensor encoder_layers_3_norm_self_att_weight = const()[name = tensor("encoder_layers_3_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85267904)))]; tensor encoder_layers_3_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85272064)))]; tensor encoder_layers_3_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85276224)))]; tensor encoder_layers_3_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85280384)))]; tensor encoder_layers_3_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85284544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86334272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86333184)))]; tensor encoder_layers_3_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86338432)))]; tensor encoder_layers_3_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86342592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87392320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87391232)))]; tensor encoder_layers_3_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87396480)))]; tensor encoder_layers_3_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87400640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88450368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88449280)))]; tensor encoder_layers_3_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88454528)))]; tensor encoder_layers_3_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88458688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89508416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89507328)))]; tensor encoder_layers_3_norm_conv_bias = const()[name = tensor("encoder_layers_3_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89512576)))]; tensor encoder_layers_3_norm_conv_weight = const()[name = tensor("encoder_layers_3_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89516736)))]; tensor encoder_layers_3_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89520896)))]; tensor encoder_layers_3_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89529152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91628480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91626368)))]; tensor encoder_layers_3_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91636736)))]; tensor encoder_layers_3_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91640896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92690624))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92689536)))]; tensor encoder_layers_3_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_3_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92694784)))]; tensor encoder_layers_3_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_3_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92698944)))]; tensor encoder_layers_3_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92703104)))]; tensor encoder_layers_3_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92719552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96918080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96913920)))]; tensor encoder_layers_3_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96934528)))]; tensor encoder_layers_3_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96938688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101134144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101133056)))]; tensor encoder_layers_3_norm_out_bias = const()[name = tensor("encoder_layers_3_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101138304)))]; tensor encoder_layers_3_norm_out_weight = const()[name = tensor("encoder_layers_3_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101142464)))]; tensor encoder_layers_4_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_4_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101146624)))]; tensor encoder_layers_4_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_4_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101150784)))]; tensor encoder_layers_4_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101154944)))]; tensor encoder_layers_4_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101171392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105369920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105365760)))]; tensor encoder_layers_4_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105386368)))]; tensor encoder_layers_4_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105390528))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109585984))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109584896)))]; tensor encoder_layers_4_norm_self_att_bias = const()[name = tensor("encoder_layers_4_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109590144)))]; tensor encoder_layers_4_norm_self_att_weight = const()[name = tensor("encoder_layers_4_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109594304)))]; tensor encoder_layers_4_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109598464)))]; tensor encoder_layers_4_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109602624)))]; tensor encoder_layers_4_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109606784)))]; tensor encoder_layers_4_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109610944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110660672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110659584)))]; tensor encoder_layers_4_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110664832)))]; tensor encoder_layers_4_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110668992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111718720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111717632)))]; tensor encoder_layers_4_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111722880)))]; tensor encoder_layers_4_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111727040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112776768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112775680)))]; tensor encoder_layers_4_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112780928)))]; tensor encoder_layers_4_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112785088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113834816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113833728)))]; tensor encoder_layers_4_norm_conv_bias = const()[name = tensor("encoder_layers_4_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113838976)))]; tensor encoder_layers_4_norm_conv_weight = const()[name = tensor("encoder_layers_4_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113843136)))]; tensor encoder_layers_4_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113847296)))]; tensor encoder_layers_4_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113855552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115954880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115952768)))]; tensor encoder_layers_4_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115963136)))]; tensor encoder_layers_4_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115967296))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117017024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117015936)))]; tensor encoder_layers_4_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_4_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117021184)))]; tensor encoder_layers_4_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_4_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117025344)))]; tensor encoder_layers_4_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117029504)))]; tensor encoder_layers_4_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117045952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121244480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121240320)))]; tensor encoder_layers_4_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121260928)))]; tensor encoder_layers_4_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121265088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125460544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125459456)))]; tensor encoder_layers_4_norm_out_bias = const()[name = tensor("encoder_layers_4_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125464704)))]; tensor encoder_layers_4_norm_out_weight = const()[name = tensor("encoder_layers_4_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125468864)))]; tensor encoder_layers_5_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_5_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125473024)))]; tensor encoder_layers_5_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_5_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125477184)))]; tensor encoder_layers_5_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125481344)))]; tensor encoder_layers_5_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125497792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129696320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129692160)))]; tensor encoder_layers_5_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129712768)))]; tensor encoder_layers_5_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129716928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133912384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133911296)))]; tensor encoder_layers_5_norm_self_att_bias = const()[name = tensor("encoder_layers_5_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133916544)))]; tensor encoder_layers_5_norm_self_att_weight = const()[name = tensor("encoder_layers_5_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133920704)))]; tensor encoder_layers_5_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133924864)))]; tensor encoder_layers_5_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133929024)))]; tensor encoder_layers_5_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133933184)))]; tensor encoder_layers_5_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133937344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134987072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134985984)))]; tensor encoder_layers_5_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134991232)))]; tensor encoder_layers_5_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134995392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136045120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136044032)))]; tensor encoder_layers_5_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136049280)))]; tensor encoder_layers_5_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136053440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137103168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137102080)))]; tensor encoder_layers_5_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137107328)))]; tensor encoder_layers_5_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137111488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138161216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138160128)))]; tensor encoder_layers_5_norm_conv_bias = const()[name = tensor("encoder_layers_5_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138165376)))]; tensor encoder_layers_5_norm_conv_weight = const()[name = tensor("encoder_layers_5_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138169536)))]; tensor encoder_layers_5_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138173696)))]; tensor encoder_layers_5_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138181952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140281280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140279168)))]; tensor encoder_layers_5_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140289536)))]; tensor encoder_layers_5_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140293696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141343424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141342336)))]; tensor encoder_layers_5_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_5_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141347584)))]; tensor encoder_layers_5_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_5_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141351744)))]; tensor encoder_layers_5_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141355904)))]; tensor encoder_layers_5_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141372352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145570880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145566720)))]; tensor encoder_layers_5_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145587328)))]; tensor encoder_layers_5_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145591488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149786944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149785856)))]; tensor encoder_layers_5_norm_out_bias = const()[name = tensor("encoder_layers_5_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149791104)))]; tensor encoder_layers_5_norm_out_weight = const()[name = tensor("encoder_layers_5_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149795264)))]; tensor encoder_layers_6_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_6_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149799424)))]; tensor encoder_layers_6_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_6_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149803584)))]; tensor encoder_layers_6_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149807744)))]; tensor encoder_layers_6_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149824192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154022720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154018560)))]; tensor encoder_layers_6_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154039168)))]; tensor encoder_layers_6_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154043328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158238784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158237696)))]; tensor encoder_layers_6_norm_self_att_bias = const()[name = tensor("encoder_layers_6_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158242944)))]; tensor encoder_layers_6_norm_self_att_weight = const()[name = tensor("encoder_layers_6_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158247104)))]; tensor encoder_layers_6_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158251264)))]; tensor encoder_layers_6_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158255424)))]; tensor encoder_layers_6_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158259584)))]; tensor encoder_layers_6_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158263744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159313472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159312384)))]; tensor encoder_layers_6_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159317632)))]; tensor encoder_layers_6_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159321792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160371520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160370432)))]; tensor encoder_layers_6_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160375680)))]; tensor encoder_layers_6_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160379840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161429568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161428480)))]; tensor encoder_layers_6_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161433728)))]; tensor encoder_layers_6_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161437888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162487616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162486528)))]; tensor encoder_layers_6_norm_conv_bias = const()[name = tensor("encoder_layers_6_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162491776)))]; tensor encoder_layers_6_norm_conv_weight = const()[name = tensor("encoder_layers_6_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162495936)))]; tensor encoder_layers_6_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162500096)))]; tensor encoder_layers_6_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162508352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164607680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164605568)))]; tensor encoder_layers_6_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164615936)))]; tensor encoder_layers_6_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164620096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165669824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165668736)))]; tensor encoder_layers_6_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_6_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165673984)))]; tensor encoder_layers_6_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_6_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165678144)))]; tensor encoder_layers_6_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165682304)))]; tensor encoder_layers_6_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165698752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169897280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169893120)))]; tensor encoder_layers_6_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169913728)))]; tensor encoder_layers_6_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_6_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169917888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174113344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174112256)))]; tensor encoder_layers_6_norm_out_bias = const()[name = tensor("encoder_layers_6_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174117504)))]; tensor encoder_layers_6_norm_out_weight = const()[name = tensor("encoder_layers_6_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174121664)))]; tensor encoder_layers_7_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_7_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174125824)))]; tensor encoder_layers_7_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_7_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174129984)))]; tensor encoder_layers_7_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174134144)))]; tensor encoder_layers_7_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174150592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178349120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178344960)))]; tensor encoder_layers_7_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178365568)))]; tensor encoder_layers_7_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178369728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182565184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182564096)))]; tensor encoder_layers_7_norm_self_att_bias = const()[name = tensor("encoder_layers_7_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182569344)))]; tensor encoder_layers_7_norm_self_att_weight = const()[name = tensor("encoder_layers_7_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182573504)))]; tensor encoder_layers_7_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182577664)))]; tensor encoder_layers_7_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182581824)))]; tensor encoder_layers_7_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182585984)))]; tensor encoder_layers_7_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182590144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183639872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183638784)))]; tensor encoder_layers_7_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183644032)))]; tensor encoder_layers_7_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183648192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184697920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184696832)))]; tensor encoder_layers_7_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184702080)))]; tensor encoder_layers_7_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184706240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185755968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185754880)))]; tensor encoder_layers_7_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185760128)))]; tensor encoder_layers_7_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185764288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186814016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186812928)))]; tensor encoder_layers_7_norm_conv_bias = const()[name = tensor("encoder_layers_7_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186818176)))]; tensor encoder_layers_7_norm_conv_weight = const()[name = tensor("encoder_layers_7_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186822336)))]; tensor encoder_layers_7_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186826496)))]; tensor encoder_layers_7_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186834752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188934080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188931968)))]; tensor encoder_layers_7_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188942336)))]; tensor encoder_layers_7_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188946496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189996224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189995136)))]; tensor encoder_layers_7_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_7_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190000384)))]; tensor encoder_layers_7_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_7_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190004544)))]; tensor encoder_layers_7_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190008704)))]; tensor encoder_layers_7_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190025152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194223680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194219520)))]; tensor encoder_layers_7_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194240128)))]; tensor encoder_layers_7_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_7_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194244288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198439744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198438656)))]; tensor encoder_layers_7_norm_out_bias = const()[name = tensor("encoder_layers_7_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198443904)))]; tensor encoder_layers_7_norm_out_weight = const()[name = tensor("encoder_layers_7_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198448064)))]; tensor encoder_layers_8_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_8_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198452224)))]; tensor encoder_layers_8_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_8_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198456384)))]; tensor encoder_layers_8_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198460544)))]; tensor encoder_layers_8_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198476992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202675520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202671360)))]; tensor encoder_layers_8_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202691968)))]; tensor encoder_layers_8_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202696128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206891584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206890496)))]; tensor encoder_layers_8_norm_self_att_bias = const()[name = tensor("encoder_layers_8_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206895744)))]; tensor encoder_layers_8_norm_self_att_weight = const()[name = tensor("encoder_layers_8_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206899904)))]; tensor encoder_layers_8_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206904064)))]; tensor encoder_layers_8_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206908224)))]; tensor encoder_layers_8_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206912384)))]; tensor encoder_layers_8_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206916544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207966272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207965184)))]; tensor encoder_layers_8_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207970432)))]; tensor encoder_layers_8_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207974592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209024320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209023232)))]; tensor encoder_layers_8_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209028480)))]; tensor encoder_layers_8_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209032640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210082368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210081280)))]; tensor encoder_layers_8_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210086528)))]; tensor encoder_layers_8_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210090688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211140416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211139328)))]; tensor encoder_layers_8_norm_conv_bias = const()[name = tensor("encoder_layers_8_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211144576)))]; tensor encoder_layers_8_norm_conv_weight = const()[name = tensor("encoder_layers_8_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211148736)))]; tensor encoder_layers_8_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211152896)))]; tensor encoder_layers_8_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211161152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213260480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213258368)))]; tensor encoder_layers_8_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213268736)))]; tensor encoder_layers_8_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213272896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214322624))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214321536)))]; tensor encoder_layers_8_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_8_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214326784)))]; tensor encoder_layers_8_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_8_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214330944)))]; tensor encoder_layers_8_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214335104)))]; tensor encoder_layers_8_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214351552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218550080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218545920)))]; tensor encoder_layers_8_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218566528)))]; tensor encoder_layers_8_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_8_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218570688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222766144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222765056)))]; tensor encoder_layers_8_norm_out_bias = const()[name = tensor("encoder_layers_8_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222770304)))]; tensor encoder_layers_8_norm_out_weight = const()[name = tensor("encoder_layers_8_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222774464)))]; tensor encoder_layers_9_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_9_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222778624)))]; tensor encoder_layers_9_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_9_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222782784)))]; tensor encoder_layers_9_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222786944)))]; tensor encoder_layers_9_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(222803392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227001920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(226997760)))]; tensor encoder_layers_9_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227018368)))]; tensor encoder_layers_9_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227022528))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231217984))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231216896)))]; tensor encoder_layers_9_norm_self_att_bias = const()[name = tensor("encoder_layers_9_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231222144)))]; tensor encoder_layers_9_norm_self_att_weight = const()[name = tensor("encoder_layers_9_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231226304)))]; tensor encoder_layers_9_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231230464)))]; tensor encoder_layers_9_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231234624)))]; tensor encoder_layers_9_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231238784)))]; tensor encoder_layers_9_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231242944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232292672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232291584)))]; tensor encoder_layers_9_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232296832)))]; tensor encoder_layers_9_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232300992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233350720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233349632)))]; tensor encoder_layers_9_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233354880)))]; tensor encoder_layers_9_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233359040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234408768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234407680)))]; tensor encoder_layers_9_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234412928)))]; tensor encoder_layers_9_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234417088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235466816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235465728)))]; tensor encoder_layers_9_norm_conv_bias = const()[name = tensor("encoder_layers_9_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235470976)))]; tensor encoder_layers_9_norm_conv_weight = const()[name = tensor("encoder_layers_9_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235475136)))]; tensor encoder_layers_9_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235479296)))]; tensor encoder_layers_9_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235487552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237586880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237584768)))]; tensor encoder_layers_9_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237595136)))]; tensor encoder_layers_9_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237599296))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238649024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238647936)))]; tensor encoder_layers_9_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_9_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238653184)))]; tensor encoder_layers_9_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_9_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238657344)))]; tensor encoder_layers_9_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238661504)))]; tensor encoder_layers_9_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238677952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242876480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242872320)))]; tensor encoder_layers_9_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242892928)))]; tensor encoder_layers_9_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_9_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242897088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247092544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247091456)))]; tensor encoder_layers_9_norm_out_bias = const()[name = tensor("encoder_layers_9_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247096704)))]; tensor encoder_layers_9_norm_out_weight = const()[name = tensor("encoder_layers_9_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247100864)))]; tensor encoder_layers_10_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_10_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247105024)))]; tensor encoder_layers_10_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_10_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247109184)))]; tensor encoder_layers_10_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247113344)))]; tensor encoder_layers_10_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247129792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251328320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251324160)))]; tensor encoder_layers_10_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251344768)))]; tensor encoder_layers_10_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251348928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255544384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255543296)))]; tensor encoder_layers_10_norm_self_att_bias = const()[name = tensor("encoder_layers_10_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255548544)))]; tensor encoder_layers_10_norm_self_att_weight = const()[name = tensor("encoder_layers_10_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255552704)))]; tensor encoder_layers_10_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255556864)))]; tensor encoder_layers_10_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255561024)))]; tensor encoder_layers_10_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255565184)))]; tensor encoder_layers_10_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255569344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256619072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256617984)))]; tensor encoder_layers_10_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256623232)))]; tensor encoder_layers_10_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256627392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257677120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257676032)))]; tensor encoder_layers_10_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257681280)))]; tensor encoder_layers_10_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257685440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258735168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258734080)))]; tensor encoder_layers_10_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258739328)))]; tensor encoder_layers_10_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258743488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259793216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259792128)))]; tensor encoder_layers_10_norm_conv_bias = const()[name = tensor("encoder_layers_10_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259797376)))]; tensor encoder_layers_10_norm_conv_weight = const()[name = tensor("encoder_layers_10_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259801536)))]; tensor encoder_layers_10_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259805696)))]; tensor encoder_layers_10_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259813952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261913280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261911168)))]; tensor encoder_layers_10_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261921536)))]; tensor encoder_layers_10_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261925696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262975424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262974336)))]; tensor encoder_layers_10_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_10_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262979584)))]; tensor encoder_layers_10_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_10_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262983744)))]; tensor encoder_layers_10_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262987904)))]; tensor encoder_layers_10_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263004352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267202880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267198720)))]; tensor encoder_layers_10_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267219328)))]; tensor encoder_layers_10_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_10_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267223488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271418944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271417856)))]; tensor encoder_layers_10_norm_out_bias = const()[name = tensor("encoder_layers_10_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271423104)))]; tensor encoder_layers_10_norm_out_weight = const()[name = tensor("encoder_layers_10_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271427264)))]; tensor encoder_layers_11_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_11_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271431424)))]; tensor encoder_layers_11_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_11_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271435584)))]; tensor encoder_layers_11_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271439744)))]; tensor encoder_layers_11_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271456192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275654720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275650560)))]; tensor encoder_layers_11_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275671168)))]; tensor encoder_layers_11_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275675328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279870784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279869696)))]; tensor encoder_layers_11_norm_self_att_bias = const()[name = tensor("encoder_layers_11_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279874944)))]; tensor encoder_layers_11_norm_self_att_weight = const()[name = tensor("encoder_layers_11_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279879104)))]; tensor encoder_layers_11_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279883264)))]; tensor encoder_layers_11_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279887424)))]; tensor encoder_layers_11_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279891584)))]; tensor encoder_layers_11_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279895744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280945472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280944384)))]; tensor encoder_layers_11_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280949632)))]; tensor encoder_layers_11_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280953792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282003520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282002432)))]; tensor encoder_layers_11_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282007680)))]; tensor encoder_layers_11_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282011840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283061568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283060480)))]; tensor encoder_layers_11_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283065728)))]; tensor encoder_layers_11_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283069888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284119616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284118528)))]; tensor encoder_layers_11_norm_conv_bias = const()[name = tensor("encoder_layers_11_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284123776)))]; tensor encoder_layers_11_norm_conv_weight = const()[name = tensor("encoder_layers_11_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284127936)))]; tensor encoder_layers_11_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284132096)))]; tensor encoder_layers_11_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284140352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286239680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286237568)))]; tensor encoder_layers_11_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286247936)))]; tensor encoder_layers_11_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286252096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287301824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287300736)))]; tensor encoder_layers_11_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_11_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287305984)))]; tensor encoder_layers_11_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_11_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287310144)))]; tensor encoder_layers_11_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287314304)))]; tensor encoder_layers_11_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287330752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291529280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291525120)))]; tensor encoder_layers_11_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291545728)))]; tensor encoder_layers_11_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_11_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291549888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295745344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295744256)))]; tensor encoder_layers_11_norm_out_bias = const()[name = tensor("encoder_layers_11_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295749504)))]; tensor encoder_layers_11_norm_out_weight = const()[name = tensor("encoder_layers_11_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295753664)))]; tensor encoder_layers_12_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_12_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295757824)))]; tensor encoder_layers_12_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_12_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295761984)))]; tensor encoder_layers_12_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295766144)))]; tensor encoder_layers_12_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295782592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(299981120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(299976960)))]; tensor encoder_layers_12_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(299997568)))]; tensor encoder_layers_12_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(300001728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304197184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304196096)))]; tensor encoder_layers_12_norm_self_att_bias = const()[name = tensor("encoder_layers_12_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304201344)))]; tensor encoder_layers_12_norm_self_att_weight = const()[name = tensor("encoder_layers_12_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304205504)))]; tensor encoder_layers_12_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304209664)))]; tensor encoder_layers_12_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304213824)))]; tensor encoder_layers_12_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304217984)))]; tensor encoder_layers_12_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304222144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305271872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305270784)))]; tensor encoder_layers_12_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305276032)))]; tensor encoder_layers_12_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305280192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306329920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306328832)))]; tensor encoder_layers_12_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306334080)))]; tensor encoder_layers_12_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306338240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307387968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307386880)))]; tensor encoder_layers_12_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307392128)))]; tensor encoder_layers_12_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307396288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308446016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308444928)))]; tensor encoder_layers_12_norm_conv_bias = const()[name = tensor("encoder_layers_12_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308450176)))]; tensor encoder_layers_12_norm_conv_weight = const()[name = tensor("encoder_layers_12_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308454336)))]; tensor encoder_layers_12_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308458496)))]; tensor encoder_layers_12_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308466752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310566080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310563968)))]; tensor encoder_layers_12_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310574336)))]; tensor encoder_layers_12_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310578496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311628224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311627136)))]; tensor encoder_layers_12_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_12_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311632384)))]; tensor encoder_layers_12_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_12_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311636544)))]; tensor encoder_layers_12_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311640704)))]; tensor encoder_layers_12_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311657152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(315855680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(315851520)))]; tensor encoder_layers_12_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(315872128)))]; tensor encoder_layers_12_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_12_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(315876288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320071744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320070656)))]; tensor encoder_layers_12_norm_out_bias = const()[name = tensor("encoder_layers_12_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320075904)))]; tensor encoder_layers_12_norm_out_weight = const()[name = tensor("encoder_layers_12_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320080064)))]; tensor encoder_layers_13_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_13_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320084224)))]; tensor encoder_layers_13_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_13_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320088384)))]; tensor encoder_layers_13_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320092544)))]; tensor encoder_layers_13_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320108992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324307520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324303360)))]; tensor encoder_layers_13_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324323968)))]; tensor encoder_layers_13_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324328128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328523584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328522496)))]; tensor encoder_layers_13_norm_self_att_bias = const()[name = tensor("encoder_layers_13_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328527744)))]; tensor encoder_layers_13_norm_self_att_weight = const()[name = tensor("encoder_layers_13_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328531904)))]; tensor encoder_layers_13_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328536064)))]; tensor encoder_layers_13_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328540224)))]; tensor encoder_layers_13_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328544384)))]; tensor encoder_layers_13_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328548544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(329598272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(329597184)))]; tensor encoder_layers_13_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(329602432)))]; tensor encoder_layers_13_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(329606592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330656320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330655232)))]; tensor encoder_layers_13_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330660480)))]; tensor encoder_layers_13_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330664640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331714368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331713280)))]; tensor encoder_layers_13_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331718528)))]; tensor encoder_layers_13_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331722688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332772416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332771328)))]; tensor encoder_layers_13_norm_conv_bias = const()[name = tensor("encoder_layers_13_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332776576)))]; tensor encoder_layers_13_norm_conv_weight = const()[name = tensor("encoder_layers_13_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332780736)))]; tensor encoder_layers_13_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332784896)))]; tensor encoder_layers_13_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332793152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334892480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334890368)))]; tensor encoder_layers_13_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334900736)))]; tensor encoder_layers_13_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334904896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335954624))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335953536)))]; tensor encoder_layers_13_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_13_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335958784)))]; tensor encoder_layers_13_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_13_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335962944)))]; tensor encoder_layers_13_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335967104)))]; tensor encoder_layers_13_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335983552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(340182080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(340177920)))]; tensor encoder_layers_13_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(340198528)))]; tensor encoder_layers_13_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_13_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(340202688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344398144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344397056)))]; tensor encoder_layers_13_norm_out_bias = const()[name = tensor("encoder_layers_13_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344402304)))]; tensor encoder_layers_13_norm_out_weight = const()[name = tensor("encoder_layers_13_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344406464)))]; tensor encoder_layers_14_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_14_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344410624)))]; tensor encoder_layers_14_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_14_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344414784)))]; tensor encoder_layers_14_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344418944)))]; tensor encoder_layers_14_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344435392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348633920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348629760)))]; tensor encoder_layers_14_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348650368)))]; tensor encoder_layers_14_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348654528))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352849984))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352848896)))]; tensor encoder_layers_14_norm_self_att_bias = const()[name = tensor("encoder_layers_14_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352854144)))]; tensor encoder_layers_14_norm_self_att_weight = const()[name = tensor("encoder_layers_14_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352858304)))]; tensor encoder_layers_14_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352862464)))]; tensor encoder_layers_14_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352866624)))]; tensor encoder_layers_14_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352870784)))]; tensor encoder_layers_14_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352874944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353924672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353923584)))]; tensor encoder_layers_14_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353928832)))]; tensor encoder_layers_14_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353932992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(354982720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(354981632)))]; tensor encoder_layers_14_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(354986880)))]; tensor encoder_layers_14_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(354991040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356040768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356039680)))]; tensor encoder_layers_14_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356044928)))]; tensor encoder_layers_14_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356049088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357098816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357097728)))]; tensor encoder_layers_14_norm_conv_bias = const()[name = tensor("encoder_layers_14_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357102976)))]; tensor encoder_layers_14_norm_conv_weight = const()[name = tensor("encoder_layers_14_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357107136)))]; tensor encoder_layers_14_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357111296)))]; tensor encoder_layers_14_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357119552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359218880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359216768)))]; tensor encoder_layers_14_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359227136)))]; tensor encoder_layers_14_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359231296))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360281024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360279936)))]; tensor encoder_layers_14_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_14_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360285184)))]; tensor encoder_layers_14_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_14_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360289344)))]; tensor encoder_layers_14_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360293504)))]; tensor encoder_layers_14_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360309952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364508480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364504320)))]; tensor encoder_layers_14_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364524928)))]; tensor encoder_layers_14_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_14_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364529088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368724544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368723456)))]; tensor encoder_layers_14_norm_out_bias = const()[name = tensor("encoder_layers_14_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368728704)))]; tensor encoder_layers_14_norm_out_weight = const()[name = tensor("encoder_layers_14_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368732864)))]; tensor encoder_layers_15_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_15_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368737024)))]; tensor encoder_layers_15_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_15_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368741184)))]; tensor encoder_layers_15_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368745344)))]; tensor encoder_layers_15_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368761792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372960320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372956160)))]; tensor encoder_layers_15_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372976768)))]; tensor encoder_layers_15_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372980928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377176384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377175296)))]; tensor encoder_layers_15_norm_self_att_bias = const()[name = tensor("encoder_layers_15_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377180544)))]; tensor encoder_layers_15_norm_self_att_weight = const()[name = tensor("encoder_layers_15_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377184704)))]; tensor encoder_layers_15_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377188864)))]; tensor encoder_layers_15_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377193024)))]; tensor encoder_layers_15_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377197184)))]; tensor encoder_layers_15_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377201344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378251072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378249984)))]; tensor encoder_layers_15_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378255232)))]; tensor encoder_layers_15_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378259392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379309120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379308032)))]; tensor encoder_layers_15_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379313280)))]; tensor encoder_layers_15_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379317440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380367168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380366080)))]; tensor encoder_layers_15_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380371328)))]; tensor encoder_layers_15_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380375488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381425216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381424128)))]; tensor encoder_layers_15_norm_conv_bias = const()[name = tensor("encoder_layers_15_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381429376)))]; tensor encoder_layers_15_norm_conv_weight = const()[name = tensor("encoder_layers_15_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381433536)))]; tensor encoder_layers_15_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381437696)))]; tensor encoder_layers_15_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381445952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383545280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383543168)))]; tensor encoder_layers_15_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383553536)))]; tensor encoder_layers_15_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383557696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384607424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384606336)))]; tensor encoder_layers_15_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_15_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384611584)))]; tensor encoder_layers_15_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_15_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384615744)))]; tensor encoder_layers_15_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384619904)))]; tensor encoder_layers_15_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384636352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388834880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388830720)))]; tensor encoder_layers_15_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388851328)))]; tensor encoder_layers_15_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_15_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388855488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393050944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393049856)))]; tensor encoder_layers_15_norm_out_bias = const()[name = tensor("encoder_layers_15_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393055104)))]; tensor encoder_layers_15_norm_out_weight = const()[name = tensor("encoder_layers_15_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393059264)))]; tensor encoder_layers_16_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_16_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393063424)))]; tensor encoder_layers_16_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_16_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393067584)))]; tensor encoder_layers_16_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393071744)))]; tensor encoder_layers_16_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393088192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397286720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397282560)))]; tensor encoder_layers_16_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397303168)))]; tensor encoder_layers_16_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397307328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401502784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401501696)))]; tensor encoder_layers_16_norm_self_att_bias = const()[name = tensor("encoder_layers_16_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401506944)))]; tensor encoder_layers_16_norm_self_att_weight = const()[name = tensor("encoder_layers_16_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401511104)))]; tensor encoder_layers_16_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401515264)))]; tensor encoder_layers_16_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401519424)))]; tensor encoder_layers_16_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401523584)))]; tensor encoder_layers_16_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401527744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402577472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402576384)))]; tensor encoder_layers_16_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402581632)))]; tensor encoder_layers_16_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402585792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403635520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403634432)))]; tensor encoder_layers_16_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403639680)))]; tensor encoder_layers_16_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403643840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(404693568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(404692480)))]; tensor encoder_layers_16_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(404697728)))]; tensor encoder_layers_16_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(404701888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405751616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405750528)))]; tensor encoder_layers_16_norm_conv_bias = const()[name = tensor("encoder_layers_16_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405755776)))]; tensor encoder_layers_16_norm_conv_weight = const()[name = tensor("encoder_layers_16_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405759936)))]; tensor encoder_layers_16_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405764096)))]; tensor encoder_layers_16_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405772352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407871680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407869568)))]; tensor encoder_layers_16_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407879936)))]; tensor encoder_layers_16_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407884096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408933824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408932736)))]; tensor encoder_layers_16_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_16_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408937984)))]; tensor encoder_layers_16_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_16_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408942144)))]; tensor encoder_layers_16_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408946304)))]; tensor encoder_layers_16_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408962752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413161280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413157120)))]; tensor encoder_layers_16_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413177728)))]; tensor encoder_layers_16_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_16_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413181888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417377344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417376256)))]; tensor encoder_layers_16_norm_out_bias = const()[name = tensor("encoder_layers_16_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417381504)))]; tensor encoder_layers_16_norm_out_weight = const()[name = tensor("encoder_layers_16_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417385664)))]; tensor encoder_layers_17_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_17_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417389824)))]; tensor encoder_layers_17_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_17_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417393984)))]; tensor encoder_layers_17_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_17_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417398144)))]; tensor encoder_layers_17_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417414592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421613120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421608960)))]; tensor encoder_layers_17_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_17_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421629568)))]; tensor encoder_layers_17_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421633728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425829184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425828096)))]; tensor encoder_layers_17_norm_self_att_bias = const()[name = tensor("encoder_layers_17_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425833344)))]; tensor encoder_layers_17_norm_self_att_weight = const()[name = tensor("encoder_layers_17_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425837504)))]; tensor encoder_layers_17_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_17_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425841664)))]; tensor encoder_layers_17_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_17_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425845824)))]; tensor encoder_layers_17_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425849984)))]; tensor encoder_layers_17_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425854144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426903872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426902784)))]; tensor encoder_layers_17_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426908032)))]; tensor encoder_layers_17_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426912192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427961920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427960832)))]; tensor encoder_layers_17_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427966080)))]; tensor encoder_layers_17_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427970240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429019968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429018880)))]; tensor encoder_layers_17_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429024128)))]; tensor encoder_layers_17_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429028288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430078016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430076928)))]; tensor encoder_layers_17_norm_conv_bias = const()[name = tensor("encoder_layers_17_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430082176)))]; tensor encoder_layers_17_norm_conv_weight = const()[name = tensor("encoder_layers_17_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430086336)))]; tensor encoder_layers_17_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_17_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430090496)))]; tensor encoder_layers_17_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430098752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432198080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432195968)))]; tensor encoder_layers_17_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_17_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432206336)))]; tensor encoder_layers_17_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432210496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433260224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433259136)))]; tensor encoder_layers_17_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_17_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433264384)))]; tensor encoder_layers_17_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_17_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433268544)))]; tensor encoder_layers_17_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_17_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433272704)))]; tensor encoder_layers_17_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433289152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437487680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437483520)))]; tensor encoder_layers_17_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_17_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437504128)))]; tensor encoder_layers_17_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_17_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437508288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441703744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441702656)))]; tensor encoder_layers_17_norm_out_bias = const()[name = tensor("encoder_layers_17_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441707904)))]; tensor encoder_layers_17_norm_out_weight = const()[name = tensor("encoder_layers_17_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441712064)))]; tensor encoder_layers_18_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_18_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441716224)))]; tensor encoder_layers_18_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_18_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441720384)))]; tensor encoder_layers_18_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_18_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441724544)))]; tensor encoder_layers_18_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441740992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445939520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445935360)))]; tensor encoder_layers_18_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_18_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445955968)))]; tensor encoder_layers_18_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445960128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450155584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450154496)))]; tensor encoder_layers_18_norm_self_att_bias = const()[name = tensor("encoder_layers_18_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450159744)))]; tensor encoder_layers_18_norm_self_att_weight = const()[name = tensor("encoder_layers_18_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450163904)))]; tensor encoder_layers_18_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_18_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450168064)))]; tensor encoder_layers_18_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_18_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450172224)))]; tensor encoder_layers_18_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450176384)))]; tensor encoder_layers_18_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450180544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451230272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451229184)))]; tensor encoder_layers_18_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451234432)))]; tensor encoder_layers_18_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451238592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452288320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452287232)))]; tensor encoder_layers_18_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452292480)))]; tensor encoder_layers_18_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452296640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(453346368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(453345280)))]; tensor encoder_layers_18_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(453350528)))]; tensor encoder_layers_18_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(453354688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454404416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454403328)))]; tensor encoder_layers_18_norm_conv_bias = const()[name = tensor("encoder_layers_18_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454408576)))]; tensor encoder_layers_18_norm_conv_weight = const()[name = tensor("encoder_layers_18_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454412736)))]; tensor encoder_layers_18_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_18_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454416896)))]; tensor encoder_layers_18_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454425152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456524480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456522368)))]; tensor encoder_layers_18_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_18_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456532736)))]; tensor encoder_layers_18_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456536896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457586624))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457585536)))]; tensor encoder_layers_18_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_18_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457590784)))]; tensor encoder_layers_18_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_18_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457594944)))]; tensor encoder_layers_18_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_18_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457599104)))]; tensor encoder_layers_18_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457615552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461814080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461809920)))]; tensor encoder_layers_18_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_18_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461830528)))]; tensor encoder_layers_18_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_18_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461834688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466030144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466029056)))]; tensor encoder_layers_18_norm_out_bias = const()[name = tensor("encoder_layers_18_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466034304)))]; tensor encoder_layers_18_norm_out_weight = const()[name = tensor("encoder_layers_18_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466038464)))]; tensor encoder_layers_19_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_19_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466042624)))]; tensor encoder_layers_19_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_19_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466046784)))]; tensor encoder_layers_19_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_19_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466050944)))]; tensor encoder_layers_19_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466067392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470265920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470261760)))]; tensor encoder_layers_19_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_19_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470282368)))]; tensor encoder_layers_19_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470286528))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474481984))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474480896)))]; tensor encoder_layers_19_norm_self_att_bias = const()[name = tensor("encoder_layers_19_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474486144)))]; tensor encoder_layers_19_norm_self_att_weight = const()[name = tensor("encoder_layers_19_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474490304)))]; tensor encoder_layers_19_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_19_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474494464)))]; tensor encoder_layers_19_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_19_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474498624)))]; tensor encoder_layers_19_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474502784)))]; tensor encoder_layers_19_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474506944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475556672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475555584)))]; tensor encoder_layers_19_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475560832)))]; tensor encoder_layers_19_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475564992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476614720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476613632)))]; tensor encoder_layers_19_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476618880)))]; tensor encoder_layers_19_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476623040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(477672768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(477671680)))]; tensor encoder_layers_19_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(477676928)))]; tensor encoder_layers_19_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(477681088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478730816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478729728)))]; tensor encoder_layers_19_norm_conv_bias = const()[name = tensor("encoder_layers_19_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478734976)))]; tensor encoder_layers_19_norm_conv_weight = const()[name = tensor("encoder_layers_19_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478739136)))]; tensor encoder_layers_19_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_19_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478743296)))]; tensor encoder_layers_19_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478751552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480850880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480848768)))]; tensor encoder_layers_19_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_19_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480859136)))]; tensor encoder_layers_19_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480863296))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481913024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481911936)))]; tensor encoder_layers_19_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_19_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481917184)))]; tensor encoder_layers_19_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_19_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481921344)))]; tensor encoder_layers_19_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_19_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481925504)))]; tensor encoder_layers_19_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481941952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486140480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486136320)))]; tensor encoder_layers_19_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_19_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486156928)))]; tensor encoder_layers_19_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_19_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486161088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490356544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490355456)))]; tensor encoder_layers_19_norm_out_bias = const()[name = tensor("encoder_layers_19_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490360704)))]; tensor encoder_layers_19_norm_out_weight = const()[name = tensor("encoder_layers_19_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490364864)))]; tensor encoder_layers_20_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_20_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490369024)))]; tensor encoder_layers_20_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_20_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490373184)))]; tensor encoder_layers_20_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_20_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490377344)))]; tensor encoder_layers_20_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490393792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494592320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494588160)))]; tensor encoder_layers_20_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_20_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494608768)))]; tensor encoder_layers_20_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494612928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498808384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498807296)))]; tensor encoder_layers_20_norm_self_att_bias = const()[name = tensor("encoder_layers_20_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498812544)))]; tensor encoder_layers_20_norm_self_att_weight = const()[name = tensor("encoder_layers_20_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498816704)))]; tensor encoder_layers_20_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_20_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498820864)))]; tensor encoder_layers_20_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_20_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498825024)))]; tensor encoder_layers_20_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498829184)))]; tensor encoder_layers_20_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498833344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499883072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499881984)))]; tensor encoder_layers_20_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499887232)))]; tensor encoder_layers_20_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499891392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500941120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500940032)))]; tensor encoder_layers_20_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500945280)))]; tensor encoder_layers_20_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500949440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501999168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501998080)))]; tensor encoder_layers_20_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(502003328)))]; tensor encoder_layers_20_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(502007488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503057216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503056128)))]; tensor encoder_layers_20_norm_conv_bias = const()[name = tensor("encoder_layers_20_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503061376)))]; tensor encoder_layers_20_norm_conv_weight = const()[name = tensor("encoder_layers_20_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503065536)))]; tensor encoder_layers_20_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_20_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503069696)))]; tensor encoder_layers_20_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(503077952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505177280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505175168)))]; tensor encoder_layers_20_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_20_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505185536)))]; tensor encoder_layers_20_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505189696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506239424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506238336)))]; tensor encoder_layers_20_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_20_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506243584)))]; tensor encoder_layers_20_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_20_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506247744)))]; tensor encoder_layers_20_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_20_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506251904)))]; tensor encoder_layers_20_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506268352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510466880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510462720)))]; tensor encoder_layers_20_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_20_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510483328)))]; tensor encoder_layers_20_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_20_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510487488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514682944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514681856)))]; tensor encoder_layers_20_norm_out_bias = const()[name = tensor("encoder_layers_20_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514687104)))]; tensor encoder_layers_20_norm_out_weight = const()[name = tensor("encoder_layers_20_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514691264)))]; tensor encoder_layers_21_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_21_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514695424)))]; tensor encoder_layers_21_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_21_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514699584)))]; tensor encoder_layers_21_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_21_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514703744)))]; tensor encoder_layers_21_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514720192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518918720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518914560)))]; tensor encoder_layers_21_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_21_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518935168)))]; tensor encoder_layers_21_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518939328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523134784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523133696)))]; tensor encoder_layers_21_norm_self_att_bias = const()[name = tensor("encoder_layers_21_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523138944)))]; tensor encoder_layers_21_norm_self_att_weight = const()[name = tensor("encoder_layers_21_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523143104)))]; tensor encoder_layers_21_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_21_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523147264)))]; tensor encoder_layers_21_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_21_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523151424)))]; tensor encoder_layers_21_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523155584)))]; tensor encoder_layers_21_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523159744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524209472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524208384)))]; tensor encoder_layers_21_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524213632)))]; tensor encoder_layers_21_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(524217792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525267520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525266432)))]; tensor encoder_layers_21_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525271680)))]; tensor encoder_layers_21_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525275840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526325568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526324480)))]; tensor encoder_layers_21_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526329728)))]; tensor encoder_layers_21_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526333888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527383616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527382528)))]; tensor encoder_layers_21_norm_conv_bias = const()[name = tensor("encoder_layers_21_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527387776)))]; tensor encoder_layers_21_norm_conv_weight = const()[name = tensor("encoder_layers_21_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527391936)))]; tensor encoder_layers_21_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_21_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527396096)))]; tensor encoder_layers_21_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527404352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529503680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529501568)))]; tensor encoder_layers_21_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_21_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529511936)))]; tensor encoder_layers_21_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529516096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530565824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530564736)))]; tensor encoder_layers_21_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_21_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530569984)))]; tensor encoder_layers_21_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_21_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530574144)))]; tensor encoder_layers_21_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_21_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530578304)))]; tensor encoder_layers_21_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530594752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(534793280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(534789120)))]; tensor encoder_layers_21_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_21_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(534809728)))]; tensor encoder_layers_21_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_21_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(534813888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539009344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539008256)))]; tensor encoder_layers_21_norm_out_bias = const()[name = tensor("encoder_layers_21_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539013504)))]; tensor encoder_layers_21_norm_out_weight = const()[name = tensor("encoder_layers_21_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539017664)))]; tensor encoder_layers_22_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_22_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539021824)))]; tensor encoder_layers_22_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_22_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539025984)))]; tensor encoder_layers_22_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_22_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539030144)))]; tensor encoder_layers_22_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(539046592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(543245120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(543240960)))]; tensor encoder_layers_22_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_22_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(543261568)))]; tensor encoder_layers_22_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(543265728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547461184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547460096)))]; tensor encoder_layers_22_norm_self_att_bias = const()[name = tensor("encoder_layers_22_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547465344)))]; tensor encoder_layers_22_norm_self_att_weight = const()[name = tensor("encoder_layers_22_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547469504)))]; tensor encoder_layers_22_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_22_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547473664)))]; tensor encoder_layers_22_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_22_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547477824)))]; tensor encoder_layers_22_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547481984)))]; tensor encoder_layers_22_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547486144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548535872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548534784)))]; tensor encoder_layers_22_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548540032)))]; tensor encoder_layers_22_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548544192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549593920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549592832)))]; tensor encoder_layers_22_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549598080)))]; tensor encoder_layers_22_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549602240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550651968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550650880)))]; tensor encoder_layers_22_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550656128)))]; tensor encoder_layers_22_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550660288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551710016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551708928)))]; tensor encoder_layers_22_norm_conv_bias = const()[name = tensor("encoder_layers_22_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551714176)))]; tensor encoder_layers_22_norm_conv_weight = const()[name = tensor("encoder_layers_22_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551718336)))]; tensor encoder_layers_22_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_22_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551722496)))]; tensor encoder_layers_22_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551730752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553830080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553827968)))]; tensor encoder_layers_22_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_22_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553838336)))]; tensor encoder_layers_22_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553842496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554892224))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554891136)))]; tensor encoder_layers_22_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_22_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554896384)))]; tensor encoder_layers_22_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_22_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554900544)))]; tensor encoder_layers_22_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_22_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554904704)))]; tensor encoder_layers_22_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554921152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559119680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559115520)))]; tensor encoder_layers_22_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_22_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559136128)))]; tensor encoder_layers_22_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_22_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559140288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563335744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563334656)))]; tensor encoder_layers_22_norm_out_bias = const()[name = tensor("encoder_layers_22_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563339904)))]; tensor encoder_layers_22_norm_out_weight = const()[name = tensor("encoder_layers_22_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563344064)))]; tensor encoder_layers_23_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_23_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563348224)))]; tensor encoder_layers_23_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_23_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563352384)))]; tensor encoder_layers_23_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_23_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563356544)))]; tensor encoder_layers_23_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563372992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567571520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567567360)))]; tensor encoder_layers_23_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_23_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567587968)))]; tensor encoder_layers_23_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567592128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571787584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571786496)))]; tensor encoder_layers_23_norm_self_att_bias = const()[name = tensor("encoder_layers_23_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571791744)))]; tensor encoder_layers_23_norm_self_att_weight = const()[name = tensor("encoder_layers_23_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571795904)))]; tensor encoder_layers_23_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_23_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571800064)))]; tensor encoder_layers_23_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_23_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571804224)))]; tensor encoder_layers_23_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571808384)))]; tensor encoder_layers_23_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571812544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572862272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572861184)))]; tensor encoder_layers_23_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572866432)))]; tensor encoder_layers_23_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572870592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573920320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573919232)))]; tensor encoder_layers_23_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573924480)))]; tensor encoder_layers_23_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573928640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574978368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574977280)))]; tensor encoder_layers_23_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574982528)))]; tensor encoder_layers_23_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574986688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(576036416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(576035328)))]; tensor encoder_layers_23_norm_conv_bias = const()[name = tensor("encoder_layers_23_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(576040576)))]; tensor encoder_layers_23_norm_conv_weight = const()[name = tensor("encoder_layers_23_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(576044736)))]; tensor encoder_layers_23_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_23_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(576048896)))]; tensor encoder_layers_23_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(576057152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578156480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578154368)))]; tensor encoder_layers_23_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_23_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578164736)))]; tensor encoder_layers_23_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578168896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579218624))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579217536)))]; tensor encoder_layers_23_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_23_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579222784)))]; tensor encoder_layers_23_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_23_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579226944)))]; tensor encoder_layers_23_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_23_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579231104)))]; tensor encoder_layers_23_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579247552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(583446080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(583441920)))]; tensor encoder_layers_23_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_23_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(583462528)))]; tensor encoder_layers_23_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_23_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(583466688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587662144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587661056)))]; tensor encoder_layers_23_norm_out_bias = const()[name = tensor("encoder_layers_23_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587666304)))]; tensor encoder_layers_23_norm_out_weight = const()[name = tensor("encoder_layers_23_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587670464)))]; tensor var_11 = const()[name = tensor("op_11"), val = tensor(0x1.4f8b58p-17)]; tensor var_13 = const()[name = tensor("op_13"), val = tensor(0x0p+0)]; tensor var_14 = const()[name = tensor("op_14"), val = tensor(-0x1.388p+13)]; tensor var_32 = const()[name = tensor("op_32"), val = tensor(-1)]; tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; tensor var_88_dtype_0 = const()[name = tensor("op_88_dtype_0"), val = tensor("fp32")]; tensor var_89_promoted = const()[name = tensor("op_89_promoted"), val = tensor(-0x1p+0)]; tensor var_88 = cast(dtype = var_88_dtype_0, x = mel_length)[name = tensor("cast_1")]; tensor var_90 = add(x = var_88, y = var_89_promoted)[name = tensor("op_90")]; tensor _inversed_92_y_0 = const()[name = tensor("_inversed_92_y_0"), val = tensor(0x1p-1)]; tensor _inversed_92 = mul(x = var_90, y = _inversed_92_y_0)[name = tensor("_inversed_92")]; tensor var_93 = const()[name = tensor("op_93"), val = tensor(0x1p+0)]; tensor lengths_1 = add(x = _inversed_92, y = var_93)[name = tensor("lengths_1")]; tensor lengths_3 = floor(x = lengths_1)[name = tensor("lengths_3")]; tensor var_97_promoted = const()[name = tensor("op_97_promoted"), val = tensor(-0x1p+0)]; tensor var_98 = add(x = lengths_3, y = var_97_promoted)[name = tensor("op_98")]; tensor _inversed_100_y_0 = const()[name = tensor("_inversed_100_y_0"), val = tensor(0x1p-1)]; tensor _inversed_100 = mul(x = var_98, y = _inversed_100_y_0)[name = tensor("_inversed_100")]; tensor var_101 = const()[name = tensor("op_101"), val = tensor(0x1p+0)]; tensor lengths_7 = add(x = _inversed_100, y = var_101)[name = tensor("lengths_7")]; tensor lengths_9 = floor(x = lengths_7)[name = tensor("lengths_9")]; tensor var_105_promoted = const()[name = tensor("op_105_promoted"), val = tensor(-0x1p+0)]; tensor var_106 = add(x = lengths_9, y = var_105_promoted)[name = tensor("op_106")]; tensor _inversed_108_y_0 = const()[name = tensor("_inversed_108_y_0"), val = tensor(0x1p-1)]; tensor _inversed_108 = mul(x = var_106, y = _inversed_108_y_0)[name = tensor("_inversed_108")]; tensor var_109 = const()[name = tensor("op_109"), val = tensor(0x1p+0)]; tensor lengths_13 = add(x = _inversed_108, y = var_109)[name = tensor("lengths_13")]; tensor lengths = floor(x = lengths_13)[name = tensor("lengths")]; tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; tensor x_1 = transpose(perm = x_1_perm_0, x = melspectrogram_features)[name = tensor("transpose_314")]; tensor input_1 = expand_dims(axes = input_1_axes_0, x = x_1)[name = tensor("input_1")]; tensor input_3_pad_type_0 = const()[name = tensor("input_3_pad_type_0"), val = tensor("custom")]; tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_3_strides_0 = const()[name = tensor("input_3_strides_0"), val = tensor([2, 2])]; tensor input_3_dilations_0 = const()[name = tensor("input_3_dilations_0"), val = tensor([1, 1])]; tensor input_3_groups_0 = const()[name = tensor("input_3_groups_0"), val = tensor(1)]; tensor input_3 = conv(bias = encoder_pre_encode_conv_0_bias, dilations = input_3_dilations_0, groups = input_3_groups_0, pad = input_3_pad_0, pad_type = input_3_pad_type_0, strides = input_3_strides_0, weight = encoder_pre_encode_conv_0_weight_quantized, x = input_1)[name = tensor("input_3")]; tensor input_5 = relu(x = input_3)[name = tensor("input_5")]; tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("custom")]; tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_7_strides_0 = const()[name = tensor("input_7_strides_0"), val = tensor([2, 2])]; tensor input_7_groups_0 = const()[name = tensor("input_7_groups_0"), val = tensor(256)]; tensor input_7_dilations_0 = const()[name = tensor("input_7_dilations_0"), val = tensor([1, 1])]; tensor input_7 = conv(bias = encoder_pre_encode_conv_2_bias, dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = encoder_pre_encode_conv_2_weight_quantized, x = input_5)[name = tensor("input_7")]; tensor input_9_pad_type_0 = const()[name = tensor("input_9_pad_type_0"), val = tensor("valid")]; tensor input_9_strides_0 = const()[name = tensor("input_9_strides_0"), val = tensor([1, 1])]; tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_9_dilations_0 = const()[name = tensor("input_9_dilations_0"), val = tensor([1, 1])]; tensor input_9_groups_0 = const()[name = tensor("input_9_groups_0"), val = tensor(1)]; tensor input_9 = conv(bias = encoder_pre_encode_conv_3_bias, dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = encoder_pre_encode_conv_3_weight_quantized, x = input_7)[name = tensor("input_9")]; tensor input_11 = relu(x = input_9)[name = tensor("input_11")]; tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("custom")]; tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([2, 2])]; tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(256)]; tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; tensor input_13 = conv(bias = encoder_pre_encode_conv_5_bias, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = encoder_pre_encode_conv_5_weight_quantized, x = input_11)[name = tensor("input_13")]; tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("valid")]; tensor input_15_strides_0 = const()[name = tensor("input_15_strides_0"), val = tensor([1, 1])]; tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_15_dilations_0 = const()[name = tensor("input_15_dilations_0"), val = tensor([1, 1])]; tensor input_15_groups_0 = const()[name = tensor("input_15_groups_0"), val = tensor(1)]; tensor input_15 = conv(bias = encoder_pre_encode_conv_6_bias, dilations = input_15_dilations_0, groups = input_15_groups_0, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = input_15_strides_0, weight = encoder_pre_encode_conv_6_weight_quantized, x = input_13)[name = tensor("input_15")]; tensor x_3 = relu(x = input_15)[name = tensor("x_3")]; tensor var_159_perm_0 = const()[name = tensor("op_159_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_160 = const()[name = tensor("op_160"), val = tensor([1, 188, -1])]; tensor var_159 = transpose(perm = var_159_perm_0, x = x_3)[name = tensor("transpose_313")]; tensor input_17 = reshape(shape = var_160, x = var_159)[name = tensor("input_17")]; tensor audio_signal_1 = linear(bias = encoder_pre_encode_out_bias, weight = encoder_pre_encode_out_weight_quantized, x = input_17)[name = tensor("linear_0")]; tensor padding_length_dtype_0 = const()[name = tensor("padding_length_dtype_0"), val = tensor("int32")]; tensor var_171 = const()[name = tensor("op_171"), val = tensor(0x1p+5)]; tensor x_5 = mul(x = audio_signal_1, y = var_171)[name = tensor("x_5")]; tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; tensor var_200_axes_0 = const()[name = tensor("op_200_axes_0"), val = tensor([-1])]; tensor encoder_length = cast(dtype = padding_length_dtype_0, x = lengths)[name = tensor("cast_0")]; tensor var_200 = expand_dims(axes = var_200_axes_0, x = encoder_length)[name = tensor("op_200")]; tensor pad_mask_1 = less(x = expand_dims_0, y = var_200)[name = tensor("pad_mask_1")]; tensor var_202_axes_0 = const()[name = tensor("op_202_axes_0"), val = tensor([1])]; tensor var_202 = expand_dims(axes = var_202_axes_0, x = pad_mask_1)[name = tensor("op_202")]; tensor var_203 = const()[name = tensor("op_203"), val = tensor([1, 188, 1])]; tensor pad_mask_for_att_mask_1 = tile(reps = var_203, x = var_202)[name = tensor("pad_mask_for_att_mask_1")]; tensor var_205_perm_0 = const()[name = tensor("op_205_perm_0"), val = tensor([0, 2, 1])]; tensor var_205 = transpose(perm = var_205_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_312")]; tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_205)[name = tensor("pad_mask_for_att_mask")]; tensor const_7 = const()[name = tensor("const_7"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_7)[name = tensor("att_mask")]; tensor mask_1 = logical_not(x = att_mask)[name = tensor("mask_1")]; tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; tensor input_21 = layer_norm(axes = input_21_axes_0, beta = encoder_layers_0_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_0_norm_feed_forward1_weight, x = x_5)[name = tensor("input_21")]; tensor input_23 = linear(bias = encoder_layers_0_feed_forward1_linear1_bias, weight = encoder_layers_0_feed_forward1_linear1_weight_quantized, x = input_21)[name = tensor("linear_1")]; tensor input_25 = silu(x = input_23)[name = tensor("input_25")]; tensor input_29 = linear(bias = encoder_layers_0_feed_forward1_linear2_bias, weight = encoder_layers_0_feed_forward1_linear2_weight_quantized, x = input_25)[name = tensor("linear_2")]; tensor var_238 = const()[name = tensor("op_238"), val = tensor(0x1p-1)]; tensor var_239 = mul(x = input_29, y = var_238)[name = tensor("op_239")]; tensor input_31 = add(x = x_5, y = var_239)[name = tensor("input_31")]; tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; tensor query_1 = layer_norm(axes = query_1_axes_0, beta = encoder_layers_0_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_0_norm_self_att_weight, x = input_31)[name = tensor("query_1")]; tensor var_255 = linear(bias = encoder_layers_0_self_attn_linear_q_bias, weight = encoder_layers_0_self_attn_linear_q_weight_quantized, x = query_1)[name = tensor("linear_3")]; tensor var_256 = const()[name = tensor("op_256"), val = tensor([1, -1, 8, 128])]; tensor q_1 = reshape(shape = var_256, x = var_255)[name = tensor("q_1")]; tensor var_260 = linear(bias = encoder_layers_0_self_attn_linear_k_bias, weight = encoder_layers_0_self_attn_linear_k_weight_quantized, x = query_1)[name = tensor("linear_4")]; tensor var_261 = const()[name = tensor("op_261"), val = tensor([1, -1, 8, 128])]; tensor k_1 = reshape(shape = var_261, x = var_260)[name = tensor("k_1")]; tensor var_265 = linear(bias = encoder_layers_0_self_attn_linear_v_bias, weight = encoder_layers_0_self_attn_linear_v_weight_quantized, x = query_1)[name = tensor("linear_5")]; tensor var_266 = const()[name = tensor("op_266"), val = tensor([1, -1, 8, 128])]; tensor v_1 = reshape(shape = var_266, x = var_265)[name = tensor("v_1")]; tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_278 = add(x = q_1, y = encoder_layers_0_self_attn_pos_bias_u)[name = tensor("op_278")]; tensor var_280 = add(x = q_1, y = encoder_layers_0_self_attn_pos_bias_v)[name = tensor("op_280")]; tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_282_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_282_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587674624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588059136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588058688)))]; tensor x_9_transpose_x_0 = const()[name = tensor("x_9_transpose_x_0"), val = tensor(false)]; tensor x_9_transpose_y_0 = const()[name = tensor("x_9_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_1 = transpose(perm = q_with_bias_v_1_perm_0, x = var_280)[name = tensor("transpose_311")]; tensor x_9 = matmul(transpose_x = x_9_transpose_x_0, transpose_y = x_9_transpose_y_0, x = q_with_bias_v_1, y = op_282_quantized)[name = tensor("x_9")]; tensor const_14 = const()[name = tensor("const_14"), val = tensor(0x0p+0)]; tensor x_11_pad_0 = const()[name = tensor("x_11_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_11_mode_0 = const()[name = tensor("x_11_mode_0"), val = tensor("constant")]; tensor x_11 = pad(constant_val = const_14, mode = x_11_mode_0, pad = x_11_pad_0, x = x_9)[name = tensor("x_11")]; tensor var_290 = const()[name = tensor("op_290"), val = tensor([1, 8, -1, 188])]; tensor x_13 = reshape(shape = var_290, x = x_11)[name = tensor("x_13")]; tensor var_294_begin_0 = const()[name = tensor("op_294_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_294_end_0 = const()[name = tensor("op_294_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_294_end_mask_0 = const()[name = tensor("op_294_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_294 = slice_by_index(begin = var_294_begin_0, end = var_294_end_0, end_mask = var_294_end_mask_0, x = x_13)[name = tensor("op_294")]; tensor var_295 = const()[name = tensor("op_295"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_1 = reshape(shape = var_295, x = var_294)[name = tensor("matrix_bd_1")]; tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; tensor transpose_96_perm_0 = const()[name = tensor("transpose_96_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_97_perm_0 = const()[name = tensor("transpose_97_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_97 = transpose(perm = transpose_97_perm_0, x = k_1)[name = tensor("transpose_309")]; tensor transpose_96 = transpose(perm = transpose_96_perm_0, x = var_278)[name = tensor("transpose_310")]; tensor matrix_ac_1 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_96, y = transpose_97)[name = tensor("matrix_ac_1")]; tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_3 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1)[name = tensor("matrix_bd_3")]; tensor var_304 = add(x = matrix_ac_1, y = matrix_bd_3)[name = tensor("op_304")]; tensor _inversed_scores_1_y_0 = const()[name = tensor("_inversed_scores_1_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_1 = mul(x = var_304, y = _inversed_scores_1_y_0)[name = tensor("_inversed_scores_1")]; tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; tensor scores_3 = select(a = var_14, b = _inversed_scores_1, cond = mask_3)[name = tensor("scores_3")]; tensor var_310 = softmax(axis = var_32, x = scores_3)[name = tensor("op_310")]; tensor input_33 = select(a = var_13, b = var_310, cond = mask_3)[name = tensor("input_33")]; tensor x_15_transpose_x_0 = const()[name = tensor("x_15_transpose_x_0"), val = tensor(false)]; tensor x_15_transpose_y_0 = const()[name = tensor("x_15_transpose_y_0"), val = tensor(false)]; tensor value_3 = transpose(perm = value_3_perm_0, x = v_1)[name = tensor("transpose_308")]; tensor x_15 = matmul(transpose_x = x_15_transpose_x_0, transpose_y = x_15_transpose_y_0, x = input_33, y = value_3)[name = tensor("x_15")]; tensor var_314_perm_0 = const()[name = tensor("op_314_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_315 = const()[name = tensor("op_315"), val = tensor([1, -1, 1024])]; tensor var_314 = transpose(perm = var_314_perm_0, x = x_15)[name = tensor("transpose_307")]; tensor input_35 = reshape(shape = var_315, x = var_314)[name = tensor("input_35")]; tensor input_37 = linear(bias = encoder_layers_0_self_attn_linear_out_bias, weight = encoder_layers_0_self_attn_linear_out_weight_quantized, x = input_35)[name = tensor("linear_7")]; tensor input_39 = add(x = input_31, y = input_37)[name = tensor("input_39")]; tensor x_19_axes_0 = const()[name = tensor("x_19_axes_0"), val = tensor([-1])]; tensor x_19 = layer_norm(axes = x_19_axes_0, beta = encoder_layers_0_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_0_norm_conv_weight, x = input_39)[name = tensor("x_19")]; tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; tensor input_41 = transpose(perm = input_41_perm_0, x = x_19)[name = tensor("transpose_306")]; tensor input_43 = conv(bias = encoder_layers_0_conv_pointwise_conv1_bias, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = encoder_layers_0_conv_pointwise_conv1_weight_quantized, x = input_41)[name = tensor("input_43")]; tensor x_21_split_num_splits_0 = const()[name = tensor("x_21_split_num_splits_0"), val = tensor(2)]; tensor x_21_split_axis_0 = const()[name = tensor("x_21_split_axis_0"), val = tensor(1)]; tensor x_21_split_0, tensor x_21_split_1 = split(axis = x_21_split_axis_0, num_splits = x_21_split_num_splits_0, x = input_43)[name = tensor("x_21_split")]; tensor x_21_split_1_sigmoid = sigmoid(x = x_21_split_1)[name = tensor("x_21_split_1_sigmoid")]; tensor x_21 = mul(x = x_21_split_0, y = x_21_split_1_sigmoid)[name = tensor("x_21")]; tensor var_339_axes_0 = const()[name = tensor("op_339_axes_0"), val = tensor([1])]; tensor var_339 = expand_dims(axes = var_339_axes_0, x = pad_mask)[name = tensor("op_339")]; tensor input_45 = select(a = var_13, b = x_21, cond = var_339)[name = tensor("input_45")]; tensor const_17 = const()[name = tensor("const_17"), val = tensor(0x0p+0)]; tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; tensor input_47 = pad(constant_val = const_17, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45)[name = tensor("input_47")]; tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(1024)]; tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; tensor const_248_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_248_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588060736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588071104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588070016)))]; tensor const_249 = const()[name = tensor("const_249"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588075264)))]; tensor input_51 = conv(bias = const_249, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_248_quantized, x = input_47)[name = tensor("input_51")]; tensor input_53 = silu(x = input_51)[name = tensor("input_53")]; tensor x_23_pad_type_0 = const()[name = tensor("x_23_pad_type_0"), val = tensor("valid")]; tensor x_23_strides_0 = const()[name = tensor("x_23_strides_0"), val = tensor([1])]; tensor x_23_pad_0 = const()[name = tensor("x_23_pad_0"), val = tensor([0, 0])]; tensor x_23_dilations_0 = const()[name = tensor("x_23_dilations_0"), val = tensor([1])]; tensor x_23_groups_0 = const()[name = tensor("x_23_groups_0"), val = tensor(1)]; tensor x_23 = conv(bias = encoder_layers_0_conv_pointwise_conv2_bias, dilations = x_23_dilations_0, groups = x_23_groups_0, pad = x_23_pad_0, pad_type = x_23_pad_type_0, strides = x_23_strides_0, weight = encoder_layers_0_conv_pointwise_conv2_weight_quantized, x = input_53)[name = tensor("x_23")]; tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; tensor input_55 = transpose(perm = input_55_perm_0, x = x_23)[name = tensor("transpose_305")]; tensor input_57 = add(x = input_39, y = input_55)[name = tensor("input_57")]; tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; tensor input_59 = layer_norm(axes = input_59_axes_0, beta = encoder_layers_0_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_0_norm_feed_forward2_weight, x = input_57)[name = tensor("input_59")]; tensor input_61 = linear(bias = encoder_layers_0_feed_forward2_linear1_bias, weight = encoder_layers_0_feed_forward2_linear1_weight_quantized, x = input_59)[name = tensor("linear_8")]; tensor input_63 = silu(x = input_61)[name = tensor("input_63")]; tensor input_67 = linear(bias = encoder_layers_0_feed_forward2_linear2_bias, weight = encoder_layers_0_feed_forward2_linear2_weight_quantized, x = input_63)[name = tensor("linear_9")]; tensor var_381 = const()[name = tensor("op_381"), val = tensor(0x1p-1)]; tensor var_382 = mul(x = input_67, y = var_381)[name = tensor("op_382")]; tensor input_69 = add(x = input_57, y = var_382)[name = tensor("input_69")]; tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; tensor input_71 = layer_norm(axes = input_71_axes_0, beta = encoder_layers_0_norm_out_bias, epsilon = var_11, gamma = encoder_layers_0_norm_out_weight, x = input_69)[name = tensor("input_71")]; tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; tensor input_73 = layer_norm(axes = input_73_axes_0, beta = encoder_layers_1_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_1_norm_feed_forward1_weight, x = input_71)[name = tensor("input_73")]; tensor input_75 = linear(bias = encoder_layers_1_feed_forward1_linear1_bias, weight = encoder_layers_1_feed_forward1_linear1_weight_quantized, x = input_73)[name = tensor("linear_10")]; tensor input_77 = silu(x = input_75)[name = tensor("input_77")]; tensor input_81 = linear(bias = encoder_layers_1_feed_forward1_linear2_bias, weight = encoder_layers_1_feed_forward1_linear2_weight_quantized, x = input_77)[name = tensor("linear_11")]; tensor var_412 = const()[name = tensor("op_412"), val = tensor(0x1p-1)]; tensor var_413 = mul(x = input_81, y = var_412)[name = tensor("op_413")]; tensor input_83 = add(x = input_71, y = var_413)[name = tensor("input_83")]; tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; tensor query_3 = layer_norm(axes = query_3_axes_0, beta = encoder_layers_1_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_1_norm_self_att_weight, x = input_83)[name = tensor("query_3")]; tensor var_429 = linear(bias = encoder_layers_1_self_attn_linear_q_bias, weight = encoder_layers_1_self_attn_linear_q_weight_quantized, x = query_3)[name = tensor("linear_12")]; tensor var_430 = const()[name = tensor("op_430"), val = tensor([1, -1, 8, 128])]; tensor q_7 = reshape(shape = var_430, x = var_429)[name = tensor("q_7")]; tensor var_434 = linear(bias = encoder_layers_1_self_attn_linear_k_bias, weight = encoder_layers_1_self_attn_linear_k_weight_quantized, x = query_3)[name = tensor("linear_13")]; tensor var_435 = const()[name = tensor("op_435"), val = tensor([1, -1, 8, 128])]; tensor k_5 = reshape(shape = var_435, x = var_434)[name = tensor("k_5")]; tensor var_439 = linear(bias = encoder_layers_1_self_attn_linear_v_bias, weight = encoder_layers_1_self_attn_linear_v_weight_quantized, x = query_3)[name = tensor("linear_14")]; tensor var_440 = const()[name = tensor("op_440"), val = tensor([1, -1, 8, 128])]; tensor v_3 = reshape(shape = var_440, x = var_439)[name = tensor("v_3")]; tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_452 = add(x = q_7, y = encoder_layers_1_self_attn_pos_bias_u)[name = tensor("op_452")]; tensor var_454 = add(x = q_7, y = encoder_layers_1_self_attn_pos_bias_v)[name = tensor("op_454")]; tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_456_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_456_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588079424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588463936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588463488)))]; tensor x_31_transpose_x_0 = const()[name = tensor("x_31_transpose_x_0"), val = tensor(false)]; tensor x_31_transpose_y_0 = const()[name = tensor("x_31_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_3 = transpose(perm = q_with_bias_v_3_perm_0, x = var_454)[name = tensor("transpose_304")]; tensor x_31 = matmul(transpose_x = x_31_transpose_x_0, transpose_y = x_31_transpose_y_0, x = q_with_bias_v_3, y = op_456_quantized)[name = tensor("x_31")]; tensor const_24 = const()[name = tensor("const_24"), val = tensor(0x0p+0)]; tensor x_33_pad_0 = const()[name = tensor("x_33_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_33_mode_0 = const()[name = tensor("x_33_mode_0"), val = tensor("constant")]; tensor x_33 = pad(constant_val = const_24, mode = x_33_mode_0, pad = x_33_pad_0, x = x_31)[name = tensor("x_33")]; tensor var_464 = const()[name = tensor("op_464"), val = tensor([1, 8, -1, 188])]; tensor x_35 = reshape(shape = var_464, x = x_33)[name = tensor("x_35")]; tensor var_468_begin_0 = const()[name = tensor("op_468_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_468_end_0 = const()[name = tensor("op_468_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_468_end_mask_0 = const()[name = tensor("op_468_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_468 = slice_by_index(begin = var_468_begin_0, end = var_468_end_0, end_mask = var_468_end_mask_0, x = x_35)[name = tensor("op_468")]; tensor var_469 = const()[name = tensor("op_469"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_5 = reshape(shape = var_469, x = var_468)[name = tensor("matrix_bd_5")]; tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; tensor transpose_98_perm_0 = const()[name = tensor("transpose_98_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_99_perm_0 = const()[name = tensor("transpose_99_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_99 = transpose(perm = transpose_99_perm_0, x = k_5)[name = tensor("transpose_302")]; tensor transpose_98 = transpose(perm = transpose_98_perm_0, x = var_452)[name = tensor("transpose_303")]; tensor matrix_ac_3 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_98, y = transpose_99)[name = tensor("matrix_ac_3")]; tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_7 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5)[name = tensor("matrix_bd_7")]; tensor var_478 = add(x = matrix_ac_3, y = matrix_bd_7)[name = tensor("op_478")]; tensor _inversed_scores_5_y_0 = const()[name = tensor("_inversed_scores_5_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_5 = mul(x = var_478, y = _inversed_scores_5_y_0)[name = tensor("_inversed_scores_5")]; tensor scores_7 = select(a = var_14, b = _inversed_scores_5, cond = mask_3)[name = tensor("scores_7")]; tensor var_484 = softmax(axis = var_32, x = scores_7)[name = tensor("op_484")]; tensor input_85 = select(a = var_13, b = var_484, cond = mask_3)[name = tensor("input_85")]; tensor x_37_transpose_x_0 = const()[name = tensor("x_37_transpose_x_0"), val = tensor(false)]; tensor x_37_transpose_y_0 = const()[name = tensor("x_37_transpose_y_0"), val = tensor(false)]; tensor value_5 = transpose(perm = value_5_perm_0, x = v_3)[name = tensor("transpose_301")]; tensor x_37 = matmul(transpose_x = x_37_transpose_x_0, transpose_y = x_37_transpose_y_0, x = input_85, y = value_5)[name = tensor("x_37")]; tensor var_488_perm_0 = const()[name = tensor("op_488_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_489 = const()[name = tensor("op_489"), val = tensor([1, -1, 1024])]; tensor var_488 = transpose(perm = var_488_perm_0, x = x_37)[name = tensor("transpose_300")]; tensor input_87 = reshape(shape = var_489, x = var_488)[name = tensor("input_87")]; tensor input_89 = linear(bias = encoder_layers_1_self_attn_linear_out_bias, weight = encoder_layers_1_self_attn_linear_out_weight_quantized, x = input_87)[name = tensor("linear_16")]; tensor input_91 = add(x = input_83, y = input_89)[name = tensor("input_91")]; tensor x_41_axes_0 = const()[name = tensor("x_41_axes_0"), val = tensor([-1])]; tensor x_41 = layer_norm(axes = x_41_axes_0, beta = encoder_layers_1_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_1_norm_conv_weight, x = input_91)[name = tensor("x_41")]; tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; tensor input_93 = transpose(perm = input_93_perm_0, x = x_41)[name = tensor("transpose_299")]; tensor input_95 = conv(bias = encoder_layers_1_conv_pointwise_conv1_bias, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = encoder_layers_1_conv_pointwise_conv1_weight_quantized, x = input_93)[name = tensor("input_95")]; tensor x_43_split_num_splits_0 = const()[name = tensor("x_43_split_num_splits_0"), val = tensor(2)]; tensor x_43_split_axis_0 = const()[name = tensor("x_43_split_axis_0"), val = tensor(1)]; tensor x_43_split_0, tensor x_43_split_1 = split(axis = x_43_split_axis_0, num_splits = x_43_split_num_splits_0, x = input_95)[name = tensor("x_43_split")]; tensor x_43_split_1_sigmoid = sigmoid(x = x_43_split_1)[name = tensor("x_43_split_1_sigmoid")]; tensor x_43 = mul(x = x_43_split_0, y = x_43_split_1_sigmoid)[name = tensor("x_43")]; tensor input_97 = select(a = var_13, b = x_43, cond = var_339)[name = tensor("input_97")]; tensor const_27 = const()[name = tensor("const_27"), val = tensor(0x0p+0)]; tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; tensor input_99 = pad(constant_val = const_27, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97)[name = tensor("input_99")]; tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(1024)]; tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; tensor const_250_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_250_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588465536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588475904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588474816)))]; tensor const_251 = const()[name = tensor("const_251"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588480064)))]; tensor input_103 = conv(bias = const_251, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_250_quantized, x = input_99)[name = tensor("input_103")]; tensor input_105 = silu(x = input_103)[name = tensor("input_105")]; tensor x_45_pad_type_0 = const()[name = tensor("x_45_pad_type_0"), val = tensor("valid")]; tensor x_45_strides_0 = const()[name = tensor("x_45_strides_0"), val = tensor([1])]; tensor x_45_pad_0 = const()[name = tensor("x_45_pad_0"), val = tensor([0, 0])]; tensor x_45_dilations_0 = const()[name = tensor("x_45_dilations_0"), val = tensor([1])]; tensor x_45_groups_0 = const()[name = tensor("x_45_groups_0"), val = tensor(1)]; tensor x_45 = conv(bias = encoder_layers_1_conv_pointwise_conv2_bias, dilations = x_45_dilations_0, groups = x_45_groups_0, pad = x_45_pad_0, pad_type = x_45_pad_type_0, strides = x_45_strides_0, weight = encoder_layers_1_conv_pointwise_conv2_weight_quantized, x = input_105)[name = tensor("x_45")]; tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; tensor input_107 = transpose(perm = input_107_perm_0, x = x_45)[name = tensor("transpose_298")]; tensor input_109 = add(x = input_91, y = input_107)[name = tensor("input_109")]; tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; tensor input_111 = layer_norm(axes = input_111_axes_0, beta = encoder_layers_1_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_1_norm_feed_forward2_weight, x = input_109)[name = tensor("input_111")]; tensor input_113 = linear(bias = encoder_layers_1_feed_forward2_linear1_bias, weight = encoder_layers_1_feed_forward2_linear1_weight_quantized, x = input_111)[name = tensor("linear_17")]; tensor input_115 = silu(x = input_113)[name = tensor("input_115")]; tensor input_119 = linear(bias = encoder_layers_1_feed_forward2_linear2_bias, weight = encoder_layers_1_feed_forward2_linear2_weight_quantized, x = input_115)[name = tensor("linear_18")]; tensor var_555 = const()[name = tensor("op_555"), val = tensor(0x1p-1)]; tensor var_556 = mul(x = input_119, y = var_555)[name = tensor("op_556")]; tensor input_121 = add(x = input_109, y = var_556)[name = tensor("input_121")]; tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; tensor input_123 = layer_norm(axes = input_123_axes_0, beta = encoder_layers_1_norm_out_bias, epsilon = var_11, gamma = encoder_layers_1_norm_out_weight, x = input_121)[name = tensor("input_123")]; tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; tensor input_125 = layer_norm(axes = input_125_axes_0, beta = encoder_layers_2_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_2_norm_feed_forward1_weight, x = input_123)[name = tensor("input_125")]; tensor input_127 = linear(bias = encoder_layers_2_feed_forward1_linear1_bias, weight = encoder_layers_2_feed_forward1_linear1_weight_quantized, x = input_125)[name = tensor("linear_19")]; tensor input_129 = silu(x = input_127)[name = tensor("input_129")]; tensor input_133 = linear(bias = encoder_layers_2_feed_forward1_linear2_bias, weight = encoder_layers_2_feed_forward1_linear2_weight_quantized, x = input_129)[name = tensor("linear_20")]; tensor var_586 = const()[name = tensor("op_586"), val = tensor(0x1p-1)]; tensor var_587 = mul(x = input_133, y = var_586)[name = tensor("op_587")]; tensor input_135 = add(x = input_123, y = var_587)[name = tensor("input_135")]; tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; tensor query_5 = layer_norm(axes = query_5_axes_0, beta = encoder_layers_2_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_2_norm_self_att_weight, x = input_135)[name = tensor("query_5")]; tensor var_603 = linear(bias = encoder_layers_2_self_attn_linear_q_bias, weight = encoder_layers_2_self_attn_linear_q_weight_quantized, x = query_5)[name = tensor("linear_21")]; tensor var_604 = const()[name = tensor("op_604"), val = tensor([1, -1, 8, 128])]; tensor q_13 = reshape(shape = var_604, x = var_603)[name = tensor("q_13")]; tensor var_608 = linear(bias = encoder_layers_2_self_attn_linear_k_bias, weight = encoder_layers_2_self_attn_linear_k_weight_quantized, x = query_5)[name = tensor("linear_22")]; tensor var_609 = const()[name = tensor("op_609"), val = tensor([1, -1, 8, 128])]; tensor k_9 = reshape(shape = var_609, x = var_608)[name = tensor("k_9")]; tensor var_613 = linear(bias = encoder_layers_2_self_attn_linear_v_bias, weight = encoder_layers_2_self_attn_linear_v_weight_quantized, x = query_5)[name = tensor("linear_23")]; tensor var_614 = const()[name = tensor("op_614"), val = tensor([1, -1, 8, 128])]; tensor v_5 = reshape(shape = var_614, x = var_613)[name = tensor("v_5")]; tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_626 = add(x = q_13, y = encoder_layers_2_self_attn_pos_bias_u)[name = tensor("op_626")]; tensor var_628 = add(x = q_13, y = encoder_layers_2_self_attn_pos_bias_v)[name = tensor("op_628")]; tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_630_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_630_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588484224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588868736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588868288)))]; tensor x_53_transpose_x_0 = const()[name = tensor("x_53_transpose_x_0"), val = tensor(false)]; tensor x_53_transpose_y_0 = const()[name = tensor("x_53_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_5 = transpose(perm = q_with_bias_v_5_perm_0, x = var_628)[name = tensor("transpose_297")]; tensor x_53 = matmul(transpose_x = x_53_transpose_x_0, transpose_y = x_53_transpose_y_0, x = q_with_bias_v_5, y = op_630_quantized)[name = tensor("x_53")]; tensor const_34 = const()[name = tensor("const_34"), val = tensor(0x0p+0)]; tensor x_55_pad_0 = const()[name = tensor("x_55_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_55_mode_0 = const()[name = tensor("x_55_mode_0"), val = tensor("constant")]; tensor x_55 = pad(constant_val = const_34, mode = x_55_mode_0, pad = x_55_pad_0, x = x_53)[name = tensor("x_55")]; tensor var_638 = const()[name = tensor("op_638"), val = tensor([1, 8, -1, 188])]; tensor x_57 = reshape(shape = var_638, x = x_55)[name = tensor("x_57")]; tensor var_642_begin_0 = const()[name = tensor("op_642_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_642_end_0 = const()[name = tensor("op_642_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_642_end_mask_0 = const()[name = tensor("op_642_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_642 = slice_by_index(begin = var_642_begin_0, end = var_642_end_0, end_mask = var_642_end_mask_0, x = x_57)[name = tensor("op_642")]; tensor var_643 = const()[name = tensor("op_643"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_9 = reshape(shape = var_643, x = var_642)[name = tensor("matrix_bd_9")]; tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; tensor transpose_100_perm_0 = const()[name = tensor("transpose_100_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_101_perm_0 = const()[name = tensor("transpose_101_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_101 = transpose(perm = transpose_101_perm_0, x = k_9)[name = tensor("transpose_295")]; tensor transpose_100 = transpose(perm = transpose_100_perm_0, x = var_626)[name = tensor("transpose_296")]; tensor matrix_ac_5 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_100, y = transpose_101)[name = tensor("matrix_ac_5")]; tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_11 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9)[name = tensor("matrix_bd_11")]; tensor var_652 = add(x = matrix_ac_5, y = matrix_bd_11)[name = tensor("op_652")]; tensor _inversed_scores_9_y_0 = const()[name = tensor("_inversed_scores_9_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_9 = mul(x = var_652, y = _inversed_scores_9_y_0)[name = tensor("_inversed_scores_9")]; tensor scores_11 = select(a = var_14, b = _inversed_scores_9, cond = mask_3)[name = tensor("scores_11")]; tensor var_658 = softmax(axis = var_32, x = scores_11)[name = tensor("op_658")]; tensor input_137 = select(a = var_13, b = var_658, cond = mask_3)[name = tensor("input_137")]; tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; tensor value_7 = transpose(perm = value_7_perm_0, x = v_5)[name = tensor("transpose_294")]; tensor x_59 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = input_137, y = value_7)[name = tensor("x_59")]; tensor var_662_perm_0 = const()[name = tensor("op_662_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_663 = const()[name = tensor("op_663"), val = tensor([1, -1, 1024])]; tensor var_662 = transpose(perm = var_662_perm_0, x = x_59)[name = tensor("transpose_293")]; tensor input_139 = reshape(shape = var_663, x = var_662)[name = tensor("input_139")]; tensor input_141 = linear(bias = encoder_layers_2_self_attn_linear_out_bias, weight = encoder_layers_2_self_attn_linear_out_weight_quantized, x = input_139)[name = tensor("linear_25")]; tensor input_143 = add(x = input_135, y = input_141)[name = tensor("input_143")]; tensor x_63_axes_0 = const()[name = tensor("x_63_axes_0"), val = tensor([-1])]; tensor x_63 = layer_norm(axes = x_63_axes_0, beta = encoder_layers_2_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_2_norm_conv_weight, x = input_143)[name = tensor("x_63")]; tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; tensor input_145 = transpose(perm = input_145_perm_0, x = x_63)[name = tensor("transpose_292")]; tensor input_147 = conv(bias = encoder_layers_2_conv_pointwise_conv1_bias, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = encoder_layers_2_conv_pointwise_conv1_weight_quantized, x = input_145)[name = tensor("input_147")]; tensor x_65_split_num_splits_0 = const()[name = tensor("x_65_split_num_splits_0"), val = tensor(2)]; tensor x_65_split_axis_0 = const()[name = tensor("x_65_split_axis_0"), val = tensor(1)]; tensor x_65_split_0, tensor x_65_split_1 = split(axis = x_65_split_axis_0, num_splits = x_65_split_num_splits_0, x = input_147)[name = tensor("x_65_split")]; tensor x_65_split_1_sigmoid = sigmoid(x = x_65_split_1)[name = tensor("x_65_split_1_sigmoid")]; tensor x_65 = mul(x = x_65_split_0, y = x_65_split_1_sigmoid)[name = tensor("x_65")]; tensor input_149 = select(a = var_13, b = x_65, cond = var_339)[name = tensor("input_149")]; tensor const_37 = const()[name = tensor("const_37"), val = tensor(0x0p+0)]; tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; tensor input_151 = pad(constant_val = const_37, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149)[name = tensor("input_151")]; tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(1024)]; tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; tensor const_252_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_252_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588870336))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588880704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588879616)))]; tensor const_253 = const()[name = tensor("const_253"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588884864)))]; tensor input_155 = conv(bias = const_253, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_252_quantized, x = input_151)[name = tensor("input_155")]; tensor input_157 = silu(x = input_155)[name = tensor("input_157")]; tensor x_67_pad_type_0 = const()[name = tensor("x_67_pad_type_0"), val = tensor("valid")]; tensor x_67_strides_0 = const()[name = tensor("x_67_strides_0"), val = tensor([1])]; tensor x_67_pad_0 = const()[name = tensor("x_67_pad_0"), val = tensor([0, 0])]; tensor x_67_dilations_0 = const()[name = tensor("x_67_dilations_0"), val = tensor([1])]; tensor x_67_groups_0 = const()[name = tensor("x_67_groups_0"), val = tensor(1)]; tensor x_67 = conv(bias = encoder_layers_2_conv_pointwise_conv2_bias, dilations = x_67_dilations_0, groups = x_67_groups_0, pad = x_67_pad_0, pad_type = x_67_pad_type_0, strides = x_67_strides_0, weight = encoder_layers_2_conv_pointwise_conv2_weight_quantized, x = input_157)[name = tensor("x_67")]; tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; tensor input_159 = transpose(perm = input_159_perm_0, x = x_67)[name = tensor("transpose_291")]; tensor input_161 = add(x = input_143, y = input_159)[name = tensor("input_161")]; tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; tensor input_163 = layer_norm(axes = input_163_axes_0, beta = encoder_layers_2_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_2_norm_feed_forward2_weight, x = input_161)[name = tensor("input_163")]; tensor input_165 = linear(bias = encoder_layers_2_feed_forward2_linear1_bias, weight = encoder_layers_2_feed_forward2_linear1_weight_quantized, x = input_163)[name = tensor("linear_26")]; tensor input_167 = silu(x = input_165)[name = tensor("input_167")]; tensor input_171 = linear(bias = encoder_layers_2_feed_forward2_linear2_bias, weight = encoder_layers_2_feed_forward2_linear2_weight_quantized, x = input_167)[name = tensor("linear_27")]; tensor var_729 = const()[name = tensor("op_729"), val = tensor(0x1p-1)]; tensor var_730 = mul(x = input_171, y = var_729)[name = tensor("op_730")]; tensor input_173 = add(x = input_161, y = var_730)[name = tensor("input_173")]; tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; tensor input_175 = layer_norm(axes = input_175_axes_0, beta = encoder_layers_2_norm_out_bias, epsilon = var_11, gamma = encoder_layers_2_norm_out_weight, x = input_173)[name = tensor("input_175")]; tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; tensor input_177 = layer_norm(axes = input_177_axes_0, beta = encoder_layers_3_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_3_norm_feed_forward1_weight, x = input_175)[name = tensor("input_177")]; tensor input_179 = linear(bias = encoder_layers_3_feed_forward1_linear1_bias, weight = encoder_layers_3_feed_forward1_linear1_weight_quantized, x = input_177)[name = tensor("linear_28")]; tensor input_181 = silu(x = input_179)[name = tensor("input_181")]; tensor input_185 = linear(bias = encoder_layers_3_feed_forward1_linear2_bias, weight = encoder_layers_3_feed_forward1_linear2_weight_quantized, x = input_181)[name = tensor("linear_29")]; tensor var_760 = const()[name = tensor("op_760"), val = tensor(0x1p-1)]; tensor var_761 = mul(x = input_185, y = var_760)[name = tensor("op_761")]; tensor input_187 = add(x = input_175, y = var_761)[name = tensor("input_187")]; tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; tensor query_7 = layer_norm(axes = query_7_axes_0, beta = encoder_layers_3_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_3_norm_self_att_weight, x = input_187)[name = tensor("query_7")]; tensor var_777 = linear(bias = encoder_layers_3_self_attn_linear_q_bias, weight = encoder_layers_3_self_attn_linear_q_weight_quantized, x = query_7)[name = tensor("linear_30")]; tensor var_778 = const()[name = tensor("op_778"), val = tensor([1, -1, 8, 128])]; tensor q_19 = reshape(shape = var_778, x = var_777)[name = tensor("q_19")]; tensor var_782 = linear(bias = encoder_layers_3_self_attn_linear_k_bias, weight = encoder_layers_3_self_attn_linear_k_weight_quantized, x = query_7)[name = tensor("linear_31")]; tensor var_783 = const()[name = tensor("op_783"), val = tensor([1, -1, 8, 128])]; tensor k_13 = reshape(shape = var_783, x = var_782)[name = tensor("k_13")]; tensor var_787 = linear(bias = encoder_layers_3_self_attn_linear_v_bias, weight = encoder_layers_3_self_attn_linear_v_weight_quantized, x = query_7)[name = tensor("linear_32")]; tensor var_788 = const()[name = tensor("op_788"), val = tensor([1, -1, 8, 128])]; tensor v_7 = reshape(shape = var_788, x = var_787)[name = tensor("v_7")]; tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_800 = add(x = q_19, y = encoder_layers_3_self_attn_pos_bias_u)[name = tensor("op_800")]; tensor var_802 = add(x = q_19, y = encoder_layers_3_self_attn_pos_bias_v)[name = tensor("op_802")]; tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_804_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_804_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588889024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589273536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589273088)))]; tensor x_75_transpose_x_0 = const()[name = tensor("x_75_transpose_x_0"), val = tensor(false)]; tensor x_75_transpose_y_0 = const()[name = tensor("x_75_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_7 = transpose(perm = q_with_bias_v_7_perm_0, x = var_802)[name = tensor("transpose_290")]; tensor x_75 = matmul(transpose_x = x_75_transpose_x_0, transpose_y = x_75_transpose_y_0, x = q_with_bias_v_7, y = op_804_quantized)[name = tensor("x_75")]; tensor const_44 = const()[name = tensor("const_44"), val = tensor(0x0p+0)]; tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_77_mode_0 = const()[name = tensor("x_77_mode_0"), val = tensor("constant")]; tensor x_77 = pad(constant_val = const_44, mode = x_77_mode_0, pad = x_77_pad_0, x = x_75)[name = tensor("x_77")]; tensor var_812 = const()[name = tensor("op_812"), val = tensor([1, 8, -1, 188])]; tensor x_79 = reshape(shape = var_812, x = x_77)[name = tensor("x_79")]; tensor var_816_begin_0 = const()[name = tensor("op_816_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_816_end_0 = const()[name = tensor("op_816_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_816_end_mask_0 = const()[name = tensor("op_816_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_816 = slice_by_index(begin = var_816_begin_0, end = var_816_end_0, end_mask = var_816_end_mask_0, x = x_79)[name = tensor("op_816")]; tensor var_817 = const()[name = tensor("op_817"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_13 = reshape(shape = var_817, x = var_816)[name = tensor("matrix_bd_13")]; tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; tensor transpose_102_perm_0 = const()[name = tensor("transpose_102_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_103_perm_0 = const()[name = tensor("transpose_103_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_103 = transpose(perm = transpose_103_perm_0, x = k_13)[name = tensor("transpose_288")]; tensor transpose_102 = transpose(perm = transpose_102_perm_0, x = var_800)[name = tensor("transpose_289")]; tensor matrix_ac_7 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_102, y = transpose_103)[name = tensor("matrix_ac_7")]; tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_15 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13)[name = tensor("matrix_bd_15")]; tensor var_826 = add(x = matrix_ac_7, y = matrix_bd_15)[name = tensor("op_826")]; tensor _inversed_scores_13_y_0 = const()[name = tensor("_inversed_scores_13_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_13 = mul(x = var_826, y = _inversed_scores_13_y_0)[name = tensor("_inversed_scores_13")]; tensor scores_15 = select(a = var_14, b = _inversed_scores_13, cond = mask_3)[name = tensor("scores_15")]; tensor var_832 = softmax(axis = var_32, x = scores_15)[name = tensor("op_832")]; tensor input_189 = select(a = var_13, b = var_832, cond = mask_3)[name = tensor("input_189")]; tensor x_81_transpose_x_0 = const()[name = tensor("x_81_transpose_x_0"), val = tensor(false)]; tensor x_81_transpose_y_0 = const()[name = tensor("x_81_transpose_y_0"), val = tensor(false)]; tensor value_9 = transpose(perm = value_9_perm_0, x = v_7)[name = tensor("transpose_287")]; tensor x_81 = matmul(transpose_x = x_81_transpose_x_0, transpose_y = x_81_transpose_y_0, x = input_189, y = value_9)[name = tensor("x_81")]; tensor var_836_perm_0 = const()[name = tensor("op_836_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_837 = const()[name = tensor("op_837"), val = tensor([1, -1, 1024])]; tensor var_836 = transpose(perm = var_836_perm_0, x = x_81)[name = tensor("transpose_286")]; tensor input_191 = reshape(shape = var_837, x = var_836)[name = tensor("input_191")]; tensor input_193 = linear(bias = encoder_layers_3_self_attn_linear_out_bias, weight = encoder_layers_3_self_attn_linear_out_weight_quantized, x = input_191)[name = tensor("linear_34")]; tensor input_195 = add(x = input_187, y = input_193)[name = tensor("input_195")]; tensor x_85_axes_0 = const()[name = tensor("x_85_axes_0"), val = tensor([-1])]; tensor x_85 = layer_norm(axes = x_85_axes_0, beta = encoder_layers_3_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_3_norm_conv_weight, x = input_195)[name = tensor("x_85")]; tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; tensor input_197 = transpose(perm = input_197_perm_0, x = x_85)[name = tensor("transpose_285")]; tensor input_199 = conv(bias = encoder_layers_3_conv_pointwise_conv1_bias, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = encoder_layers_3_conv_pointwise_conv1_weight_quantized, x = input_197)[name = tensor("input_199")]; tensor x_87_split_num_splits_0 = const()[name = tensor("x_87_split_num_splits_0"), val = tensor(2)]; tensor x_87_split_axis_0 = const()[name = tensor("x_87_split_axis_0"), val = tensor(1)]; tensor x_87_split_0, tensor x_87_split_1 = split(axis = x_87_split_axis_0, num_splits = x_87_split_num_splits_0, x = input_199)[name = tensor("x_87_split")]; tensor x_87_split_1_sigmoid = sigmoid(x = x_87_split_1)[name = tensor("x_87_split_1_sigmoid")]; tensor x_87 = mul(x = x_87_split_0, y = x_87_split_1_sigmoid)[name = tensor("x_87")]; tensor input_201 = select(a = var_13, b = x_87, cond = var_339)[name = tensor("input_201")]; tensor const_47 = const()[name = tensor("const_47"), val = tensor(0x0p+0)]; tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; tensor input_203 = pad(constant_val = const_47, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201)[name = tensor("input_203")]; tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(1024)]; tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; tensor const_254_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_254_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589275136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589285504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589284416)))]; tensor const_255 = const()[name = tensor("const_255"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589289664)))]; tensor input_207 = conv(bias = const_255, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_254_quantized, x = input_203)[name = tensor("input_207")]; tensor input_209 = silu(x = input_207)[name = tensor("input_209")]; tensor x_89_pad_type_0 = const()[name = tensor("x_89_pad_type_0"), val = tensor("valid")]; tensor x_89_strides_0 = const()[name = tensor("x_89_strides_0"), val = tensor([1])]; tensor x_89_pad_0 = const()[name = tensor("x_89_pad_0"), val = tensor([0, 0])]; tensor x_89_dilations_0 = const()[name = tensor("x_89_dilations_0"), val = tensor([1])]; tensor x_89_groups_0 = const()[name = tensor("x_89_groups_0"), val = tensor(1)]; tensor x_89 = conv(bias = encoder_layers_3_conv_pointwise_conv2_bias, dilations = x_89_dilations_0, groups = x_89_groups_0, pad = x_89_pad_0, pad_type = x_89_pad_type_0, strides = x_89_strides_0, weight = encoder_layers_3_conv_pointwise_conv2_weight_quantized, x = input_209)[name = tensor("x_89")]; tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; tensor input_211 = transpose(perm = input_211_perm_0, x = x_89)[name = tensor("transpose_284")]; tensor input_213 = add(x = input_195, y = input_211)[name = tensor("input_213")]; tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; tensor input_215 = layer_norm(axes = input_215_axes_0, beta = encoder_layers_3_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_3_norm_feed_forward2_weight, x = input_213)[name = tensor("input_215")]; tensor input_217 = linear(bias = encoder_layers_3_feed_forward2_linear1_bias, weight = encoder_layers_3_feed_forward2_linear1_weight_quantized, x = input_215)[name = tensor("linear_35")]; tensor input_219 = silu(x = input_217)[name = tensor("input_219")]; tensor input_223 = linear(bias = encoder_layers_3_feed_forward2_linear2_bias, weight = encoder_layers_3_feed_forward2_linear2_weight_quantized, x = input_219)[name = tensor("linear_36")]; tensor var_903 = const()[name = tensor("op_903"), val = tensor(0x1p-1)]; tensor var_904 = mul(x = input_223, y = var_903)[name = tensor("op_904")]; tensor input_225 = add(x = input_213, y = var_904)[name = tensor("input_225")]; tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; tensor input_227 = layer_norm(axes = input_227_axes_0, beta = encoder_layers_3_norm_out_bias, epsilon = var_11, gamma = encoder_layers_3_norm_out_weight, x = input_225)[name = tensor("input_227")]; tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; tensor input_229 = layer_norm(axes = input_229_axes_0, beta = encoder_layers_4_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_4_norm_feed_forward1_weight, x = input_227)[name = tensor("input_229")]; tensor input_231 = linear(bias = encoder_layers_4_feed_forward1_linear1_bias, weight = encoder_layers_4_feed_forward1_linear1_weight_quantized, x = input_229)[name = tensor("linear_37")]; tensor input_233 = silu(x = input_231)[name = tensor("input_233")]; tensor input_237 = linear(bias = encoder_layers_4_feed_forward1_linear2_bias, weight = encoder_layers_4_feed_forward1_linear2_weight_quantized, x = input_233)[name = tensor("linear_38")]; tensor var_934 = const()[name = tensor("op_934"), val = tensor(0x1p-1)]; tensor var_935 = mul(x = input_237, y = var_934)[name = tensor("op_935")]; tensor input_239 = add(x = input_227, y = var_935)[name = tensor("input_239")]; tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; tensor query_9 = layer_norm(axes = query_9_axes_0, beta = encoder_layers_4_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_4_norm_self_att_weight, x = input_239)[name = tensor("query_9")]; tensor var_951 = linear(bias = encoder_layers_4_self_attn_linear_q_bias, weight = encoder_layers_4_self_attn_linear_q_weight_quantized, x = query_9)[name = tensor("linear_39")]; tensor var_952 = const()[name = tensor("op_952"), val = tensor([1, -1, 8, 128])]; tensor q_25 = reshape(shape = var_952, x = var_951)[name = tensor("q_25")]; tensor var_956 = linear(bias = encoder_layers_4_self_attn_linear_k_bias, weight = encoder_layers_4_self_attn_linear_k_weight_quantized, x = query_9)[name = tensor("linear_40")]; tensor var_957 = const()[name = tensor("op_957"), val = tensor([1, -1, 8, 128])]; tensor k_17 = reshape(shape = var_957, x = var_956)[name = tensor("k_17")]; tensor var_961 = linear(bias = encoder_layers_4_self_attn_linear_v_bias, weight = encoder_layers_4_self_attn_linear_v_weight_quantized, x = query_9)[name = tensor("linear_41")]; tensor var_962 = const()[name = tensor("op_962"), val = tensor([1, -1, 8, 128])]; tensor v_9 = reshape(shape = var_962, x = var_961)[name = tensor("v_9")]; tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_974 = add(x = q_25, y = encoder_layers_4_self_attn_pos_bias_u)[name = tensor("op_974")]; tensor var_976 = add(x = q_25, y = encoder_layers_4_self_attn_pos_bias_v)[name = tensor("op_976")]; tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_978_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_978_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589293824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589678336))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589677888)))]; tensor x_97_transpose_x_0 = const()[name = tensor("x_97_transpose_x_0"), val = tensor(false)]; tensor x_97_transpose_y_0 = const()[name = tensor("x_97_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_9 = transpose(perm = q_with_bias_v_9_perm_0, x = var_976)[name = tensor("transpose_283")]; tensor x_97 = matmul(transpose_x = x_97_transpose_x_0, transpose_y = x_97_transpose_y_0, x = q_with_bias_v_9, y = op_978_quantized)[name = tensor("x_97")]; tensor const_54 = const()[name = tensor("const_54"), val = tensor(0x0p+0)]; tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_99_mode_0 = const()[name = tensor("x_99_mode_0"), val = tensor("constant")]; tensor x_99 = pad(constant_val = const_54, mode = x_99_mode_0, pad = x_99_pad_0, x = x_97)[name = tensor("x_99")]; tensor var_986 = const()[name = tensor("op_986"), val = tensor([1, 8, -1, 188])]; tensor x_101 = reshape(shape = var_986, x = x_99)[name = tensor("x_101")]; tensor var_990_begin_0 = const()[name = tensor("op_990_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_990_end_0 = const()[name = tensor("op_990_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_990_end_mask_0 = const()[name = tensor("op_990_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_990 = slice_by_index(begin = var_990_begin_0, end = var_990_end_0, end_mask = var_990_end_mask_0, x = x_101)[name = tensor("op_990")]; tensor var_991 = const()[name = tensor("op_991"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_17 = reshape(shape = var_991, x = var_990)[name = tensor("matrix_bd_17")]; tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; tensor transpose_104_perm_0 = const()[name = tensor("transpose_104_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_105_perm_0 = const()[name = tensor("transpose_105_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_105 = transpose(perm = transpose_105_perm_0, x = k_17)[name = tensor("transpose_281")]; tensor transpose_104 = transpose(perm = transpose_104_perm_0, x = var_974)[name = tensor("transpose_282")]; tensor matrix_ac_9 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_104, y = transpose_105)[name = tensor("matrix_ac_9")]; tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_19 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17)[name = tensor("matrix_bd_19")]; tensor var_1000 = add(x = matrix_ac_9, y = matrix_bd_19)[name = tensor("op_1000")]; tensor _inversed_scores_17_y_0 = const()[name = tensor("_inversed_scores_17_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_17 = mul(x = var_1000, y = _inversed_scores_17_y_0)[name = tensor("_inversed_scores_17")]; tensor scores_19 = select(a = var_14, b = _inversed_scores_17, cond = mask_3)[name = tensor("scores_19")]; tensor var_1006 = softmax(axis = var_32, x = scores_19)[name = tensor("op_1006")]; tensor input_241 = select(a = var_13, b = var_1006, cond = mask_3)[name = tensor("input_241")]; tensor x_103_transpose_x_0 = const()[name = tensor("x_103_transpose_x_0"), val = tensor(false)]; tensor x_103_transpose_y_0 = const()[name = tensor("x_103_transpose_y_0"), val = tensor(false)]; tensor value_11 = transpose(perm = value_11_perm_0, x = v_9)[name = tensor("transpose_280")]; tensor x_103 = matmul(transpose_x = x_103_transpose_x_0, transpose_y = x_103_transpose_y_0, x = input_241, y = value_11)[name = tensor("x_103")]; tensor var_1010_perm_0 = const()[name = tensor("op_1010_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1011 = const()[name = tensor("op_1011"), val = tensor([1, -1, 1024])]; tensor var_1010 = transpose(perm = var_1010_perm_0, x = x_103)[name = tensor("transpose_279")]; tensor input_243 = reshape(shape = var_1011, x = var_1010)[name = tensor("input_243")]; tensor input_245 = linear(bias = encoder_layers_4_self_attn_linear_out_bias, weight = encoder_layers_4_self_attn_linear_out_weight_quantized, x = input_243)[name = tensor("linear_43")]; tensor input_247 = add(x = input_239, y = input_245)[name = tensor("input_247")]; tensor x_107_axes_0 = const()[name = tensor("x_107_axes_0"), val = tensor([-1])]; tensor x_107 = layer_norm(axes = x_107_axes_0, beta = encoder_layers_4_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_4_norm_conv_weight, x = input_247)[name = tensor("x_107")]; tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; tensor input_249 = transpose(perm = input_249_perm_0, x = x_107)[name = tensor("transpose_278")]; tensor input_251 = conv(bias = encoder_layers_4_conv_pointwise_conv1_bias, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = encoder_layers_4_conv_pointwise_conv1_weight_quantized, x = input_249)[name = tensor("input_251")]; tensor x_109_split_num_splits_0 = const()[name = tensor("x_109_split_num_splits_0"), val = tensor(2)]; tensor x_109_split_axis_0 = const()[name = tensor("x_109_split_axis_0"), val = tensor(1)]; tensor x_109_split_0, tensor x_109_split_1 = split(axis = x_109_split_axis_0, num_splits = x_109_split_num_splits_0, x = input_251)[name = tensor("x_109_split")]; tensor x_109_split_1_sigmoid = sigmoid(x = x_109_split_1)[name = tensor("x_109_split_1_sigmoid")]; tensor x_109 = mul(x = x_109_split_0, y = x_109_split_1_sigmoid)[name = tensor("x_109")]; tensor input_253 = select(a = var_13, b = x_109, cond = var_339)[name = tensor("input_253")]; tensor const_57 = const()[name = tensor("const_57"), val = tensor(0x0p+0)]; tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; tensor input_255 = pad(constant_val = const_57, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253)[name = tensor("input_255")]; tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(1024)]; tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; tensor const_256_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_256_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589679936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589690304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589689216)))]; tensor const_257 = const()[name = tensor("const_257"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589694464)))]; tensor input_259 = conv(bias = const_257, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_256_quantized, x = input_255)[name = tensor("input_259")]; tensor input_261 = silu(x = input_259)[name = tensor("input_261")]; tensor x_111_pad_type_0 = const()[name = tensor("x_111_pad_type_0"), val = tensor("valid")]; tensor x_111_strides_0 = const()[name = tensor("x_111_strides_0"), val = tensor([1])]; tensor x_111_pad_0 = const()[name = tensor("x_111_pad_0"), val = tensor([0, 0])]; tensor x_111_dilations_0 = const()[name = tensor("x_111_dilations_0"), val = tensor([1])]; tensor x_111_groups_0 = const()[name = tensor("x_111_groups_0"), val = tensor(1)]; tensor x_111 = conv(bias = encoder_layers_4_conv_pointwise_conv2_bias, dilations = x_111_dilations_0, groups = x_111_groups_0, pad = x_111_pad_0, pad_type = x_111_pad_type_0, strides = x_111_strides_0, weight = encoder_layers_4_conv_pointwise_conv2_weight_quantized, x = input_261)[name = tensor("x_111")]; tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; tensor input_263 = transpose(perm = input_263_perm_0, x = x_111)[name = tensor("transpose_277")]; tensor input_265 = add(x = input_247, y = input_263)[name = tensor("input_265")]; tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; tensor input_267 = layer_norm(axes = input_267_axes_0, beta = encoder_layers_4_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_4_norm_feed_forward2_weight, x = input_265)[name = tensor("input_267")]; tensor input_269 = linear(bias = encoder_layers_4_feed_forward2_linear1_bias, weight = encoder_layers_4_feed_forward2_linear1_weight_quantized, x = input_267)[name = tensor("linear_44")]; tensor input_271 = silu(x = input_269)[name = tensor("input_271")]; tensor input_275 = linear(bias = encoder_layers_4_feed_forward2_linear2_bias, weight = encoder_layers_4_feed_forward2_linear2_weight_quantized, x = input_271)[name = tensor("linear_45")]; tensor var_1077 = const()[name = tensor("op_1077"), val = tensor(0x1p-1)]; tensor var_1078 = mul(x = input_275, y = var_1077)[name = tensor("op_1078")]; tensor input_277 = add(x = input_265, y = var_1078)[name = tensor("input_277")]; tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; tensor input_279 = layer_norm(axes = input_279_axes_0, beta = encoder_layers_4_norm_out_bias, epsilon = var_11, gamma = encoder_layers_4_norm_out_weight, x = input_277)[name = tensor("input_279")]; tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; tensor input_281 = layer_norm(axes = input_281_axes_0, beta = encoder_layers_5_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_5_norm_feed_forward1_weight, x = input_279)[name = tensor("input_281")]; tensor input_283 = linear(bias = encoder_layers_5_feed_forward1_linear1_bias, weight = encoder_layers_5_feed_forward1_linear1_weight_quantized, x = input_281)[name = tensor("linear_46")]; tensor input_285 = silu(x = input_283)[name = tensor("input_285")]; tensor input_289 = linear(bias = encoder_layers_5_feed_forward1_linear2_bias, weight = encoder_layers_5_feed_forward1_linear2_weight_quantized, x = input_285)[name = tensor("linear_47")]; tensor var_1108 = const()[name = tensor("op_1108"), val = tensor(0x1p-1)]; tensor var_1109 = mul(x = input_289, y = var_1108)[name = tensor("op_1109")]; tensor input_291 = add(x = input_279, y = var_1109)[name = tensor("input_291")]; tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; tensor query_11 = layer_norm(axes = query_11_axes_0, beta = encoder_layers_5_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_5_norm_self_att_weight, x = input_291)[name = tensor("query_11")]; tensor var_1125 = linear(bias = encoder_layers_5_self_attn_linear_q_bias, weight = encoder_layers_5_self_attn_linear_q_weight_quantized, x = query_11)[name = tensor("linear_48")]; tensor var_1126 = const()[name = tensor("op_1126"), val = tensor([1, -1, 8, 128])]; tensor q_31 = reshape(shape = var_1126, x = var_1125)[name = tensor("q_31")]; tensor var_1130 = linear(bias = encoder_layers_5_self_attn_linear_k_bias, weight = encoder_layers_5_self_attn_linear_k_weight_quantized, x = query_11)[name = tensor("linear_49")]; tensor var_1131 = const()[name = tensor("op_1131"), val = tensor([1, -1, 8, 128])]; tensor k_21 = reshape(shape = var_1131, x = var_1130)[name = tensor("k_21")]; tensor var_1135 = linear(bias = encoder_layers_5_self_attn_linear_v_bias, weight = encoder_layers_5_self_attn_linear_v_weight_quantized, x = query_11)[name = tensor("linear_50")]; tensor var_1136 = const()[name = tensor("op_1136"), val = tensor([1, -1, 8, 128])]; tensor v_11 = reshape(shape = var_1136, x = var_1135)[name = tensor("v_11")]; tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_1148 = add(x = q_31, y = encoder_layers_5_self_attn_pos_bias_u)[name = tensor("op_1148")]; tensor var_1150 = add(x = q_31, y = encoder_layers_5_self_attn_pos_bias_v)[name = tensor("op_1150")]; tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_1152_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1152_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589698624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590083136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590082688)))]; tensor x_119_transpose_x_0 = const()[name = tensor("x_119_transpose_x_0"), val = tensor(false)]; tensor x_119_transpose_y_0 = const()[name = tensor("x_119_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_11 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1150)[name = tensor("transpose_276")]; tensor x_119 = matmul(transpose_x = x_119_transpose_x_0, transpose_y = x_119_transpose_y_0, x = q_with_bias_v_11, y = op_1152_quantized)[name = tensor("x_119")]; tensor const_64 = const()[name = tensor("const_64"), val = tensor(0x0p+0)]; tensor x_121_pad_0 = const()[name = tensor("x_121_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_121_mode_0 = const()[name = tensor("x_121_mode_0"), val = tensor("constant")]; tensor x_121 = pad(constant_val = const_64, mode = x_121_mode_0, pad = x_121_pad_0, x = x_119)[name = tensor("x_121")]; tensor var_1160 = const()[name = tensor("op_1160"), val = tensor([1, 8, -1, 188])]; tensor x_123 = reshape(shape = var_1160, x = x_121)[name = tensor("x_123")]; tensor var_1164_begin_0 = const()[name = tensor("op_1164_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1164_end_0 = const()[name = tensor("op_1164_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1164_end_mask_0 = const()[name = tensor("op_1164_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1164 = slice_by_index(begin = var_1164_begin_0, end = var_1164_end_0, end_mask = var_1164_end_mask_0, x = x_123)[name = tensor("op_1164")]; tensor var_1165 = const()[name = tensor("op_1165"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_21 = reshape(shape = var_1165, x = var_1164)[name = tensor("matrix_bd_21")]; tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; tensor transpose_106_perm_0 = const()[name = tensor("transpose_106_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_107_perm_0 = const()[name = tensor("transpose_107_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_107 = transpose(perm = transpose_107_perm_0, x = k_21)[name = tensor("transpose_274")]; tensor transpose_106 = transpose(perm = transpose_106_perm_0, x = var_1148)[name = tensor("transpose_275")]; tensor matrix_ac_11 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_106, y = transpose_107)[name = tensor("matrix_ac_11")]; tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_23 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21)[name = tensor("matrix_bd_23")]; tensor var_1174 = add(x = matrix_ac_11, y = matrix_bd_23)[name = tensor("op_1174")]; tensor _inversed_scores_21_y_0 = const()[name = tensor("_inversed_scores_21_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_21 = mul(x = var_1174, y = _inversed_scores_21_y_0)[name = tensor("_inversed_scores_21")]; tensor scores_23 = select(a = var_14, b = _inversed_scores_21, cond = mask_3)[name = tensor("scores_23")]; tensor var_1180 = softmax(axis = var_32, x = scores_23)[name = tensor("op_1180")]; tensor input_293 = select(a = var_13, b = var_1180, cond = mask_3)[name = tensor("input_293")]; tensor x_125_transpose_x_0 = const()[name = tensor("x_125_transpose_x_0"), val = tensor(false)]; tensor x_125_transpose_y_0 = const()[name = tensor("x_125_transpose_y_0"), val = tensor(false)]; tensor value_13 = transpose(perm = value_13_perm_0, x = v_11)[name = tensor("transpose_273")]; tensor x_125 = matmul(transpose_x = x_125_transpose_x_0, transpose_y = x_125_transpose_y_0, x = input_293, y = value_13)[name = tensor("x_125")]; tensor var_1184_perm_0 = const()[name = tensor("op_1184_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1185 = const()[name = tensor("op_1185"), val = tensor([1, -1, 1024])]; tensor var_1184 = transpose(perm = var_1184_perm_0, x = x_125)[name = tensor("transpose_272")]; tensor input_295 = reshape(shape = var_1185, x = var_1184)[name = tensor("input_295")]; tensor input_297 = linear(bias = encoder_layers_5_self_attn_linear_out_bias, weight = encoder_layers_5_self_attn_linear_out_weight_quantized, x = input_295)[name = tensor("linear_52")]; tensor input_299 = add(x = input_291, y = input_297)[name = tensor("input_299")]; tensor x_129_axes_0 = const()[name = tensor("x_129_axes_0"), val = tensor([-1])]; tensor x_129 = layer_norm(axes = x_129_axes_0, beta = encoder_layers_5_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_5_norm_conv_weight, x = input_299)[name = tensor("x_129")]; tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; tensor input_301 = transpose(perm = input_301_perm_0, x = x_129)[name = tensor("transpose_271")]; tensor input_303 = conv(bias = encoder_layers_5_conv_pointwise_conv1_bias, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = encoder_layers_5_conv_pointwise_conv1_weight_quantized, x = input_301)[name = tensor("input_303")]; tensor x_131_split_num_splits_0 = const()[name = tensor("x_131_split_num_splits_0"), val = tensor(2)]; tensor x_131_split_axis_0 = const()[name = tensor("x_131_split_axis_0"), val = tensor(1)]; tensor x_131_split_0, tensor x_131_split_1 = split(axis = x_131_split_axis_0, num_splits = x_131_split_num_splits_0, x = input_303)[name = tensor("x_131_split")]; tensor x_131_split_1_sigmoid = sigmoid(x = x_131_split_1)[name = tensor("x_131_split_1_sigmoid")]; tensor x_131 = mul(x = x_131_split_0, y = x_131_split_1_sigmoid)[name = tensor("x_131")]; tensor input_305 = select(a = var_13, b = x_131, cond = var_339)[name = tensor("input_305")]; tensor const_67 = const()[name = tensor("const_67"), val = tensor(0x0p+0)]; tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; tensor input_307 = pad(constant_val = const_67, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305)[name = tensor("input_307")]; tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(1024)]; tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; tensor const_258_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_258_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590084736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590095104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590094016)))]; tensor const_259 = const()[name = tensor("const_259"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590099264)))]; tensor input_311 = conv(bias = const_259, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_258_quantized, x = input_307)[name = tensor("input_311")]; tensor input_313 = silu(x = input_311)[name = tensor("input_313")]; tensor x_133_pad_type_0 = const()[name = tensor("x_133_pad_type_0"), val = tensor("valid")]; tensor x_133_strides_0 = const()[name = tensor("x_133_strides_0"), val = tensor([1])]; tensor x_133_pad_0 = const()[name = tensor("x_133_pad_0"), val = tensor([0, 0])]; tensor x_133_dilations_0 = const()[name = tensor("x_133_dilations_0"), val = tensor([1])]; tensor x_133_groups_0 = const()[name = tensor("x_133_groups_0"), val = tensor(1)]; tensor x_133 = conv(bias = encoder_layers_5_conv_pointwise_conv2_bias, dilations = x_133_dilations_0, groups = x_133_groups_0, pad = x_133_pad_0, pad_type = x_133_pad_type_0, strides = x_133_strides_0, weight = encoder_layers_5_conv_pointwise_conv2_weight_quantized, x = input_313)[name = tensor("x_133")]; tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; tensor input_315 = transpose(perm = input_315_perm_0, x = x_133)[name = tensor("transpose_270")]; tensor input_317 = add(x = input_299, y = input_315)[name = tensor("input_317")]; tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; tensor input_319 = layer_norm(axes = input_319_axes_0, beta = encoder_layers_5_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_5_norm_feed_forward2_weight, x = input_317)[name = tensor("input_319")]; tensor input_321 = linear(bias = encoder_layers_5_feed_forward2_linear1_bias, weight = encoder_layers_5_feed_forward2_linear1_weight_quantized, x = input_319)[name = tensor("linear_53")]; tensor input_323 = silu(x = input_321)[name = tensor("input_323")]; tensor input_327 = linear(bias = encoder_layers_5_feed_forward2_linear2_bias, weight = encoder_layers_5_feed_forward2_linear2_weight_quantized, x = input_323)[name = tensor("linear_54")]; tensor var_1251 = const()[name = tensor("op_1251"), val = tensor(0x1p-1)]; tensor var_1252 = mul(x = input_327, y = var_1251)[name = tensor("op_1252")]; tensor input_329 = add(x = input_317, y = var_1252)[name = tensor("input_329")]; tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; tensor input_331 = layer_norm(axes = input_331_axes_0, beta = encoder_layers_5_norm_out_bias, epsilon = var_11, gamma = encoder_layers_5_norm_out_weight, x = input_329)[name = tensor("input_331")]; tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; tensor input_333 = layer_norm(axes = input_333_axes_0, beta = encoder_layers_6_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_6_norm_feed_forward1_weight, x = input_331)[name = tensor("input_333")]; tensor input_335 = linear(bias = encoder_layers_6_feed_forward1_linear1_bias, weight = encoder_layers_6_feed_forward1_linear1_weight_quantized, x = input_333)[name = tensor("linear_55")]; tensor input_337 = silu(x = input_335)[name = tensor("input_337")]; tensor input_341 = linear(bias = encoder_layers_6_feed_forward1_linear2_bias, weight = encoder_layers_6_feed_forward1_linear2_weight_quantized, x = input_337)[name = tensor("linear_56")]; tensor var_1282 = const()[name = tensor("op_1282"), val = tensor(0x1p-1)]; tensor var_1283 = mul(x = input_341, y = var_1282)[name = tensor("op_1283")]; tensor input_343 = add(x = input_331, y = var_1283)[name = tensor("input_343")]; tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; tensor query_13 = layer_norm(axes = query_13_axes_0, beta = encoder_layers_6_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_6_norm_self_att_weight, x = input_343)[name = tensor("query_13")]; tensor var_1299 = linear(bias = encoder_layers_6_self_attn_linear_q_bias, weight = encoder_layers_6_self_attn_linear_q_weight_quantized, x = query_13)[name = tensor("linear_57")]; tensor var_1300 = const()[name = tensor("op_1300"), val = tensor([1, -1, 8, 128])]; tensor q_37 = reshape(shape = var_1300, x = var_1299)[name = tensor("q_37")]; tensor var_1304 = linear(bias = encoder_layers_6_self_attn_linear_k_bias, weight = encoder_layers_6_self_attn_linear_k_weight_quantized, x = query_13)[name = tensor("linear_58")]; tensor var_1305 = const()[name = tensor("op_1305"), val = tensor([1, -1, 8, 128])]; tensor k_25 = reshape(shape = var_1305, x = var_1304)[name = tensor("k_25")]; tensor var_1309 = linear(bias = encoder_layers_6_self_attn_linear_v_bias, weight = encoder_layers_6_self_attn_linear_v_weight_quantized, x = query_13)[name = tensor("linear_59")]; tensor var_1310 = const()[name = tensor("op_1310"), val = tensor([1, -1, 8, 128])]; tensor v_13 = reshape(shape = var_1310, x = var_1309)[name = tensor("v_13")]; tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_1322 = add(x = q_37, y = encoder_layers_6_self_attn_pos_bias_u)[name = tensor("op_1322")]; tensor var_1324 = add(x = q_37, y = encoder_layers_6_self_attn_pos_bias_v)[name = tensor("op_1324")]; tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_1326_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1326_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590103424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590487936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590487488)))]; tensor x_141_transpose_x_0 = const()[name = tensor("x_141_transpose_x_0"), val = tensor(false)]; tensor x_141_transpose_y_0 = const()[name = tensor("x_141_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_13 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1324)[name = tensor("transpose_269")]; tensor x_141 = matmul(transpose_x = x_141_transpose_x_0, transpose_y = x_141_transpose_y_0, x = q_with_bias_v_13, y = op_1326_quantized)[name = tensor("x_141")]; tensor const_74 = const()[name = tensor("const_74"), val = tensor(0x0p+0)]; tensor x_143_pad_0 = const()[name = tensor("x_143_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_143_mode_0 = const()[name = tensor("x_143_mode_0"), val = tensor("constant")]; tensor x_143 = pad(constant_val = const_74, mode = x_143_mode_0, pad = x_143_pad_0, x = x_141)[name = tensor("x_143")]; tensor var_1334 = const()[name = tensor("op_1334"), val = tensor([1, 8, -1, 188])]; tensor x_145 = reshape(shape = var_1334, x = x_143)[name = tensor("x_145")]; tensor var_1338_begin_0 = const()[name = tensor("op_1338_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1338_end_0 = const()[name = tensor("op_1338_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1338_end_mask_0 = const()[name = tensor("op_1338_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1338 = slice_by_index(begin = var_1338_begin_0, end = var_1338_end_0, end_mask = var_1338_end_mask_0, x = x_145)[name = tensor("op_1338")]; tensor var_1339 = const()[name = tensor("op_1339"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_25 = reshape(shape = var_1339, x = var_1338)[name = tensor("matrix_bd_25")]; tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; tensor transpose_108_perm_0 = const()[name = tensor("transpose_108_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_109_perm_0 = const()[name = tensor("transpose_109_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_109 = transpose(perm = transpose_109_perm_0, x = k_25)[name = tensor("transpose_267")]; tensor transpose_108 = transpose(perm = transpose_108_perm_0, x = var_1322)[name = tensor("transpose_268")]; tensor matrix_ac_13 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_108, y = transpose_109)[name = tensor("matrix_ac_13")]; tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_27 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25)[name = tensor("matrix_bd_27")]; tensor var_1348 = add(x = matrix_ac_13, y = matrix_bd_27)[name = tensor("op_1348")]; tensor _inversed_scores_25_y_0 = const()[name = tensor("_inversed_scores_25_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_25 = mul(x = var_1348, y = _inversed_scores_25_y_0)[name = tensor("_inversed_scores_25")]; tensor scores_27 = select(a = var_14, b = _inversed_scores_25, cond = mask_3)[name = tensor("scores_27")]; tensor var_1354 = softmax(axis = var_32, x = scores_27)[name = tensor("op_1354")]; tensor input_345 = select(a = var_13, b = var_1354, cond = mask_3)[name = tensor("input_345")]; tensor x_147_transpose_x_0 = const()[name = tensor("x_147_transpose_x_0"), val = tensor(false)]; tensor x_147_transpose_y_0 = const()[name = tensor("x_147_transpose_y_0"), val = tensor(false)]; tensor value_15 = transpose(perm = value_15_perm_0, x = v_13)[name = tensor("transpose_266")]; tensor x_147 = matmul(transpose_x = x_147_transpose_x_0, transpose_y = x_147_transpose_y_0, x = input_345, y = value_15)[name = tensor("x_147")]; tensor var_1358_perm_0 = const()[name = tensor("op_1358_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1359 = const()[name = tensor("op_1359"), val = tensor([1, -1, 1024])]; tensor var_1358 = transpose(perm = var_1358_perm_0, x = x_147)[name = tensor("transpose_265")]; tensor input_347 = reshape(shape = var_1359, x = var_1358)[name = tensor("input_347")]; tensor input_349 = linear(bias = encoder_layers_6_self_attn_linear_out_bias, weight = encoder_layers_6_self_attn_linear_out_weight_quantized, x = input_347)[name = tensor("linear_61")]; tensor input_351 = add(x = input_343, y = input_349)[name = tensor("input_351")]; tensor x_151_axes_0 = const()[name = tensor("x_151_axes_0"), val = tensor([-1])]; tensor x_151 = layer_norm(axes = x_151_axes_0, beta = encoder_layers_6_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_6_norm_conv_weight, x = input_351)[name = tensor("x_151")]; tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; tensor input_353 = transpose(perm = input_353_perm_0, x = x_151)[name = tensor("transpose_264")]; tensor input_355 = conv(bias = encoder_layers_6_conv_pointwise_conv1_bias, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = encoder_layers_6_conv_pointwise_conv1_weight_quantized, x = input_353)[name = tensor("input_355")]; tensor x_153_split_num_splits_0 = const()[name = tensor("x_153_split_num_splits_0"), val = tensor(2)]; tensor x_153_split_axis_0 = const()[name = tensor("x_153_split_axis_0"), val = tensor(1)]; tensor x_153_split_0, tensor x_153_split_1 = split(axis = x_153_split_axis_0, num_splits = x_153_split_num_splits_0, x = input_355)[name = tensor("x_153_split")]; tensor x_153_split_1_sigmoid = sigmoid(x = x_153_split_1)[name = tensor("x_153_split_1_sigmoid")]; tensor x_153 = mul(x = x_153_split_0, y = x_153_split_1_sigmoid)[name = tensor("x_153")]; tensor input_357 = select(a = var_13, b = x_153, cond = var_339)[name = tensor("input_357")]; tensor const_77 = const()[name = tensor("const_77"), val = tensor(0x0p+0)]; tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; tensor input_359 = pad(constant_val = const_77, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357)[name = tensor("input_359")]; tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(1024)]; tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; tensor const_260_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_260_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590489536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590499904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590498816)))]; tensor const_261 = const()[name = tensor("const_261"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590504064)))]; tensor input_363 = conv(bias = const_261, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_260_quantized, x = input_359)[name = tensor("input_363")]; tensor input_365 = silu(x = input_363)[name = tensor("input_365")]; tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; tensor x_155 = conv(bias = encoder_layers_6_conv_pointwise_conv2_bias, dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = encoder_layers_6_conv_pointwise_conv2_weight_quantized, x = input_365)[name = tensor("x_155")]; tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; tensor input_367 = transpose(perm = input_367_perm_0, x = x_155)[name = tensor("transpose_263")]; tensor input_369 = add(x = input_351, y = input_367)[name = tensor("input_369")]; tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; tensor input_371 = layer_norm(axes = input_371_axes_0, beta = encoder_layers_6_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_6_norm_feed_forward2_weight, x = input_369)[name = tensor("input_371")]; tensor input_373 = linear(bias = encoder_layers_6_feed_forward2_linear1_bias, weight = encoder_layers_6_feed_forward2_linear1_weight_quantized, x = input_371)[name = tensor("linear_62")]; tensor input_375 = silu(x = input_373)[name = tensor("input_375")]; tensor input_379 = linear(bias = encoder_layers_6_feed_forward2_linear2_bias, weight = encoder_layers_6_feed_forward2_linear2_weight_quantized, x = input_375)[name = tensor("linear_63")]; tensor var_1425 = const()[name = tensor("op_1425"), val = tensor(0x1p-1)]; tensor var_1426 = mul(x = input_379, y = var_1425)[name = tensor("op_1426")]; tensor input_381 = add(x = input_369, y = var_1426)[name = tensor("input_381")]; tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; tensor input_383 = layer_norm(axes = input_383_axes_0, beta = encoder_layers_6_norm_out_bias, epsilon = var_11, gamma = encoder_layers_6_norm_out_weight, x = input_381)[name = tensor("input_383")]; tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; tensor input_385 = layer_norm(axes = input_385_axes_0, beta = encoder_layers_7_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_7_norm_feed_forward1_weight, x = input_383)[name = tensor("input_385")]; tensor input_387 = linear(bias = encoder_layers_7_feed_forward1_linear1_bias, weight = encoder_layers_7_feed_forward1_linear1_weight_quantized, x = input_385)[name = tensor("linear_64")]; tensor input_389 = silu(x = input_387)[name = tensor("input_389")]; tensor input_393 = linear(bias = encoder_layers_7_feed_forward1_linear2_bias, weight = encoder_layers_7_feed_forward1_linear2_weight_quantized, x = input_389)[name = tensor("linear_65")]; tensor var_1456 = const()[name = tensor("op_1456"), val = tensor(0x1p-1)]; tensor var_1457 = mul(x = input_393, y = var_1456)[name = tensor("op_1457")]; tensor input_395 = add(x = input_383, y = var_1457)[name = tensor("input_395")]; tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; tensor query_15 = layer_norm(axes = query_15_axes_0, beta = encoder_layers_7_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_7_norm_self_att_weight, x = input_395)[name = tensor("query_15")]; tensor var_1473 = linear(bias = encoder_layers_7_self_attn_linear_q_bias, weight = encoder_layers_7_self_attn_linear_q_weight_quantized, x = query_15)[name = tensor("linear_66")]; tensor var_1474 = const()[name = tensor("op_1474"), val = tensor([1, -1, 8, 128])]; tensor q_43 = reshape(shape = var_1474, x = var_1473)[name = tensor("q_43")]; tensor var_1478 = linear(bias = encoder_layers_7_self_attn_linear_k_bias, weight = encoder_layers_7_self_attn_linear_k_weight_quantized, x = query_15)[name = tensor("linear_67")]; tensor var_1479 = const()[name = tensor("op_1479"), val = tensor([1, -1, 8, 128])]; tensor k_29 = reshape(shape = var_1479, x = var_1478)[name = tensor("k_29")]; tensor var_1483 = linear(bias = encoder_layers_7_self_attn_linear_v_bias, weight = encoder_layers_7_self_attn_linear_v_weight_quantized, x = query_15)[name = tensor("linear_68")]; tensor var_1484 = const()[name = tensor("op_1484"), val = tensor([1, -1, 8, 128])]; tensor v_15 = reshape(shape = var_1484, x = var_1483)[name = tensor("v_15")]; tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_1496 = add(x = q_43, y = encoder_layers_7_self_attn_pos_bias_u)[name = tensor("op_1496")]; tensor var_1498 = add(x = q_43, y = encoder_layers_7_self_attn_pos_bias_v)[name = tensor("op_1498")]; tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_1500_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1500_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590508224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590892736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590892288)))]; tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_15 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1498)[name = tensor("transpose_262")]; tensor x_163 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_15, y = op_1500_quantized)[name = tensor("x_163")]; tensor const_84 = const()[name = tensor("const_84"), val = tensor(0x0p+0)]; tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; tensor x_165 = pad(constant_val = const_84, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163)[name = tensor("x_165")]; tensor var_1508 = const()[name = tensor("op_1508"), val = tensor([1, 8, -1, 188])]; tensor x_167 = reshape(shape = var_1508, x = x_165)[name = tensor("x_167")]; tensor var_1512_begin_0 = const()[name = tensor("op_1512_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1512_end_0 = const()[name = tensor("op_1512_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1512_end_mask_0 = const()[name = tensor("op_1512_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1512 = slice_by_index(begin = var_1512_begin_0, end = var_1512_end_0, end_mask = var_1512_end_mask_0, x = x_167)[name = tensor("op_1512")]; tensor var_1513 = const()[name = tensor("op_1513"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_29 = reshape(shape = var_1513, x = var_1512)[name = tensor("matrix_bd_29")]; tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; tensor transpose_110_perm_0 = const()[name = tensor("transpose_110_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_111_perm_0 = const()[name = tensor("transpose_111_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_111 = transpose(perm = transpose_111_perm_0, x = k_29)[name = tensor("transpose_260")]; tensor transpose_110 = transpose(perm = transpose_110_perm_0, x = var_1496)[name = tensor("transpose_261")]; tensor matrix_ac_15 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_110, y = transpose_111)[name = tensor("matrix_ac_15")]; tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_31 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29)[name = tensor("matrix_bd_31")]; tensor var_1522 = add(x = matrix_ac_15, y = matrix_bd_31)[name = tensor("op_1522")]; tensor _inversed_scores_29_y_0 = const()[name = tensor("_inversed_scores_29_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_29 = mul(x = var_1522, y = _inversed_scores_29_y_0)[name = tensor("_inversed_scores_29")]; tensor scores_31 = select(a = var_14, b = _inversed_scores_29, cond = mask_3)[name = tensor("scores_31")]; tensor var_1528 = softmax(axis = var_32, x = scores_31)[name = tensor("op_1528")]; tensor input_397 = select(a = var_13, b = var_1528, cond = mask_3)[name = tensor("input_397")]; tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; tensor value_17 = transpose(perm = value_17_perm_0, x = v_15)[name = tensor("transpose_259")]; tensor x_169 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_397, y = value_17)[name = tensor("x_169")]; tensor var_1532_perm_0 = const()[name = tensor("op_1532_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1533 = const()[name = tensor("op_1533"), val = tensor([1, -1, 1024])]; tensor var_1532 = transpose(perm = var_1532_perm_0, x = x_169)[name = tensor("transpose_258")]; tensor input_399 = reshape(shape = var_1533, x = var_1532)[name = tensor("input_399")]; tensor input_401 = linear(bias = encoder_layers_7_self_attn_linear_out_bias, weight = encoder_layers_7_self_attn_linear_out_weight_quantized, x = input_399)[name = tensor("linear_70")]; tensor input_403 = add(x = input_395, y = input_401)[name = tensor("input_403")]; tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; tensor x_173 = layer_norm(axes = x_173_axes_0, beta = encoder_layers_7_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_7_norm_conv_weight, x = input_403)[name = tensor("x_173")]; tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; tensor input_405 = transpose(perm = input_405_perm_0, x = x_173)[name = tensor("transpose_257")]; tensor input_407 = conv(bias = encoder_layers_7_conv_pointwise_conv1_bias, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = encoder_layers_7_conv_pointwise_conv1_weight_quantized, x = input_405)[name = tensor("input_407")]; tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; tensor x_175_split_0, tensor x_175_split_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_407)[name = tensor("x_175_split")]; tensor x_175_split_1_sigmoid = sigmoid(x = x_175_split_1)[name = tensor("x_175_split_1_sigmoid")]; tensor x_175 = mul(x = x_175_split_0, y = x_175_split_1_sigmoid)[name = tensor("x_175")]; tensor input_409 = select(a = var_13, b = x_175, cond = var_339)[name = tensor("input_409")]; tensor const_87 = const()[name = tensor("const_87"), val = tensor(0x0p+0)]; tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; tensor input_411 = pad(constant_val = const_87, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409)[name = tensor("input_411")]; tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(1024)]; tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; tensor const_262_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_262_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590894336))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590904704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590903616)))]; tensor const_263 = const()[name = tensor("const_263"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590908864)))]; tensor input_415 = conv(bias = const_263, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_262_quantized, x = input_411)[name = tensor("input_415")]; tensor input_417 = silu(x = input_415)[name = tensor("input_417")]; tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(1)]; tensor x_177 = conv(bias = encoder_layers_7_conv_pointwise_conv2_bias, dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = encoder_layers_7_conv_pointwise_conv2_weight_quantized, x = input_417)[name = tensor("x_177")]; tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; tensor input_419 = transpose(perm = input_419_perm_0, x = x_177)[name = tensor("transpose_256")]; tensor input_421 = add(x = input_403, y = input_419)[name = tensor("input_421")]; tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; tensor input_423 = layer_norm(axes = input_423_axes_0, beta = encoder_layers_7_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_7_norm_feed_forward2_weight, x = input_421)[name = tensor("input_423")]; tensor input_425 = linear(bias = encoder_layers_7_feed_forward2_linear1_bias, weight = encoder_layers_7_feed_forward2_linear1_weight_quantized, x = input_423)[name = tensor("linear_71")]; tensor input_427 = silu(x = input_425)[name = tensor("input_427")]; tensor input_431 = linear(bias = encoder_layers_7_feed_forward2_linear2_bias, weight = encoder_layers_7_feed_forward2_linear2_weight_quantized, x = input_427)[name = tensor("linear_72")]; tensor var_1599 = const()[name = tensor("op_1599"), val = tensor(0x1p-1)]; tensor var_1600 = mul(x = input_431, y = var_1599)[name = tensor("op_1600")]; tensor input_433 = add(x = input_421, y = var_1600)[name = tensor("input_433")]; tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; tensor input_435 = layer_norm(axes = input_435_axes_0, beta = encoder_layers_7_norm_out_bias, epsilon = var_11, gamma = encoder_layers_7_norm_out_weight, x = input_433)[name = tensor("input_435")]; tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; tensor input_437 = layer_norm(axes = input_437_axes_0, beta = encoder_layers_8_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_8_norm_feed_forward1_weight, x = input_435)[name = tensor("input_437")]; tensor input_439 = linear(bias = encoder_layers_8_feed_forward1_linear1_bias, weight = encoder_layers_8_feed_forward1_linear1_weight_quantized, x = input_437)[name = tensor("linear_73")]; tensor input_441 = silu(x = input_439)[name = tensor("input_441")]; tensor input_445 = linear(bias = encoder_layers_8_feed_forward1_linear2_bias, weight = encoder_layers_8_feed_forward1_linear2_weight_quantized, x = input_441)[name = tensor("linear_74")]; tensor var_1630 = const()[name = tensor("op_1630"), val = tensor(0x1p-1)]; tensor var_1631 = mul(x = input_445, y = var_1630)[name = tensor("op_1631")]; tensor input_447 = add(x = input_435, y = var_1631)[name = tensor("input_447")]; tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; tensor query_17 = layer_norm(axes = query_17_axes_0, beta = encoder_layers_8_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_8_norm_self_att_weight, x = input_447)[name = tensor("query_17")]; tensor var_1647 = linear(bias = encoder_layers_8_self_attn_linear_q_bias, weight = encoder_layers_8_self_attn_linear_q_weight_quantized, x = query_17)[name = tensor("linear_75")]; tensor var_1648 = const()[name = tensor("op_1648"), val = tensor([1, -1, 8, 128])]; tensor q_49 = reshape(shape = var_1648, x = var_1647)[name = tensor("q_49")]; tensor var_1652 = linear(bias = encoder_layers_8_self_attn_linear_k_bias, weight = encoder_layers_8_self_attn_linear_k_weight_quantized, x = query_17)[name = tensor("linear_76")]; tensor var_1653 = const()[name = tensor("op_1653"), val = tensor([1, -1, 8, 128])]; tensor k_33 = reshape(shape = var_1653, x = var_1652)[name = tensor("k_33")]; tensor var_1657 = linear(bias = encoder_layers_8_self_attn_linear_v_bias, weight = encoder_layers_8_self_attn_linear_v_weight_quantized, x = query_17)[name = tensor("linear_77")]; tensor var_1658 = const()[name = tensor("op_1658"), val = tensor([1, -1, 8, 128])]; tensor v_17 = reshape(shape = var_1658, x = var_1657)[name = tensor("v_17")]; tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_1670 = add(x = q_49, y = encoder_layers_8_self_attn_pos_bias_u)[name = tensor("op_1670")]; tensor var_1672 = add(x = q_49, y = encoder_layers_8_self_attn_pos_bias_v)[name = tensor("op_1672")]; tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_1674_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1674_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590913024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591297536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591297088)))]; tensor x_185_transpose_x_0 = const()[name = tensor("x_185_transpose_x_0"), val = tensor(false)]; tensor x_185_transpose_y_0 = const()[name = tensor("x_185_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_17 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1672)[name = tensor("transpose_255")]; tensor x_185 = matmul(transpose_x = x_185_transpose_x_0, transpose_y = x_185_transpose_y_0, x = q_with_bias_v_17, y = op_1674_quantized)[name = tensor("x_185")]; tensor const_94 = const()[name = tensor("const_94"), val = tensor(0x0p+0)]; tensor x_187_pad_0 = const()[name = tensor("x_187_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_187_mode_0 = const()[name = tensor("x_187_mode_0"), val = tensor("constant")]; tensor x_187 = pad(constant_val = const_94, mode = x_187_mode_0, pad = x_187_pad_0, x = x_185)[name = tensor("x_187")]; tensor var_1682 = const()[name = tensor("op_1682"), val = tensor([1, 8, -1, 188])]; tensor x_189 = reshape(shape = var_1682, x = x_187)[name = tensor("x_189")]; tensor var_1686_begin_0 = const()[name = tensor("op_1686_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1686_end_0 = const()[name = tensor("op_1686_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1686_end_mask_0 = const()[name = tensor("op_1686_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1686 = slice_by_index(begin = var_1686_begin_0, end = var_1686_end_0, end_mask = var_1686_end_mask_0, x = x_189)[name = tensor("op_1686")]; tensor var_1687 = const()[name = tensor("op_1687"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_33 = reshape(shape = var_1687, x = var_1686)[name = tensor("matrix_bd_33")]; tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; tensor transpose_112_perm_0 = const()[name = tensor("transpose_112_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_113_perm_0 = const()[name = tensor("transpose_113_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_113 = transpose(perm = transpose_113_perm_0, x = k_33)[name = tensor("transpose_253")]; tensor transpose_112 = transpose(perm = transpose_112_perm_0, x = var_1670)[name = tensor("transpose_254")]; tensor matrix_ac_17 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_112, y = transpose_113)[name = tensor("matrix_ac_17")]; tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_35 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33)[name = tensor("matrix_bd_35")]; tensor var_1696 = add(x = matrix_ac_17, y = matrix_bd_35)[name = tensor("op_1696")]; tensor _inversed_scores_33_y_0 = const()[name = tensor("_inversed_scores_33_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_33 = mul(x = var_1696, y = _inversed_scores_33_y_0)[name = tensor("_inversed_scores_33")]; tensor scores_35 = select(a = var_14, b = _inversed_scores_33, cond = mask_3)[name = tensor("scores_35")]; tensor var_1702 = softmax(axis = var_32, x = scores_35)[name = tensor("op_1702")]; tensor input_449 = select(a = var_13, b = var_1702, cond = mask_3)[name = tensor("input_449")]; tensor x_191_transpose_x_0 = const()[name = tensor("x_191_transpose_x_0"), val = tensor(false)]; tensor x_191_transpose_y_0 = const()[name = tensor("x_191_transpose_y_0"), val = tensor(false)]; tensor value_19 = transpose(perm = value_19_perm_0, x = v_17)[name = tensor("transpose_252")]; tensor x_191 = matmul(transpose_x = x_191_transpose_x_0, transpose_y = x_191_transpose_y_0, x = input_449, y = value_19)[name = tensor("x_191")]; tensor var_1706_perm_0 = const()[name = tensor("op_1706_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1707 = const()[name = tensor("op_1707"), val = tensor([1, -1, 1024])]; tensor var_1706 = transpose(perm = var_1706_perm_0, x = x_191)[name = tensor("transpose_251")]; tensor input_451 = reshape(shape = var_1707, x = var_1706)[name = tensor("input_451")]; tensor input_453 = linear(bias = encoder_layers_8_self_attn_linear_out_bias, weight = encoder_layers_8_self_attn_linear_out_weight_quantized, x = input_451)[name = tensor("linear_79")]; tensor input_455 = add(x = input_447, y = input_453)[name = tensor("input_455")]; tensor x_195_axes_0 = const()[name = tensor("x_195_axes_0"), val = tensor([-1])]; tensor x_195 = layer_norm(axes = x_195_axes_0, beta = encoder_layers_8_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_8_norm_conv_weight, x = input_455)[name = tensor("x_195")]; tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; tensor input_457 = transpose(perm = input_457_perm_0, x = x_195)[name = tensor("transpose_250")]; tensor input_459 = conv(bias = encoder_layers_8_conv_pointwise_conv1_bias, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = encoder_layers_8_conv_pointwise_conv1_weight_quantized, x = input_457)[name = tensor("input_459")]; tensor x_197_split_num_splits_0 = const()[name = tensor("x_197_split_num_splits_0"), val = tensor(2)]; tensor x_197_split_axis_0 = const()[name = tensor("x_197_split_axis_0"), val = tensor(1)]; tensor x_197_split_0, tensor x_197_split_1 = split(axis = x_197_split_axis_0, num_splits = x_197_split_num_splits_0, x = input_459)[name = tensor("x_197_split")]; tensor x_197_split_1_sigmoid = sigmoid(x = x_197_split_1)[name = tensor("x_197_split_1_sigmoid")]; tensor x_197 = mul(x = x_197_split_0, y = x_197_split_1_sigmoid)[name = tensor("x_197")]; tensor input_461 = select(a = var_13, b = x_197, cond = var_339)[name = tensor("input_461")]; tensor const_97 = const()[name = tensor("const_97"), val = tensor(0x0p+0)]; tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; tensor input_463 = pad(constant_val = const_97, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461)[name = tensor("input_463")]; tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(1024)]; tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; tensor const_264_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_264_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591299136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591309504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591308416)))]; tensor const_265 = const()[name = tensor("const_265"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591313664)))]; tensor input_467 = conv(bias = const_265, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_264_quantized, x = input_463)[name = tensor("input_467")]; tensor input_469 = silu(x = input_467)[name = tensor("input_469")]; tensor x_199_pad_type_0 = const()[name = tensor("x_199_pad_type_0"), val = tensor("valid")]; tensor x_199_strides_0 = const()[name = tensor("x_199_strides_0"), val = tensor([1])]; tensor x_199_pad_0 = const()[name = tensor("x_199_pad_0"), val = tensor([0, 0])]; tensor x_199_dilations_0 = const()[name = tensor("x_199_dilations_0"), val = tensor([1])]; tensor x_199_groups_0 = const()[name = tensor("x_199_groups_0"), val = tensor(1)]; tensor x_199 = conv(bias = encoder_layers_8_conv_pointwise_conv2_bias, dilations = x_199_dilations_0, groups = x_199_groups_0, pad = x_199_pad_0, pad_type = x_199_pad_type_0, strides = x_199_strides_0, weight = encoder_layers_8_conv_pointwise_conv2_weight_quantized, x = input_469)[name = tensor("x_199")]; tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; tensor input_471 = transpose(perm = input_471_perm_0, x = x_199)[name = tensor("transpose_249")]; tensor input_473 = add(x = input_455, y = input_471)[name = tensor("input_473")]; tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; tensor input_475 = layer_norm(axes = input_475_axes_0, beta = encoder_layers_8_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_8_norm_feed_forward2_weight, x = input_473)[name = tensor("input_475")]; tensor input_477 = linear(bias = encoder_layers_8_feed_forward2_linear1_bias, weight = encoder_layers_8_feed_forward2_linear1_weight_quantized, x = input_475)[name = tensor("linear_80")]; tensor input_479 = silu(x = input_477)[name = tensor("input_479")]; tensor input_483 = linear(bias = encoder_layers_8_feed_forward2_linear2_bias, weight = encoder_layers_8_feed_forward2_linear2_weight_quantized, x = input_479)[name = tensor("linear_81")]; tensor var_1773 = const()[name = tensor("op_1773"), val = tensor(0x1p-1)]; tensor var_1774 = mul(x = input_483, y = var_1773)[name = tensor("op_1774")]; tensor input_485 = add(x = input_473, y = var_1774)[name = tensor("input_485")]; tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; tensor input_487 = layer_norm(axes = input_487_axes_0, beta = encoder_layers_8_norm_out_bias, epsilon = var_11, gamma = encoder_layers_8_norm_out_weight, x = input_485)[name = tensor("input_487")]; tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; tensor input_489 = layer_norm(axes = input_489_axes_0, beta = encoder_layers_9_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_9_norm_feed_forward1_weight, x = input_487)[name = tensor("input_489")]; tensor input_491 = linear(bias = encoder_layers_9_feed_forward1_linear1_bias, weight = encoder_layers_9_feed_forward1_linear1_weight_quantized, x = input_489)[name = tensor("linear_82")]; tensor input_493 = silu(x = input_491)[name = tensor("input_493")]; tensor input_497 = linear(bias = encoder_layers_9_feed_forward1_linear2_bias, weight = encoder_layers_9_feed_forward1_linear2_weight_quantized, x = input_493)[name = tensor("linear_83")]; tensor var_1804 = const()[name = tensor("op_1804"), val = tensor(0x1p-1)]; tensor var_1805 = mul(x = input_497, y = var_1804)[name = tensor("op_1805")]; tensor input_499 = add(x = input_487, y = var_1805)[name = tensor("input_499")]; tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; tensor query_19 = layer_norm(axes = query_19_axes_0, beta = encoder_layers_9_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_9_norm_self_att_weight, x = input_499)[name = tensor("query_19")]; tensor var_1821 = linear(bias = encoder_layers_9_self_attn_linear_q_bias, weight = encoder_layers_9_self_attn_linear_q_weight_quantized, x = query_19)[name = tensor("linear_84")]; tensor var_1822 = const()[name = tensor("op_1822"), val = tensor([1, -1, 8, 128])]; tensor q_55 = reshape(shape = var_1822, x = var_1821)[name = tensor("q_55")]; tensor var_1826 = linear(bias = encoder_layers_9_self_attn_linear_k_bias, weight = encoder_layers_9_self_attn_linear_k_weight_quantized, x = query_19)[name = tensor("linear_85")]; tensor var_1827 = const()[name = tensor("op_1827"), val = tensor([1, -1, 8, 128])]; tensor k_37 = reshape(shape = var_1827, x = var_1826)[name = tensor("k_37")]; tensor var_1831 = linear(bias = encoder_layers_9_self_attn_linear_v_bias, weight = encoder_layers_9_self_attn_linear_v_weight_quantized, x = query_19)[name = tensor("linear_86")]; tensor var_1832 = const()[name = tensor("op_1832"), val = tensor([1, -1, 8, 128])]; tensor v_19 = reshape(shape = var_1832, x = var_1831)[name = tensor("v_19")]; tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_1844 = add(x = q_55, y = encoder_layers_9_self_attn_pos_bias_u)[name = tensor("op_1844")]; tensor var_1846 = add(x = q_55, y = encoder_layers_9_self_attn_pos_bias_v)[name = tensor("op_1846")]; tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_1848_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1848_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591317824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591702336))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591701888)))]; tensor x_207_transpose_x_0 = const()[name = tensor("x_207_transpose_x_0"), val = tensor(false)]; tensor x_207_transpose_y_0 = const()[name = tensor("x_207_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_19 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1846)[name = tensor("transpose_248")]; tensor x_207 = matmul(transpose_x = x_207_transpose_x_0, transpose_y = x_207_transpose_y_0, x = q_with_bias_v_19, y = op_1848_quantized)[name = tensor("x_207")]; tensor const_104 = const()[name = tensor("const_104"), val = tensor(0x0p+0)]; tensor x_209_pad_0 = const()[name = tensor("x_209_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_209_mode_0 = const()[name = tensor("x_209_mode_0"), val = tensor("constant")]; tensor x_209 = pad(constant_val = const_104, mode = x_209_mode_0, pad = x_209_pad_0, x = x_207)[name = tensor("x_209")]; tensor var_1856 = const()[name = tensor("op_1856"), val = tensor([1, 8, -1, 188])]; tensor x_211 = reshape(shape = var_1856, x = x_209)[name = tensor("x_211")]; tensor var_1860_begin_0 = const()[name = tensor("op_1860_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1860_end_0 = const()[name = tensor("op_1860_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1860_end_mask_0 = const()[name = tensor("op_1860_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1860 = slice_by_index(begin = var_1860_begin_0, end = var_1860_end_0, end_mask = var_1860_end_mask_0, x = x_211)[name = tensor("op_1860")]; tensor var_1861 = const()[name = tensor("op_1861"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_37 = reshape(shape = var_1861, x = var_1860)[name = tensor("matrix_bd_37")]; tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; tensor transpose_114_perm_0 = const()[name = tensor("transpose_114_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_115_perm_0 = const()[name = tensor("transpose_115_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_115 = transpose(perm = transpose_115_perm_0, x = k_37)[name = tensor("transpose_246")]; tensor transpose_114 = transpose(perm = transpose_114_perm_0, x = var_1844)[name = tensor("transpose_247")]; tensor matrix_ac_19 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_114, y = transpose_115)[name = tensor("matrix_ac_19")]; tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_39 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37)[name = tensor("matrix_bd_39")]; tensor var_1870 = add(x = matrix_ac_19, y = matrix_bd_39)[name = tensor("op_1870")]; tensor _inversed_scores_37_y_0 = const()[name = tensor("_inversed_scores_37_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_37 = mul(x = var_1870, y = _inversed_scores_37_y_0)[name = tensor("_inversed_scores_37")]; tensor scores_39 = select(a = var_14, b = _inversed_scores_37, cond = mask_3)[name = tensor("scores_39")]; tensor var_1876 = softmax(axis = var_32, x = scores_39)[name = tensor("op_1876")]; tensor input_501 = select(a = var_13, b = var_1876, cond = mask_3)[name = tensor("input_501")]; tensor x_213_transpose_x_0 = const()[name = tensor("x_213_transpose_x_0"), val = tensor(false)]; tensor x_213_transpose_y_0 = const()[name = tensor("x_213_transpose_y_0"), val = tensor(false)]; tensor value_21 = transpose(perm = value_21_perm_0, x = v_19)[name = tensor("transpose_245")]; tensor x_213 = matmul(transpose_x = x_213_transpose_x_0, transpose_y = x_213_transpose_y_0, x = input_501, y = value_21)[name = tensor("x_213")]; tensor var_1880_perm_0 = const()[name = tensor("op_1880_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1881 = const()[name = tensor("op_1881"), val = tensor([1, -1, 1024])]; tensor var_1880 = transpose(perm = var_1880_perm_0, x = x_213)[name = tensor("transpose_244")]; tensor input_503 = reshape(shape = var_1881, x = var_1880)[name = tensor("input_503")]; tensor input_505 = linear(bias = encoder_layers_9_self_attn_linear_out_bias, weight = encoder_layers_9_self_attn_linear_out_weight_quantized, x = input_503)[name = tensor("linear_88")]; tensor input_507 = add(x = input_499, y = input_505)[name = tensor("input_507")]; tensor x_217_axes_0 = const()[name = tensor("x_217_axes_0"), val = tensor([-1])]; tensor x_217 = layer_norm(axes = x_217_axes_0, beta = encoder_layers_9_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_9_norm_conv_weight, x = input_507)[name = tensor("x_217")]; tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; tensor input_509 = transpose(perm = input_509_perm_0, x = x_217)[name = tensor("transpose_243")]; tensor input_511 = conv(bias = encoder_layers_9_conv_pointwise_conv1_bias, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = encoder_layers_9_conv_pointwise_conv1_weight_quantized, x = input_509)[name = tensor("input_511")]; tensor x_219_split_num_splits_0 = const()[name = tensor("x_219_split_num_splits_0"), val = tensor(2)]; tensor x_219_split_axis_0 = const()[name = tensor("x_219_split_axis_0"), val = tensor(1)]; tensor x_219_split_0, tensor x_219_split_1 = split(axis = x_219_split_axis_0, num_splits = x_219_split_num_splits_0, x = input_511)[name = tensor("x_219_split")]; tensor x_219_split_1_sigmoid = sigmoid(x = x_219_split_1)[name = tensor("x_219_split_1_sigmoid")]; tensor x_219 = mul(x = x_219_split_0, y = x_219_split_1_sigmoid)[name = tensor("x_219")]; tensor input_513 = select(a = var_13, b = x_219, cond = var_339)[name = tensor("input_513")]; tensor const_107 = const()[name = tensor("const_107"), val = tensor(0x0p+0)]; tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; tensor input_515 = pad(constant_val = const_107, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513)[name = tensor("input_515")]; tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(1024)]; tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; tensor const_266_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_266_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591703936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591714304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591713216)))]; tensor const_267 = const()[name = tensor("const_267"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591718464)))]; tensor input_519 = conv(bias = const_267, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_266_quantized, x = input_515)[name = tensor("input_519")]; tensor input_521 = silu(x = input_519)[name = tensor("input_521")]; tensor x_221_pad_type_0 = const()[name = tensor("x_221_pad_type_0"), val = tensor("valid")]; tensor x_221_strides_0 = const()[name = tensor("x_221_strides_0"), val = tensor([1])]; tensor x_221_pad_0 = const()[name = tensor("x_221_pad_0"), val = tensor([0, 0])]; tensor x_221_dilations_0 = const()[name = tensor("x_221_dilations_0"), val = tensor([1])]; tensor x_221_groups_0 = const()[name = tensor("x_221_groups_0"), val = tensor(1)]; tensor x_221 = conv(bias = encoder_layers_9_conv_pointwise_conv2_bias, dilations = x_221_dilations_0, groups = x_221_groups_0, pad = x_221_pad_0, pad_type = x_221_pad_type_0, strides = x_221_strides_0, weight = encoder_layers_9_conv_pointwise_conv2_weight_quantized, x = input_521)[name = tensor("x_221")]; tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; tensor input_523 = transpose(perm = input_523_perm_0, x = x_221)[name = tensor("transpose_242")]; tensor input_525 = add(x = input_507, y = input_523)[name = tensor("input_525")]; tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; tensor input_527 = layer_norm(axes = input_527_axes_0, beta = encoder_layers_9_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_9_norm_feed_forward2_weight, x = input_525)[name = tensor("input_527")]; tensor input_529 = linear(bias = encoder_layers_9_feed_forward2_linear1_bias, weight = encoder_layers_9_feed_forward2_linear1_weight_quantized, x = input_527)[name = tensor("linear_89")]; tensor input_531 = silu(x = input_529)[name = tensor("input_531")]; tensor input_535 = linear(bias = encoder_layers_9_feed_forward2_linear2_bias, weight = encoder_layers_9_feed_forward2_linear2_weight_quantized, x = input_531)[name = tensor("linear_90")]; tensor var_1947 = const()[name = tensor("op_1947"), val = tensor(0x1p-1)]; tensor var_1948 = mul(x = input_535, y = var_1947)[name = tensor("op_1948")]; tensor input_537 = add(x = input_525, y = var_1948)[name = tensor("input_537")]; tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; tensor input_539 = layer_norm(axes = input_539_axes_0, beta = encoder_layers_9_norm_out_bias, epsilon = var_11, gamma = encoder_layers_9_norm_out_weight, x = input_537)[name = tensor("input_539")]; tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; tensor input_541 = layer_norm(axes = input_541_axes_0, beta = encoder_layers_10_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_10_norm_feed_forward1_weight, x = input_539)[name = tensor("input_541")]; tensor input_543 = linear(bias = encoder_layers_10_feed_forward1_linear1_bias, weight = encoder_layers_10_feed_forward1_linear1_weight_quantized, x = input_541)[name = tensor("linear_91")]; tensor input_545 = silu(x = input_543)[name = tensor("input_545")]; tensor input_549 = linear(bias = encoder_layers_10_feed_forward1_linear2_bias, weight = encoder_layers_10_feed_forward1_linear2_weight_quantized, x = input_545)[name = tensor("linear_92")]; tensor var_1978 = const()[name = tensor("op_1978"), val = tensor(0x1p-1)]; tensor var_1979 = mul(x = input_549, y = var_1978)[name = tensor("op_1979")]; tensor input_551 = add(x = input_539, y = var_1979)[name = tensor("input_551")]; tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; tensor query_21 = layer_norm(axes = query_21_axes_0, beta = encoder_layers_10_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_10_norm_self_att_weight, x = input_551)[name = tensor("query_21")]; tensor var_1995 = linear(bias = encoder_layers_10_self_attn_linear_q_bias, weight = encoder_layers_10_self_attn_linear_q_weight_quantized, x = query_21)[name = tensor("linear_93")]; tensor var_1996 = const()[name = tensor("op_1996"), val = tensor([1, -1, 8, 128])]; tensor q_61 = reshape(shape = var_1996, x = var_1995)[name = tensor("q_61")]; tensor var_2000 = linear(bias = encoder_layers_10_self_attn_linear_k_bias, weight = encoder_layers_10_self_attn_linear_k_weight_quantized, x = query_21)[name = tensor("linear_94")]; tensor var_2001 = const()[name = tensor("op_2001"), val = tensor([1, -1, 8, 128])]; tensor k_41 = reshape(shape = var_2001, x = var_2000)[name = tensor("k_41")]; tensor var_2005 = linear(bias = encoder_layers_10_self_attn_linear_v_bias, weight = encoder_layers_10_self_attn_linear_v_weight_quantized, x = query_21)[name = tensor("linear_95")]; tensor var_2006 = const()[name = tensor("op_2006"), val = tensor([1, -1, 8, 128])]; tensor v_21 = reshape(shape = var_2006, x = var_2005)[name = tensor("v_21")]; tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_2018 = add(x = q_61, y = encoder_layers_10_self_attn_pos_bias_u)[name = tensor("op_2018")]; tensor var_2020 = add(x = q_61, y = encoder_layers_10_self_attn_pos_bias_v)[name = tensor("op_2020")]; tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_2022_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2022_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(591722624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592107136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592106688)))]; tensor x_229_transpose_x_0 = const()[name = tensor("x_229_transpose_x_0"), val = tensor(false)]; tensor x_229_transpose_y_0 = const()[name = tensor("x_229_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_21 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2020)[name = tensor("transpose_241")]; tensor x_229 = matmul(transpose_x = x_229_transpose_x_0, transpose_y = x_229_transpose_y_0, x = q_with_bias_v_21, y = op_2022_quantized)[name = tensor("x_229")]; tensor const_114 = const()[name = tensor("const_114"), val = tensor(0x0p+0)]; tensor x_231_pad_0 = const()[name = tensor("x_231_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_231_mode_0 = const()[name = tensor("x_231_mode_0"), val = tensor("constant")]; tensor x_231 = pad(constant_val = const_114, mode = x_231_mode_0, pad = x_231_pad_0, x = x_229)[name = tensor("x_231")]; tensor var_2030 = const()[name = tensor("op_2030"), val = tensor([1, 8, -1, 188])]; tensor x_233 = reshape(shape = var_2030, x = x_231)[name = tensor("x_233")]; tensor var_2034_begin_0 = const()[name = tensor("op_2034_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2034_end_0 = const()[name = tensor("op_2034_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2034_end_mask_0 = const()[name = tensor("op_2034_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2034 = slice_by_index(begin = var_2034_begin_0, end = var_2034_end_0, end_mask = var_2034_end_mask_0, x = x_233)[name = tensor("op_2034")]; tensor var_2035 = const()[name = tensor("op_2035"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_41 = reshape(shape = var_2035, x = var_2034)[name = tensor("matrix_bd_41")]; tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; tensor transpose_116_perm_0 = const()[name = tensor("transpose_116_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_117_perm_0 = const()[name = tensor("transpose_117_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_117 = transpose(perm = transpose_117_perm_0, x = k_41)[name = tensor("transpose_239")]; tensor transpose_116 = transpose(perm = transpose_116_perm_0, x = var_2018)[name = tensor("transpose_240")]; tensor matrix_ac_21 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_116, y = transpose_117)[name = tensor("matrix_ac_21")]; tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_43 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41)[name = tensor("matrix_bd_43")]; tensor var_2044 = add(x = matrix_ac_21, y = matrix_bd_43)[name = tensor("op_2044")]; tensor _inversed_scores_41_y_0 = const()[name = tensor("_inversed_scores_41_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_41 = mul(x = var_2044, y = _inversed_scores_41_y_0)[name = tensor("_inversed_scores_41")]; tensor scores_43 = select(a = var_14, b = _inversed_scores_41, cond = mask_3)[name = tensor("scores_43")]; tensor var_2050 = softmax(axis = var_32, x = scores_43)[name = tensor("op_2050")]; tensor input_553 = select(a = var_13, b = var_2050, cond = mask_3)[name = tensor("input_553")]; tensor x_235_transpose_x_0 = const()[name = tensor("x_235_transpose_x_0"), val = tensor(false)]; tensor x_235_transpose_y_0 = const()[name = tensor("x_235_transpose_y_0"), val = tensor(false)]; tensor value_23 = transpose(perm = value_23_perm_0, x = v_21)[name = tensor("transpose_238")]; tensor x_235 = matmul(transpose_x = x_235_transpose_x_0, transpose_y = x_235_transpose_y_0, x = input_553, y = value_23)[name = tensor("x_235")]; tensor var_2054_perm_0 = const()[name = tensor("op_2054_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2055 = const()[name = tensor("op_2055"), val = tensor([1, -1, 1024])]; tensor var_2054 = transpose(perm = var_2054_perm_0, x = x_235)[name = tensor("transpose_237")]; tensor input_555 = reshape(shape = var_2055, x = var_2054)[name = tensor("input_555")]; tensor input_557 = linear(bias = encoder_layers_10_self_attn_linear_out_bias, weight = encoder_layers_10_self_attn_linear_out_weight_quantized, x = input_555)[name = tensor("linear_97")]; tensor input_559 = add(x = input_551, y = input_557)[name = tensor("input_559")]; tensor x_239_axes_0 = const()[name = tensor("x_239_axes_0"), val = tensor([-1])]; tensor x_239 = layer_norm(axes = x_239_axes_0, beta = encoder_layers_10_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_10_norm_conv_weight, x = input_559)[name = tensor("x_239")]; tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; tensor input_561 = transpose(perm = input_561_perm_0, x = x_239)[name = tensor("transpose_236")]; tensor input_563 = conv(bias = encoder_layers_10_conv_pointwise_conv1_bias, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = encoder_layers_10_conv_pointwise_conv1_weight_quantized, x = input_561)[name = tensor("input_563")]; tensor x_241_split_num_splits_0 = const()[name = tensor("x_241_split_num_splits_0"), val = tensor(2)]; tensor x_241_split_axis_0 = const()[name = tensor("x_241_split_axis_0"), val = tensor(1)]; tensor x_241_split_0, tensor x_241_split_1 = split(axis = x_241_split_axis_0, num_splits = x_241_split_num_splits_0, x = input_563)[name = tensor("x_241_split")]; tensor x_241_split_1_sigmoid = sigmoid(x = x_241_split_1)[name = tensor("x_241_split_1_sigmoid")]; tensor x_241 = mul(x = x_241_split_0, y = x_241_split_1_sigmoid)[name = tensor("x_241")]; tensor input_565 = select(a = var_13, b = x_241, cond = var_339)[name = tensor("input_565")]; tensor const_117 = const()[name = tensor("const_117"), val = tensor(0x0p+0)]; tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; tensor input_567 = pad(constant_val = const_117, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565)[name = tensor("input_567")]; tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(1024)]; tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; tensor const_268_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_268_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592108736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592119104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592118016)))]; tensor const_269 = const()[name = tensor("const_269"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592123264)))]; tensor input_571 = conv(bias = const_269, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_268_quantized, x = input_567)[name = tensor("input_571")]; tensor input_573 = silu(x = input_571)[name = tensor("input_573")]; tensor x_243_pad_type_0 = const()[name = tensor("x_243_pad_type_0"), val = tensor("valid")]; tensor x_243_strides_0 = const()[name = tensor("x_243_strides_0"), val = tensor([1])]; tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0])]; tensor x_243_dilations_0 = const()[name = tensor("x_243_dilations_0"), val = tensor([1])]; tensor x_243_groups_0 = const()[name = tensor("x_243_groups_0"), val = tensor(1)]; tensor x_243 = conv(bias = encoder_layers_10_conv_pointwise_conv2_bias, dilations = x_243_dilations_0, groups = x_243_groups_0, pad = x_243_pad_0, pad_type = x_243_pad_type_0, strides = x_243_strides_0, weight = encoder_layers_10_conv_pointwise_conv2_weight_quantized, x = input_573)[name = tensor("x_243")]; tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; tensor input_575 = transpose(perm = input_575_perm_0, x = x_243)[name = tensor("transpose_235")]; tensor input_577 = add(x = input_559, y = input_575)[name = tensor("input_577")]; tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; tensor input_579 = layer_norm(axes = input_579_axes_0, beta = encoder_layers_10_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_10_norm_feed_forward2_weight, x = input_577)[name = tensor("input_579")]; tensor input_581 = linear(bias = encoder_layers_10_feed_forward2_linear1_bias, weight = encoder_layers_10_feed_forward2_linear1_weight_quantized, x = input_579)[name = tensor("linear_98")]; tensor input_583 = silu(x = input_581)[name = tensor("input_583")]; tensor input_587 = linear(bias = encoder_layers_10_feed_forward2_linear2_bias, weight = encoder_layers_10_feed_forward2_linear2_weight_quantized, x = input_583)[name = tensor("linear_99")]; tensor var_2121 = const()[name = tensor("op_2121"), val = tensor(0x1p-1)]; tensor var_2122 = mul(x = input_587, y = var_2121)[name = tensor("op_2122")]; tensor input_589 = add(x = input_577, y = var_2122)[name = tensor("input_589")]; tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; tensor input_591 = layer_norm(axes = input_591_axes_0, beta = encoder_layers_10_norm_out_bias, epsilon = var_11, gamma = encoder_layers_10_norm_out_weight, x = input_589)[name = tensor("input_591")]; tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; tensor input_593 = layer_norm(axes = input_593_axes_0, beta = encoder_layers_11_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_11_norm_feed_forward1_weight, x = input_591)[name = tensor("input_593")]; tensor input_595 = linear(bias = encoder_layers_11_feed_forward1_linear1_bias, weight = encoder_layers_11_feed_forward1_linear1_weight_quantized, x = input_593)[name = tensor("linear_100")]; tensor input_597 = silu(x = input_595)[name = tensor("input_597")]; tensor input_601 = linear(bias = encoder_layers_11_feed_forward1_linear2_bias, weight = encoder_layers_11_feed_forward1_linear2_weight_quantized, x = input_597)[name = tensor("linear_101")]; tensor var_2152 = const()[name = tensor("op_2152"), val = tensor(0x1p-1)]; tensor var_2153 = mul(x = input_601, y = var_2152)[name = tensor("op_2153")]; tensor input_603 = add(x = input_591, y = var_2153)[name = tensor("input_603")]; tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; tensor query_23 = layer_norm(axes = query_23_axes_0, beta = encoder_layers_11_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_11_norm_self_att_weight, x = input_603)[name = tensor("query_23")]; tensor var_2169 = linear(bias = encoder_layers_11_self_attn_linear_q_bias, weight = encoder_layers_11_self_attn_linear_q_weight_quantized, x = query_23)[name = tensor("linear_102")]; tensor var_2170 = const()[name = tensor("op_2170"), val = tensor([1, -1, 8, 128])]; tensor q_67 = reshape(shape = var_2170, x = var_2169)[name = tensor("q_67")]; tensor var_2174 = linear(bias = encoder_layers_11_self_attn_linear_k_bias, weight = encoder_layers_11_self_attn_linear_k_weight_quantized, x = query_23)[name = tensor("linear_103")]; tensor var_2175 = const()[name = tensor("op_2175"), val = tensor([1, -1, 8, 128])]; tensor k_45 = reshape(shape = var_2175, x = var_2174)[name = tensor("k_45")]; tensor var_2179 = linear(bias = encoder_layers_11_self_attn_linear_v_bias, weight = encoder_layers_11_self_attn_linear_v_weight_quantized, x = query_23)[name = tensor("linear_104")]; tensor var_2180 = const()[name = tensor("op_2180"), val = tensor([1, -1, 8, 128])]; tensor v_23 = reshape(shape = var_2180, x = var_2179)[name = tensor("v_23")]; tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_2192 = add(x = q_67, y = encoder_layers_11_self_attn_pos_bias_u)[name = tensor("op_2192")]; tensor var_2194 = add(x = q_67, y = encoder_layers_11_self_attn_pos_bias_v)[name = tensor("op_2194")]; tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_2196_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2196_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592127424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592511936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592511488)))]; tensor x_251_transpose_x_0 = const()[name = tensor("x_251_transpose_x_0"), val = tensor(false)]; tensor x_251_transpose_y_0 = const()[name = tensor("x_251_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_23 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2194)[name = tensor("transpose_234")]; tensor x_251 = matmul(transpose_x = x_251_transpose_x_0, transpose_y = x_251_transpose_y_0, x = q_with_bias_v_23, y = op_2196_quantized)[name = tensor("x_251")]; tensor const_124 = const()[name = tensor("const_124"), val = tensor(0x0p+0)]; tensor x_253_pad_0 = const()[name = tensor("x_253_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_253_mode_0 = const()[name = tensor("x_253_mode_0"), val = tensor("constant")]; tensor x_253 = pad(constant_val = const_124, mode = x_253_mode_0, pad = x_253_pad_0, x = x_251)[name = tensor("x_253")]; tensor var_2204 = const()[name = tensor("op_2204"), val = tensor([1, 8, -1, 188])]; tensor x_255 = reshape(shape = var_2204, x = x_253)[name = tensor("x_255")]; tensor var_2208_begin_0 = const()[name = tensor("op_2208_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2208_end_0 = const()[name = tensor("op_2208_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2208_end_mask_0 = const()[name = tensor("op_2208_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2208 = slice_by_index(begin = var_2208_begin_0, end = var_2208_end_0, end_mask = var_2208_end_mask_0, x = x_255)[name = tensor("op_2208")]; tensor var_2209 = const()[name = tensor("op_2209"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_45 = reshape(shape = var_2209, x = var_2208)[name = tensor("matrix_bd_45")]; tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; tensor transpose_118_perm_0 = const()[name = tensor("transpose_118_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_119_perm_0 = const()[name = tensor("transpose_119_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_119 = transpose(perm = transpose_119_perm_0, x = k_45)[name = tensor("transpose_232")]; tensor transpose_118 = transpose(perm = transpose_118_perm_0, x = var_2192)[name = tensor("transpose_233")]; tensor matrix_ac_23 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_118, y = transpose_119)[name = tensor("matrix_ac_23")]; tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_47 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45)[name = tensor("matrix_bd_47")]; tensor var_2218 = add(x = matrix_ac_23, y = matrix_bd_47)[name = tensor("op_2218")]; tensor _inversed_scores_45_y_0 = const()[name = tensor("_inversed_scores_45_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_45 = mul(x = var_2218, y = _inversed_scores_45_y_0)[name = tensor("_inversed_scores_45")]; tensor scores_47 = select(a = var_14, b = _inversed_scores_45, cond = mask_3)[name = tensor("scores_47")]; tensor var_2224 = softmax(axis = var_32, x = scores_47)[name = tensor("op_2224")]; tensor input_605 = select(a = var_13, b = var_2224, cond = mask_3)[name = tensor("input_605")]; tensor x_257_transpose_x_0 = const()[name = tensor("x_257_transpose_x_0"), val = tensor(false)]; tensor x_257_transpose_y_0 = const()[name = tensor("x_257_transpose_y_0"), val = tensor(false)]; tensor value_25 = transpose(perm = value_25_perm_0, x = v_23)[name = tensor("transpose_231")]; tensor x_257 = matmul(transpose_x = x_257_transpose_x_0, transpose_y = x_257_transpose_y_0, x = input_605, y = value_25)[name = tensor("x_257")]; tensor var_2228_perm_0 = const()[name = tensor("op_2228_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2229 = const()[name = tensor("op_2229"), val = tensor([1, -1, 1024])]; tensor var_2228 = transpose(perm = var_2228_perm_0, x = x_257)[name = tensor("transpose_230")]; tensor input_607 = reshape(shape = var_2229, x = var_2228)[name = tensor("input_607")]; tensor input_609 = linear(bias = encoder_layers_11_self_attn_linear_out_bias, weight = encoder_layers_11_self_attn_linear_out_weight_quantized, x = input_607)[name = tensor("linear_106")]; tensor input_611 = add(x = input_603, y = input_609)[name = tensor("input_611")]; tensor x_261_axes_0 = const()[name = tensor("x_261_axes_0"), val = tensor([-1])]; tensor x_261 = layer_norm(axes = x_261_axes_0, beta = encoder_layers_11_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_11_norm_conv_weight, x = input_611)[name = tensor("x_261")]; tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; tensor input_613 = transpose(perm = input_613_perm_0, x = x_261)[name = tensor("transpose_229")]; tensor input_615 = conv(bias = encoder_layers_11_conv_pointwise_conv1_bias, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = encoder_layers_11_conv_pointwise_conv1_weight_quantized, x = input_613)[name = tensor("input_615")]; tensor x_263_split_num_splits_0 = const()[name = tensor("x_263_split_num_splits_0"), val = tensor(2)]; tensor x_263_split_axis_0 = const()[name = tensor("x_263_split_axis_0"), val = tensor(1)]; tensor x_263_split_0, tensor x_263_split_1 = split(axis = x_263_split_axis_0, num_splits = x_263_split_num_splits_0, x = input_615)[name = tensor("x_263_split")]; tensor x_263_split_1_sigmoid = sigmoid(x = x_263_split_1)[name = tensor("x_263_split_1_sigmoid")]; tensor x_263 = mul(x = x_263_split_0, y = x_263_split_1_sigmoid)[name = tensor("x_263")]; tensor input_617 = select(a = var_13, b = x_263, cond = var_339)[name = tensor("input_617")]; tensor const_127 = const()[name = tensor("const_127"), val = tensor(0x0p+0)]; tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; tensor input_619 = pad(constant_val = const_127, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617)[name = tensor("input_619")]; tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(1024)]; tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; tensor const_270_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_270_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592513536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592523904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592522816)))]; tensor const_271 = const()[name = tensor("const_271"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592528064)))]; tensor input_623 = conv(bias = const_271, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_270_quantized, x = input_619)[name = tensor("input_623")]; tensor input_625 = silu(x = input_623)[name = tensor("input_625")]; tensor x_265_pad_type_0 = const()[name = tensor("x_265_pad_type_0"), val = tensor("valid")]; tensor x_265_strides_0 = const()[name = tensor("x_265_strides_0"), val = tensor([1])]; tensor x_265_pad_0 = const()[name = tensor("x_265_pad_0"), val = tensor([0, 0])]; tensor x_265_dilations_0 = const()[name = tensor("x_265_dilations_0"), val = tensor([1])]; tensor x_265_groups_0 = const()[name = tensor("x_265_groups_0"), val = tensor(1)]; tensor x_265 = conv(bias = encoder_layers_11_conv_pointwise_conv2_bias, dilations = x_265_dilations_0, groups = x_265_groups_0, pad = x_265_pad_0, pad_type = x_265_pad_type_0, strides = x_265_strides_0, weight = encoder_layers_11_conv_pointwise_conv2_weight_quantized, x = input_625)[name = tensor("x_265")]; tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; tensor input_627 = transpose(perm = input_627_perm_0, x = x_265)[name = tensor("transpose_228")]; tensor input_629 = add(x = input_611, y = input_627)[name = tensor("input_629")]; tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; tensor input_631 = layer_norm(axes = input_631_axes_0, beta = encoder_layers_11_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_11_norm_feed_forward2_weight, x = input_629)[name = tensor("input_631")]; tensor input_633 = linear(bias = encoder_layers_11_feed_forward2_linear1_bias, weight = encoder_layers_11_feed_forward2_linear1_weight_quantized, x = input_631)[name = tensor("linear_107")]; tensor input_635 = silu(x = input_633)[name = tensor("input_635")]; tensor input_639 = linear(bias = encoder_layers_11_feed_forward2_linear2_bias, weight = encoder_layers_11_feed_forward2_linear2_weight_quantized, x = input_635)[name = tensor("linear_108")]; tensor var_2295 = const()[name = tensor("op_2295"), val = tensor(0x1p-1)]; tensor var_2296 = mul(x = input_639, y = var_2295)[name = tensor("op_2296")]; tensor input_641 = add(x = input_629, y = var_2296)[name = tensor("input_641")]; tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; tensor input_643 = layer_norm(axes = input_643_axes_0, beta = encoder_layers_11_norm_out_bias, epsilon = var_11, gamma = encoder_layers_11_norm_out_weight, x = input_641)[name = tensor("input_643")]; tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; tensor input_645 = layer_norm(axes = input_645_axes_0, beta = encoder_layers_12_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_12_norm_feed_forward1_weight, x = input_643)[name = tensor("input_645")]; tensor input_647 = linear(bias = encoder_layers_12_feed_forward1_linear1_bias, weight = encoder_layers_12_feed_forward1_linear1_weight_quantized, x = input_645)[name = tensor("linear_109")]; tensor input_649 = silu(x = input_647)[name = tensor("input_649")]; tensor input_653 = linear(bias = encoder_layers_12_feed_forward1_linear2_bias, weight = encoder_layers_12_feed_forward1_linear2_weight_quantized, x = input_649)[name = tensor("linear_110")]; tensor var_2326 = const()[name = tensor("op_2326"), val = tensor(0x1p-1)]; tensor var_2327 = mul(x = input_653, y = var_2326)[name = tensor("op_2327")]; tensor input_655 = add(x = input_643, y = var_2327)[name = tensor("input_655")]; tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; tensor query_25 = layer_norm(axes = query_25_axes_0, beta = encoder_layers_12_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_12_norm_self_att_weight, x = input_655)[name = tensor("query_25")]; tensor var_2343 = linear(bias = encoder_layers_12_self_attn_linear_q_bias, weight = encoder_layers_12_self_attn_linear_q_weight_quantized, x = query_25)[name = tensor("linear_111")]; tensor var_2344 = const()[name = tensor("op_2344"), val = tensor([1, -1, 8, 128])]; tensor q_73 = reshape(shape = var_2344, x = var_2343)[name = tensor("q_73")]; tensor var_2348 = linear(bias = encoder_layers_12_self_attn_linear_k_bias, weight = encoder_layers_12_self_attn_linear_k_weight_quantized, x = query_25)[name = tensor("linear_112")]; tensor var_2349 = const()[name = tensor("op_2349"), val = tensor([1, -1, 8, 128])]; tensor k_49 = reshape(shape = var_2349, x = var_2348)[name = tensor("k_49")]; tensor var_2353 = linear(bias = encoder_layers_12_self_attn_linear_v_bias, weight = encoder_layers_12_self_attn_linear_v_weight_quantized, x = query_25)[name = tensor("linear_113")]; tensor var_2354 = const()[name = tensor("op_2354"), val = tensor([1, -1, 8, 128])]; tensor v_25 = reshape(shape = var_2354, x = var_2353)[name = tensor("v_25")]; tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_2366 = add(x = q_73, y = encoder_layers_12_self_attn_pos_bias_u)[name = tensor("op_2366")]; tensor var_2368 = add(x = q_73, y = encoder_layers_12_self_attn_pos_bias_v)[name = tensor("op_2368")]; tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_2370_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2370_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592532224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592916736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592916288)))]; tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_25 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2368)[name = tensor("transpose_227")]; tensor x_273 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = q_with_bias_v_25, y = op_2370_quantized)[name = tensor("x_273")]; tensor const_134 = const()[name = tensor("const_134"), val = tensor(0x0p+0)]; tensor x_275_pad_0 = const()[name = tensor("x_275_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_275_mode_0 = const()[name = tensor("x_275_mode_0"), val = tensor("constant")]; tensor x_275 = pad(constant_val = const_134, mode = x_275_mode_0, pad = x_275_pad_0, x = x_273)[name = tensor("x_275")]; tensor var_2378 = const()[name = tensor("op_2378"), val = tensor([1, 8, -1, 188])]; tensor x_277 = reshape(shape = var_2378, x = x_275)[name = tensor("x_277")]; tensor var_2382_begin_0 = const()[name = tensor("op_2382_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2382_end_0 = const()[name = tensor("op_2382_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2382_end_mask_0 = const()[name = tensor("op_2382_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2382 = slice_by_index(begin = var_2382_begin_0, end = var_2382_end_0, end_mask = var_2382_end_mask_0, x = x_277)[name = tensor("op_2382")]; tensor var_2383 = const()[name = tensor("op_2383"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_49 = reshape(shape = var_2383, x = var_2382)[name = tensor("matrix_bd_49")]; tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; tensor transpose_120_perm_0 = const()[name = tensor("transpose_120_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_121_perm_0 = const()[name = tensor("transpose_121_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_121 = transpose(perm = transpose_121_perm_0, x = k_49)[name = tensor("transpose_225")]; tensor transpose_120 = transpose(perm = transpose_120_perm_0, x = var_2366)[name = tensor("transpose_226")]; tensor matrix_ac_25 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_120, y = transpose_121)[name = tensor("matrix_ac_25")]; tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_51 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49)[name = tensor("matrix_bd_51")]; tensor var_2392 = add(x = matrix_ac_25, y = matrix_bd_51)[name = tensor("op_2392")]; tensor _inversed_scores_49_y_0 = const()[name = tensor("_inversed_scores_49_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_49 = mul(x = var_2392, y = _inversed_scores_49_y_0)[name = tensor("_inversed_scores_49")]; tensor scores_51 = select(a = var_14, b = _inversed_scores_49, cond = mask_3)[name = tensor("scores_51")]; tensor var_2398 = softmax(axis = var_32, x = scores_51)[name = tensor("op_2398")]; tensor input_657 = select(a = var_13, b = var_2398, cond = mask_3)[name = tensor("input_657")]; tensor x_279_transpose_x_0 = const()[name = tensor("x_279_transpose_x_0"), val = tensor(false)]; tensor x_279_transpose_y_0 = const()[name = tensor("x_279_transpose_y_0"), val = tensor(false)]; tensor value_27 = transpose(perm = value_27_perm_0, x = v_25)[name = tensor("transpose_224")]; tensor x_279 = matmul(transpose_x = x_279_transpose_x_0, transpose_y = x_279_transpose_y_0, x = input_657, y = value_27)[name = tensor("x_279")]; tensor var_2402_perm_0 = const()[name = tensor("op_2402_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2403 = const()[name = tensor("op_2403"), val = tensor([1, -1, 1024])]; tensor var_2402 = transpose(perm = var_2402_perm_0, x = x_279)[name = tensor("transpose_223")]; tensor input_659 = reshape(shape = var_2403, x = var_2402)[name = tensor("input_659")]; tensor input_661 = linear(bias = encoder_layers_12_self_attn_linear_out_bias, weight = encoder_layers_12_self_attn_linear_out_weight_quantized, x = input_659)[name = tensor("linear_115")]; tensor input_663 = add(x = input_655, y = input_661)[name = tensor("input_663")]; tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; tensor x_283 = layer_norm(axes = x_283_axes_0, beta = encoder_layers_12_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_12_norm_conv_weight, x = input_663)[name = tensor("x_283")]; tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; tensor input_665 = transpose(perm = input_665_perm_0, x = x_283)[name = tensor("transpose_222")]; tensor input_667 = conv(bias = encoder_layers_12_conv_pointwise_conv1_bias, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = encoder_layers_12_conv_pointwise_conv1_weight_quantized, x = input_665)[name = tensor("input_667")]; tensor x_285_split_num_splits_0 = const()[name = tensor("x_285_split_num_splits_0"), val = tensor(2)]; tensor x_285_split_axis_0 = const()[name = tensor("x_285_split_axis_0"), val = tensor(1)]; tensor x_285_split_0, tensor x_285_split_1 = split(axis = x_285_split_axis_0, num_splits = x_285_split_num_splits_0, x = input_667)[name = tensor("x_285_split")]; tensor x_285_split_1_sigmoid = sigmoid(x = x_285_split_1)[name = tensor("x_285_split_1_sigmoid")]; tensor x_285 = mul(x = x_285_split_0, y = x_285_split_1_sigmoid)[name = tensor("x_285")]; tensor input_669 = select(a = var_13, b = x_285, cond = var_339)[name = tensor("input_669")]; tensor const_137 = const()[name = tensor("const_137"), val = tensor(0x0p+0)]; tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; tensor input_671 = pad(constant_val = const_137, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669)[name = tensor("input_671")]; tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(1024)]; tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; tensor const_272_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_272_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592918336))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592928704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592927616)))]; tensor const_273 = const()[name = tensor("const_273"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592932864)))]; tensor input_675 = conv(bias = const_273, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_272_quantized, x = input_671)[name = tensor("input_675")]; tensor input_677 = silu(x = input_675)[name = tensor("input_677")]; tensor x_287_pad_type_0 = const()[name = tensor("x_287_pad_type_0"), val = tensor("valid")]; tensor x_287_strides_0 = const()[name = tensor("x_287_strides_0"), val = tensor([1])]; tensor x_287_pad_0 = const()[name = tensor("x_287_pad_0"), val = tensor([0, 0])]; tensor x_287_dilations_0 = const()[name = tensor("x_287_dilations_0"), val = tensor([1])]; tensor x_287_groups_0 = const()[name = tensor("x_287_groups_0"), val = tensor(1)]; tensor x_287 = conv(bias = encoder_layers_12_conv_pointwise_conv2_bias, dilations = x_287_dilations_0, groups = x_287_groups_0, pad = x_287_pad_0, pad_type = x_287_pad_type_0, strides = x_287_strides_0, weight = encoder_layers_12_conv_pointwise_conv2_weight_quantized, x = input_677)[name = tensor("x_287")]; tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; tensor input_679 = transpose(perm = input_679_perm_0, x = x_287)[name = tensor("transpose_221")]; tensor input_681 = add(x = input_663, y = input_679)[name = tensor("input_681")]; tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; tensor input_683 = layer_norm(axes = input_683_axes_0, beta = encoder_layers_12_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_12_norm_feed_forward2_weight, x = input_681)[name = tensor("input_683")]; tensor input_685 = linear(bias = encoder_layers_12_feed_forward2_linear1_bias, weight = encoder_layers_12_feed_forward2_linear1_weight_quantized, x = input_683)[name = tensor("linear_116")]; tensor input_687 = silu(x = input_685)[name = tensor("input_687")]; tensor input_691 = linear(bias = encoder_layers_12_feed_forward2_linear2_bias, weight = encoder_layers_12_feed_forward2_linear2_weight_quantized, x = input_687)[name = tensor("linear_117")]; tensor var_2469 = const()[name = tensor("op_2469"), val = tensor(0x1p-1)]; tensor var_2470 = mul(x = input_691, y = var_2469)[name = tensor("op_2470")]; tensor input_693 = add(x = input_681, y = var_2470)[name = tensor("input_693")]; tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; tensor input_695 = layer_norm(axes = input_695_axes_0, beta = encoder_layers_12_norm_out_bias, epsilon = var_11, gamma = encoder_layers_12_norm_out_weight, x = input_693)[name = tensor("input_695")]; tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; tensor input_697 = layer_norm(axes = input_697_axes_0, beta = encoder_layers_13_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_13_norm_feed_forward1_weight, x = input_695)[name = tensor("input_697")]; tensor input_699 = linear(bias = encoder_layers_13_feed_forward1_linear1_bias, weight = encoder_layers_13_feed_forward1_linear1_weight_quantized, x = input_697)[name = tensor("linear_118")]; tensor input_701 = silu(x = input_699)[name = tensor("input_701")]; tensor input_705 = linear(bias = encoder_layers_13_feed_forward1_linear2_bias, weight = encoder_layers_13_feed_forward1_linear2_weight_quantized, x = input_701)[name = tensor("linear_119")]; tensor var_2500 = const()[name = tensor("op_2500"), val = tensor(0x1p-1)]; tensor var_2501 = mul(x = input_705, y = var_2500)[name = tensor("op_2501")]; tensor input_707 = add(x = input_695, y = var_2501)[name = tensor("input_707")]; tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; tensor query_27 = layer_norm(axes = query_27_axes_0, beta = encoder_layers_13_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_13_norm_self_att_weight, x = input_707)[name = tensor("query_27")]; tensor var_2517 = linear(bias = encoder_layers_13_self_attn_linear_q_bias, weight = encoder_layers_13_self_attn_linear_q_weight_quantized, x = query_27)[name = tensor("linear_120")]; tensor var_2518 = const()[name = tensor("op_2518"), val = tensor([1, -1, 8, 128])]; tensor q_79 = reshape(shape = var_2518, x = var_2517)[name = tensor("q_79")]; tensor var_2522 = linear(bias = encoder_layers_13_self_attn_linear_k_bias, weight = encoder_layers_13_self_attn_linear_k_weight_quantized, x = query_27)[name = tensor("linear_121")]; tensor var_2523 = const()[name = tensor("op_2523"), val = tensor([1, -1, 8, 128])]; tensor k_53 = reshape(shape = var_2523, x = var_2522)[name = tensor("k_53")]; tensor var_2527 = linear(bias = encoder_layers_13_self_attn_linear_v_bias, weight = encoder_layers_13_self_attn_linear_v_weight_quantized, x = query_27)[name = tensor("linear_122")]; tensor var_2528 = const()[name = tensor("op_2528"), val = tensor([1, -1, 8, 128])]; tensor v_27 = reshape(shape = var_2528, x = var_2527)[name = tensor("v_27")]; tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_2540 = add(x = q_79, y = encoder_layers_13_self_attn_pos_bias_u)[name = tensor("op_2540")]; tensor var_2542 = add(x = q_79, y = encoder_layers_13_self_attn_pos_bias_v)[name = tensor("op_2542")]; tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_2544_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2544_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592937024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593321536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593321088)))]; tensor x_295_transpose_x_0 = const()[name = tensor("x_295_transpose_x_0"), val = tensor(false)]; tensor x_295_transpose_y_0 = const()[name = tensor("x_295_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_27 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2542)[name = tensor("transpose_220")]; tensor x_295 = matmul(transpose_x = x_295_transpose_x_0, transpose_y = x_295_transpose_y_0, x = q_with_bias_v_27, y = op_2544_quantized)[name = tensor("x_295")]; tensor const_144 = const()[name = tensor("const_144"), val = tensor(0x0p+0)]; tensor x_297_pad_0 = const()[name = tensor("x_297_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_297_mode_0 = const()[name = tensor("x_297_mode_0"), val = tensor("constant")]; tensor x_297 = pad(constant_val = const_144, mode = x_297_mode_0, pad = x_297_pad_0, x = x_295)[name = tensor("x_297")]; tensor var_2552 = const()[name = tensor("op_2552"), val = tensor([1, 8, -1, 188])]; tensor x_299 = reshape(shape = var_2552, x = x_297)[name = tensor("x_299")]; tensor var_2556_begin_0 = const()[name = tensor("op_2556_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2556_end_0 = const()[name = tensor("op_2556_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2556_end_mask_0 = const()[name = tensor("op_2556_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2556 = slice_by_index(begin = var_2556_begin_0, end = var_2556_end_0, end_mask = var_2556_end_mask_0, x = x_299)[name = tensor("op_2556")]; tensor var_2557 = const()[name = tensor("op_2557"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_53 = reshape(shape = var_2557, x = var_2556)[name = tensor("matrix_bd_53")]; tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; tensor transpose_122_perm_0 = const()[name = tensor("transpose_122_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_123_perm_0 = const()[name = tensor("transpose_123_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_123 = transpose(perm = transpose_123_perm_0, x = k_53)[name = tensor("transpose_218")]; tensor transpose_122 = transpose(perm = transpose_122_perm_0, x = var_2540)[name = tensor("transpose_219")]; tensor matrix_ac_27 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_122, y = transpose_123)[name = tensor("matrix_ac_27")]; tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_55 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53)[name = tensor("matrix_bd_55")]; tensor var_2566 = add(x = matrix_ac_27, y = matrix_bd_55)[name = tensor("op_2566")]; tensor _inversed_scores_53_y_0 = const()[name = tensor("_inversed_scores_53_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_53 = mul(x = var_2566, y = _inversed_scores_53_y_0)[name = tensor("_inversed_scores_53")]; tensor scores_55 = select(a = var_14, b = _inversed_scores_53, cond = mask_3)[name = tensor("scores_55")]; tensor var_2572 = softmax(axis = var_32, x = scores_55)[name = tensor("op_2572")]; tensor input_709 = select(a = var_13, b = var_2572, cond = mask_3)[name = tensor("input_709")]; tensor x_301_transpose_x_0 = const()[name = tensor("x_301_transpose_x_0"), val = tensor(false)]; tensor x_301_transpose_y_0 = const()[name = tensor("x_301_transpose_y_0"), val = tensor(false)]; tensor value_29 = transpose(perm = value_29_perm_0, x = v_27)[name = tensor("transpose_217")]; tensor x_301 = matmul(transpose_x = x_301_transpose_x_0, transpose_y = x_301_transpose_y_0, x = input_709, y = value_29)[name = tensor("x_301")]; tensor var_2576_perm_0 = const()[name = tensor("op_2576_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2577 = const()[name = tensor("op_2577"), val = tensor([1, -1, 1024])]; tensor var_2576 = transpose(perm = var_2576_perm_0, x = x_301)[name = tensor("transpose_216")]; tensor input_711 = reshape(shape = var_2577, x = var_2576)[name = tensor("input_711")]; tensor input_713 = linear(bias = encoder_layers_13_self_attn_linear_out_bias, weight = encoder_layers_13_self_attn_linear_out_weight_quantized, x = input_711)[name = tensor("linear_124")]; tensor input_715 = add(x = input_707, y = input_713)[name = tensor("input_715")]; tensor x_305_axes_0 = const()[name = tensor("x_305_axes_0"), val = tensor([-1])]; tensor x_305 = layer_norm(axes = x_305_axes_0, beta = encoder_layers_13_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_13_norm_conv_weight, x = input_715)[name = tensor("x_305")]; tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; tensor input_717 = transpose(perm = input_717_perm_0, x = x_305)[name = tensor("transpose_215")]; tensor input_719 = conv(bias = encoder_layers_13_conv_pointwise_conv1_bias, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = encoder_layers_13_conv_pointwise_conv1_weight_quantized, x = input_717)[name = tensor("input_719")]; tensor x_307_split_num_splits_0 = const()[name = tensor("x_307_split_num_splits_0"), val = tensor(2)]; tensor x_307_split_axis_0 = const()[name = tensor("x_307_split_axis_0"), val = tensor(1)]; tensor x_307_split_0, tensor x_307_split_1 = split(axis = x_307_split_axis_0, num_splits = x_307_split_num_splits_0, x = input_719)[name = tensor("x_307_split")]; tensor x_307_split_1_sigmoid = sigmoid(x = x_307_split_1)[name = tensor("x_307_split_1_sigmoid")]; tensor x_307 = mul(x = x_307_split_0, y = x_307_split_1_sigmoid)[name = tensor("x_307")]; tensor input_721 = select(a = var_13, b = x_307, cond = var_339)[name = tensor("input_721")]; tensor const_147 = const()[name = tensor("const_147"), val = tensor(0x0p+0)]; tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; tensor input_723 = pad(constant_val = const_147, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721)[name = tensor("input_723")]; tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(1024)]; tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; tensor const_274_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_274_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593323136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593333504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593332416)))]; tensor const_275 = const()[name = tensor("const_275"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593337664)))]; tensor input_727 = conv(bias = const_275, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_274_quantized, x = input_723)[name = tensor("input_727")]; tensor input_729 = silu(x = input_727)[name = tensor("input_729")]; tensor x_309_pad_type_0 = const()[name = tensor("x_309_pad_type_0"), val = tensor("valid")]; tensor x_309_strides_0 = const()[name = tensor("x_309_strides_0"), val = tensor([1])]; tensor x_309_pad_0 = const()[name = tensor("x_309_pad_0"), val = tensor([0, 0])]; tensor x_309_dilations_0 = const()[name = tensor("x_309_dilations_0"), val = tensor([1])]; tensor x_309_groups_0 = const()[name = tensor("x_309_groups_0"), val = tensor(1)]; tensor x_309 = conv(bias = encoder_layers_13_conv_pointwise_conv2_bias, dilations = x_309_dilations_0, groups = x_309_groups_0, pad = x_309_pad_0, pad_type = x_309_pad_type_0, strides = x_309_strides_0, weight = encoder_layers_13_conv_pointwise_conv2_weight_quantized, x = input_729)[name = tensor("x_309")]; tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; tensor input_731 = transpose(perm = input_731_perm_0, x = x_309)[name = tensor("transpose_214")]; tensor input_733 = add(x = input_715, y = input_731)[name = tensor("input_733")]; tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; tensor input_735 = layer_norm(axes = input_735_axes_0, beta = encoder_layers_13_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_13_norm_feed_forward2_weight, x = input_733)[name = tensor("input_735")]; tensor input_737 = linear(bias = encoder_layers_13_feed_forward2_linear1_bias, weight = encoder_layers_13_feed_forward2_linear1_weight_quantized, x = input_735)[name = tensor("linear_125")]; tensor input_739 = silu(x = input_737)[name = tensor("input_739")]; tensor input_743 = linear(bias = encoder_layers_13_feed_forward2_linear2_bias, weight = encoder_layers_13_feed_forward2_linear2_weight_quantized, x = input_739)[name = tensor("linear_126")]; tensor var_2643 = const()[name = tensor("op_2643"), val = tensor(0x1p-1)]; tensor var_2644 = mul(x = input_743, y = var_2643)[name = tensor("op_2644")]; tensor input_745 = add(x = input_733, y = var_2644)[name = tensor("input_745")]; tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; tensor input_747 = layer_norm(axes = input_747_axes_0, beta = encoder_layers_13_norm_out_bias, epsilon = var_11, gamma = encoder_layers_13_norm_out_weight, x = input_745)[name = tensor("input_747")]; tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; tensor input_749 = layer_norm(axes = input_749_axes_0, beta = encoder_layers_14_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_14_norm_feed_forward1_weight, x = input_747)[name = tensor("input_749")]; tensor input_751 = linear(bias = encoder_layers_14_feed_forward1_linear1_bias, weight = encoder_layers_14_feed_forward1_linear1_weight_quantized, x = input_749)[name = tensor("linear_127")]; tensor input_753 = silu(x = input_751)[name = tensor("input_753")]; tensor input_757 = linear(bias = encoder_layers_14_feed_forward1_linear2_bias, weight = encoder_layers_14_feed_forward1_linear2_weight_quantized, x = input_753)[name = tensor("linear_128")]; tensor var_2674 = const()[name = tensor("op_2674"), val = tensor(0x1p-1)]; tensor var_2675 = mul(x = input_757, y = var_2674)[name = tensor("op_2675")]; tensor input_759 = add(x = input_747, y = var_2675)[name = tensor("input_759")]; tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; tensor query_29 = layer_norm(axes = query_29_axes_0, beta = encoder_layers_14_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_14_norm_self_att_weight, x = input_759)[name = tensor("query_29")]; tensor var_2691 = linear(bias = encoder_layers_14_self_attn_linear_q_bias, weight = encoder_layers_14_self_attn_linear_q_weight_quantized, x = query_29)[name = tensor("linear_129")]; tensor var_2692 = const()[name = tensor("op_2692"), val = tensor([1, -1, 8, 128])]; tensor q_85 = reshape(shape = var_2692, x = var_2691)[name = tensor("q_85")]; tensor var_2696 = linear(bias = encoder_layers_14_self_attn_linear_k_bias, weight = encoder_layers_14_self_attn_linear_k_weight_quantized, x = query_29)[name = tensor("linear_130")]; tensor var_2697 = const()[name = tensor("op_2697"), val = tensor([1, -1, 8, 128])]; tensor k_57 = reshape(shape = var_2697, x = var_2696)[name = tensor("k_57")]; tensor var_2701 = linear(bias = encoder_layers_14_self_attn_linear_v_bias, weight = encoder_layers_14_self_attn_linear_v_weight_quantized, x = query_29)[name = tensor("linear_131")]; tensor var_2702 = const()[name = tensor("op_2702"), val = tensor([1, -1, 8, 128])]; tensor v_29 = reshape(shape = var_2702, x = var_2701)[name = tensor("v_29")]; tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_2714 = add(x = q_85, y = encoder_layers_14_self_attn_pos_bias_u)[name = tensor("op_2714")]; tensor var_2716 = add(x = q_85, y = encoder_layers_14_self_attn_pos_bias_v)[name = tensor("op_2716")]; tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_2718_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2718_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593341824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593726336))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593725888)))]; tensor x_317_transpose_x_0 = const()[name = tensor("x_317_transpose_x_0"), val = tensor(false)]; tensor x_317_transpose_y_0 = const()[name = tensor("x_317_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_29 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2716)[name = tensor("transpose_213")]; tensor x_317 = matmul(transpose_x = x_317_transpose_x_0, transpose_y = x_317_transpose_y_0, x = q_with_bias_v_29, y = op_2718_quantized)[name = tensor("x_317")]; tensor const_154 = const()[name = tensor("const_154"), val = tensor(0x0p+0)]; tensor x_319_pad_0 = const()[name = tensor("x_319_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_319_mode_0 = const()[name = tensor("x_319_mode_0"), val = tensor("constant")]; tensor x_319 = pad(constant_val = const_154, mode = x_319_mode_0, pad = x_319_pad_0, x = x_317)[name = tensor("x_319")]; tensor var_2726 = const()[name = tensor("op_2726"), val = tensor([1, 8, -1, 188])]; tensor x_321 = reshape(shape = var_2726, x = x_319)[name = tensor("x_321")]; tensor var_2730_begin_0 = const()[name = tensor("op_2730_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2730_end_0 = const()[name = tensor("op_2730_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2730_end_mask_0 = const()[name = tensor("op_2730_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2730 = slice_by_index(begin = var_2730_begin_0, end = var_2730_end_0, end_mask = var_2730_end_mask_0, x = x_321)[name = tensor("op_2730")]; tensor var_2731 = const()[name = tensor("op_2731"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_57 = reshape(shape = var_2731, x = var_2730)[name = tensor("matrix_bd_57")]; tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; tensor transpose_124_perm_0 = const()[name = tensor("transpose_124_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_125_perm_0 = const()[name = tensor("transpose_125_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_125 = transpose(perm = transpose_125_perm_0, x = k_57)[name = tensor("transpose_211")]; tensor transpose_124 = transpose(perm = transpose_124_perm_0, x = var_2714)[name = tensor("transpose_212")]; tensor matrix_ac_29 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_124, y = transpose_125)[name = tensor("matrix_ac_29")]; tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_59 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57)[name = tensor("matrix_bd_59")]; tensor var_2740 = add(x = matrix_ac_29, y = matrix_bd_59)[name = tensor("op_2740")]; tensor _inversed_scores_57_y_0 = const()[name = tensor("_inversed_scores_57_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_57 = mul(x = var_2740, y = _inversed_scores_57_y_0)[name = tensor("_inversed_scores_57")]; tensor scores_59 = select(a = var_14, b = _inversed_scores_57, cond = mask_3)[name = tensor("scores_59")]; tensor var_2746 = softmax(axis = var_32, x = scores_59)[name = tensor("op_2746")]; tensor input_761 = select(a = var_13, b = var_2746, cond = mask_3)[name = tensor("input_761")]; tensor x_323_transpose_x_0 = const()[name = tensor("x_323_transpose_x_0"), val = tensor(false)]; tensor x_323_transpose_y_0 = const()[name = tensor("x_323_transpose_y_0"), val = tensor(false)]; tensor value_31 = transpose(perm = value_31_perm_0, x = v_29)[name = tensor("transpose_210")]; tensor x_323 = matmul(transpose_x = x_323_transpose_x_0, transpose_y = x_323_transpose_y_0, x = input_761, y = value_31)[name = tensor("x_323")]; tensor var_2750_perm_0 = const()[name = tensor("op_2750_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2751 = const()[name = tensor("op_2751"), val = tensor([1, -1, 1024])]; tensor var_2750 = transpose(perm = var_2750_perm_0, x = x_323)[name = tensor("transpose_209")]; tensor input_763 = reshape(shape = var_2751, x = var_2750)[name = tensor("input_763")]; tensor input_765 = linear(bias = encoder_layers_14_self_attn_linear_out_bias, weight = encoder_layers_14_self_attn_linear_out_weight_quantized, x = input_763)[name = tensor("linear_133")]; tensor input_767 = add(x = input_759, y = input_765)[name = tensor("input_767")]; tensor x_327_axes_0 = const()[name = tensor("x_327_axes_0"), val = tensor([-1])]; tensor x_327 = layer_norm(axes = x_327_axes_0, beta = encoder_layers_14_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_14_norm_conv_weight, x = input_767)[name = tensor("x_327")]; tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; tensor input_769 = transpose(perm = input_769_perm_0, x = x_327)[name = tensor("transpose_208")]; tensor input_771 = conv(bias = encoder_layers_14_conv_pointwise_conv1_bias, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = encoder_layers_14_conv_pointwise_conv1_weight_quantized, x = input_769)[name = tensor("input_771")]; tensor x_329_split_num_splits_0 = const()[name = tensor("x_329_split_num_splits_0"), val = tensor(2)]; tensor x_329_split_axis_0 = const()[name = tensor("x_329_split_axis_0"), val = tensor(1)]; tensor x_329_split_0, tensor x_329_split_1 = split(axis = x_329_split_axis_0, num_splits = x_329_split_num_splits_0, x = input_771)[name = tensor("x_329_split")]; tensor x_329_split_1_sigmoid = sigmoid(x = x_329_split_1)[name = tensor("x_329_split_1_sigmoid")]; tensor x_329 = mul(x = x_329_split_0, y = x_329_split_1_sigmoid)[name = tensor("x_329")]; tensor input_773 = select(a = var_13, b = x_329, cond = var_339)[name = tensor("input_773")]; tensor const_157 = const()[name = tensor("const_157"), val = tensor(0x0p+0)]; tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; tensor input_775 = pad(constant_val = const_157, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773)[name = tensor("input_775")]; tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(1024)]; tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; tensor const_276_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_276_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593727936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593738304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593737216)))]; tensor const_277 = const()[name = tensor("const_277"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593742464)))]; tensor input_779 = conv(bias = const_277, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_276_quantized, x = input_775)[name = tensor("input_779")]; tensor input_781 = silu(x = input_779)[name = tensor("input_781")]; tensor x_331_pad_type_0 = const()[name = tensor("x_331_pad_type_0"), val = tensor("valid")]; tensor x_331_strides_0 = const()[name = tensor("x_331_strides_0"), val = tensor([1])]; tensor x_331_pad_0 = const()[name = tensor("x_331_pad_0"), val = tensor([0, 0])]; tensor x_331_dilations_0 = const()[name = tensor("x_331_dilations_0"), val = tensor([1])]; tensor x_331_groups_0 = const()[name = tensor("x_331_groups_0"), val = tensor(1)]; tensor x_331 = conv(bias = encoder_layers_14_conv_pointwise_conv2_bias, dilations = x_331_dilations_0, groups = x_331_groups_0, pad = x_331_pad_0, pad_type = x_331_pad_type_0, strides = x_331_strides_0, weight = encoder_layers_14_conv_pointwise_conv2_weight_quantized, x = input_781)[name = tensor("x_331")]; tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; tensor input_783 = transpose(perm = input_783_perm_0, x = x_331)[name = tensor("transpose_207")]; tensor input_785 = add(x = input_767, y = input_783)[name = tensor("input_785")]; tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; tensor input_787 = layer_norm(axes = input_787_axes_0, beta = encoder_layers_14_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_14_norm_feed_forward2_weight, x = input_785)[name = tensor("input_787")]; tensor input_789 = linear(bias = encoder_layers_14_feed_forward2_linear1_bias, weight = encoder_layers_14_feed_forward2_linear1_weight_quantized, x = input_787)[name = tensor("linear_134")]; tensor input_791 = silu(x = input_789)[name = tensor("input_791")]; tensor input_795 = linear(bias = encoder_layers_14_feed_forward2_linear2_bias, weight = encoder_layers_14_feed_forward2_linear2_weight_quantized, x = input_791)[name = tensor("linear_135")]; tensor var_2817 = const()[name = tensor("op_2817"), val = tensor(0x1p-1)]; tensor var_2818 = mul(x = input_795, y = var_2817)[name = tensor("op_2818")]; tensor input_797 = add(x = input_785, y = var_2818)[name = tensor("input_797")]; tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; tensor input_799 = layer_norm(axes = input_799_axes_0, beta = encoder_layers_14_norm_out_bias, epsilon = var_11, gamma = encoder_layers_14_norm_out_weight, x = input_797)[name = tensor("input_799")]; tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; tensor input_801 = layer_norm(axes = input_801_axes_0, beta = encoder_layers_15_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_15_norm_feed_forward1_weight, x = input_799)[name = tensor("input_801")]; tensor input_803 = linear(bias = encoder_layers_15_feed_forward1_linear1_bias, weight = encoder_layers_15_feed_forward1_linear1_weight_quantized, x = input_801)[name = tensor("linear_136")]; tensor input_805 = silu(x = input_803)[name = tensor("input_805")]; tensor input_809 = linear(bias = encoder_layers_15_feed_forward1_linear2_bias, weight = encoder_layers_15_feed_forward1_linear2_weight_quantized, x = input_805)[name = tensor("linear_137")]; tensor var_2848 = const()[name = tensor("op_2848"), val = tensor(0x1p-1)]; tensor var_2849 = mul(x = input_809, y = var_2848)[name = tensor("op_2849")]; tensor input_811 = add(x = input_799, y = var_2849)[name = tensor("input_811")]; tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; tensor query_31 = layer_norm(axes = query_31_axes_0, beta = encoder_layers_15_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_15_norm_self_att_weight, x = input_811)[name = tensor("query_31")]; tensor var_2865 = linear(bias = encoder_layers_15_self_attn_linear_q_bias, weight = encoder_layers_15_self_attn_linear_q_weight_quantized, x = query_31)[name = tensor("linear_138")]; tensor var_2866 = const()[name = tensor("op_2866"), val = tensor([1, -1, 8, 128])]; tensor q_91 = reshape(shape = var_2866, x = var_2865)[name = tensor("q_91")]; tensor var_2870 = linear(bias = encoder_layers_15_self_attn_linear_k_bias, weight = encoder_layers_15_self_attn_linear_k_weight_quantized, x = query_31)[name = tensor("linear_139")]; tensor var_2871 = const()[name = tensor("op_2871"), val = tensor([1, -1, 8, 128])]; tensor k_61 = reshape(shape = var_2871, x = var_2870)[name = tensor("k_61")]; tensor var_2875 = linear(bias = encoder_layers_15_self_attn_linear_v_bias, weight = encoder_layers_15_self_attn_linear_v_weight_quantized, x = query_31)[name = tensor("linear_140")]; tensor var_2876 = const()[name = tensor("op_2876"), val = tensor([1, -1, 8, 128])]; tensor v_31 = reshape(shape = var_2876, x = var_2875)[name = tensor("v_31")]; tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_2888 = add(x = q_91, y = encoder_layers_15_self_attn_pos_bias_u)[name = tensor("op_2888")]; tensor var_2890 = add(x = q_91, y = encoder_layers_15_self_attn_pos_bias_v)[name = tensor("op_2890")]; tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_2892_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2892_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593746624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594131136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594130688)))]; tensor x_339_transpose_x_0 = const()[name = tensor("x_339_transpose_x_0"), val = tensor(false)]; tensor x_339_transpose_y_0 = const()[name = tensor("x_339_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_31 = transpose(perm = q_with_bias_v_31_perm_0, x = var_2890)[name = tensor("transpose_206")]; tensor x_339 = matmul(transpose_x = x_339_transpose_x_0, transpose_y = x_339_transpose_y_0, x = q_with_bias_v_31, y = op_2892_quantized)[name = tensor("x_339")]; tensor const_164 = const()[name = tensor("const_164"), val = tensor(0x0p+0)]; tensor x_341_pad_0 = const()[name = tensor("x_341_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_341_mode_0 = const()[name = tensor("x_341_mode_0"), val = tensor("constant")]; tensor x_341 = pad(constant_val = const_164, mode = x_341_mode_0, pad = x_341_pad_0, x = x_339)[name = tensor("x_341")]; tensor var_2900 = const()[name = tensor("op_2900"), val = tensor([1, 8, -1, 188])]; tensor x_343 = reshape(shape = var_2900, x = x_341)[name = tensor("x_343")]; tensor var_2904_begin_0 = const()[name = tensor("op_2904_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2904_end_0 = const()[name = tensor("op_2904_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2904_end_mask_0 = const()[name = tensor("op_2904_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2904 = slice_by_index(begin = var_2904_begin_0, end = var_2904_end_0, end_mask = var_2904_end_mask_0, x = x_343)[name = tensor("op_2904")]; tensor var_2905 = const()[name = tensor("op_2905"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_61 = reshape(shape = var_2905, x = var_2904)[name = tensor("matrix_bd_61")]; tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; tensor transpose_126_perm_0 = const()[name = tensor("transpose_126_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_127_perm_0 = const()[name = tensor("transpose_127_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_127 = transpose(perm = transpose_127_perm_0, x = k_61)[name = tensor("transpose_204")]; tensor transpose_126 = transpose(perm = transpose_126_perm_0, x = var_2888)[name = tensor("transpose_205")]; tensor matrix_ac_31 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_126, y = transpose_127)[name = tensor("matrix_ac_31")]; tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_63 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61)[name = tensor("matrix_bd_63")]; tensor var_2914 = add(x = matrix_ac_31, y = matrix_bd_63)[name = tensor("op_2914")]; tensor _inversed_scores_61_y_0 = const()[name = tensor("_inversed_scores_61_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_61 = mul(x = var_2914, y = _inversed_scores_61_y_0)[name = tensor("_inversed_scores_61")]; tensor scores_63 = select(a = var_14, b = _inversed_scores_61, cond = mask_3)[name = tensor("scores_63")]; tensor var_2920 = softmax(axis = var_32, x = scores_63)[name = tensor("op_2920")]; tensor input_813 = select(a = var_13, b = var_2920, cond = mask_3)[name = tensor("input_813")]; tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; tensor value_33 = transpose(perm = value_33_perm_0, x = v_31)[name = tensor("transpose_203")]; tensor x_345 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = input_813, y = value_33)[name = tensor("x_345")]; tensor var_2924_perm_0 = const()[name = tensor("op_2924_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2925 = const()[name = tensor("op_2925"), val = tensor([1, -1, 1024])]; tensor var_2924 = transpose(perm = var_2924_perm_0, x = x_345)[name = tensor("transpose_202")]; tensor input_815 = reshape(shape = var_2925, x = var_2924)[name = tensor("input_815")]; tensor input_817 = linear(bias = encoder_layers_15_self_attn_linear_out_bias, weight = encoder_layers_15_self_attn_linear_out_weight_quantized, x = input_815)[name = tensor("linear_142")]; tensor input_819 = add(x = input_811, y = input_817)[name = tensor("input_819")]; tensor x_349_axes_0 = const()[name = tensor("x_349_axes_0"), val = tensor([-1])]; tensor x_349 = layer_norm(axes = x_349_axes_0, beta = encoder_layers_15_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_15_norm_conv_weight, x = input_819)[name = tensor("x_349")]; tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; tensor input_821 = transpose(perm = input_821_perm_0, x = x_349)[name = tensor("transpose_201")]; tensor input_823 = conv(bias = encoder_layers_15_conv_pointwise_conv1_bias, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = encoder_layers_15_conv_pointwise_conv1_weight_quantized, x = input_821)[name = tensor("input_823")]; tensor x_351_split_num_splits_0 = const()[name = tensor("x_351_split_num_splits_0"), val = tensor(2)]; tensor x_351_split_axis_0 = const()[name = tensor("x_351_split_axis_0"), val = tensor(1)]; tensor x_351_split_0, tensor x_351_split_1 = split(axis = x_351_split_axis_0, num_splits = x_351_split_num_splits_0, x = input_823)[name = tensor("x_351_split")]; tensor x_351_split_1_sigmoid = sigmoid(x = x_351_split_1)[name = tensor("x_351_split_1_sigmoid")]; tensor x_351 = mul(x = x_351_split_0, y = x_351_split_1_sigmoid)[name = tensor("x_351")]; tensor input_825 = select(a = var_13, b = x_351, cond = var_339)[name = tensor("input_825")]; tensor const_167 = const()[name = tensor("const_167"), val = tensor(0x0p+0)]; tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; tensor input_827 = pad(constant_val = const_167, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825)[name = tensor("input_827")]; tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(1024)]; tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; tensor const_278_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_278_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594132736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594143104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594142016)))]; tensor const_279 = const()[name = tensor("const_279"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594147264)))]; tensor input_831 = conv(bias = const_279, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_278_quantized, x = input_827)[name = tensor("input_831")]; tensor input_833 = silu(x = input_831)[name = tensor("input_833")]; tensor x_353_pad_type_0 = const()[name = tensor("x_353_pad_type_0"), val = tensor("valid")]; tensor x_353_strides_0 = const()[name = tensor("x_353_strides_0"), val = tensor([1])]; tensor x_353_pad_0 = const()[name = tensor("x_353_pad_0"), val = tensor([0, 0])]; tensor x_353_dilations_0 = const()[name = tensor("x_353_dilations_0"), val = tensor([1])]; tensor x_353_groups_0 = const()[name = tensor("x_353_groups_0"), val = tensor(1)]; tensor x_353 = conv(bias = encoder_layers_15_conv_pointwise_conv2_bias, dilations = x_353_dilations_0, groups = x_353_groups_0, pad = x_353_pad_0, pad_type = x_353_pad_type_0, strides = x_353_strides_0, weight = encoder_layers_15_conv_pointwise_conv2_weight_quantized, x = input_833)[name = tensor("x_353")]; tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; tensor input_835 = transpose(perm = input_835_perm_0, x = x_353)[name = tensor("transpose_200")]; tensor input_837 = add(x = input_819, y = input_835)[name = tensor("input_837")]; tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; tensor input_839 = layer_norm(axes = input_839_axes_0, beta = encoder_layers_15_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_15_norm_feed_forward2_weight, x = input_837)[name = tensor("input_839")]; tensor input_841 = linear(bias = encoder_layers_15_feed_forward2_linear1_bias, weight = encoder_layers_15_feed_forward2_linear1_weight_quantized, x = input_839)[name = tensor("linear_143")]; tensor input_843 = silu(x = input_841)[name = tensor("input_843")]; tensor input_847 = linear(bias = encoder_layers_15_feed_forward2_linear2_bias, weight = encoder_layers_15_feed_forward2_linear2_weight_quantized, x = input_843)[name = tensor("linear_144")]; tensor var_2991 = const()[name = tensor("op_2991"), val = tensor(0x1p-1)]; tensor var_2992 = mul(x = input_847, y = var_2991)[name = tensor("op_2992")]; tensor input_849 = add(x = input_837, y = var_2992)[name = tensor("input_849")]; tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; tensor input_851 = layer_norm(axes = input_851_axes_0, beta = encoder_layers_15_norm_out_bias, epsilon = var_11, gamma = encoder_layers_15_norm_out_weight, x = input_849)[name = tensor("input_851")]; tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; tensor input_853 = layer_norm(axes = input_853_axes_0, beta = encoder_layers_16_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_16_norm_feed_forward1_weight, x = input_851)[name = tensor("input_853")]; tensor input_855 = linear(bias = encoder_layers_16_feed_forward1_linear1_bias, weight = encoder_layers_16_feed_forward1_linear1_weight_quantized, x = input_853)[name = tensor("linear_145")]; tensor input_857 = silu(x = input_855)[name = tensor("input_857")]; tensor input_861 = linear(bias = encoder_layers_16_feed_forward1_linear2_bias, weight = encoder_layers_16_feed_forward1_linear2_weight_quantized, x = input_857)[name = tensor("linear_146")]; tensor var_3022 = const()[name = tensor("op_3022"), val = tensor(0x1p-1)]; tensor var_3023 = mul(x = input_861, y = var_3022)[name = tensor("op_3023")]; tensor input_863 = add(x = input_851, y = var_3023)[name = tensor("input_863")]; tensor query_33_axes_0 = const()[name = tensor("query_33_axes_0"), val = tensor([-1])]; tensor query_33 = layer_norm(axes = query_33_axes_0, beta = encoder_layers_16_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_16_norm_self_att_weight, x = input_863)[name = tensor("query_33")]; tensor var_3039 = linear(bias = encoder_layers_16_self_attn_linear_q_bias, weight = encoder_layers_16_self_attn_linear_q_weight_quantized, x = query_33)[name = tensor("linear_147")]; tensor var_3040 = const()[name = tensor("op_3040"), val = tensor([1, -1, 8, 128])]; tensor q_97 = reshape(shape = var_3040, x = var_3039)[name = tensor("q_97")]; tensor var_3044 = linear(bias = encoder_layers_16_self_attn_linear_k_bias, weight = encoder_layers_16_self_attn_linear_k_weight_quantized, x = query_33)[name = tensor("linear_148")]; tensor var_3045 = const()[name = tensor("op_3045"), val = tensor([1, -1, 8, 128])]; tensor k_65 = reshape(shape = var_3045, x = var_3044)[name = tensor("k_65")]; tensor var_3049 = linear(bias = encoder_layers_16_self_attn_linear_v_bias, weight = encoder_layers_16_self_attn_linear_v_weight_quantized, x = query_33)[name = tensor("linear_149")]; tensor var_3050 = const()[name = tensor("op_3050"), val = tensor([1, -1, 8, 128])]; tensor v_33 = reshape(shape = var_3050, x = var_3049)[name = tensor("v_33")]; tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_3062 = add(x = q_97, y = encoder_layers_16_self_attn_pos_bias_u)[name = tensor("op_3062")]; tensor var_3064 = add(x = q_97, y = encoder_layers_16_self_attn_pos_bias_v)[name = tensor("op_3064")]; tensor q_with_bias_v_33_perm_0 = const()[name = tensor("q_with_bias_v_33_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_3066_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3066_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594151424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594535936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594535488)))]; tensor x_361_transpose_x_0 = const()[name = tensor("x_361_transpose_x_0"), val = tensor(false)]; tensor x_361_transpose_y_0 = const()[name = tensor("x_361_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_33 = transpose(perm = q_with_bias_v_33_perm_0, x = var_3064)[name = tensor("transpose_199")]; tensor x_361 = matmul(transpose_x = x_361_transpose_x_0, transpose_y = x_361_transpose_y_0, x = q_with_bias_v_33, y = op_3066_quantized)[name = tensor("x_361")]; tensor const_174 = const()[name = tensor("const_174"), val = tensor(0x0p+0)]; tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_363_mode_0 = const()[name = tensor("x_363_mode_0"), val = tensor("constant")]; tensor x_363 = pad(constant_val = const_174, mode = x_363_mode_0, pad = x_363_pad_0, x = x_361)[name = tensor("x_363")]; tensor var_3074 = const()[name = tensor("op_3074"), val = tensor([1, 8, -1, 188])]; tensor x_365 = reshape(shape = var_3074, x = x_363)[name = tensor("x_365")]; tensor var_3078_begin_0 = const()[name = tensor("op_3078_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3078_end_0 = const()[name = tensor("op_3078_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3078_end_mask_0 = const()[name = tensor("op_3078_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3078 = slice_by_index(begin = var_3078_begin_0, end = var_3078_end_0, end_mask = var_3078_end_mask_0, x = x_365)[name = tensor("op_3078")]; tensor var_3079 = const()[name = tensor("op_3079"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_65 = reshape(shape = var_3079, x = var_3078)[name = tensor("matrix_bd_65")]; tensor matrix_ac_33_transpose_x_0 = const()[name = tensor("matrix_ac_33_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_33_transpose_y_0 = const()[name = tensor("matrix_ac_33_transpose_y_0"), val = tensor(false)]; tensor transpose_128_perm_0 = const()[name = tensor("transpose_128_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_129_perm_0 = const()[name = tensor("transpose_129_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_129 = transpose(perm = transpose_129_perm_0, x = k_65)[name = tensor("transpose_197")]; tensor transpose_128 = transpose(perm = transpose_128_perm_0, x = var_3062)[name = tensor("transpose_198")]; tensor matrix_ac_33 = matmul(transpose_x = matrix_ac_33_transpose_x_0, transpose_y = matrix_ac_33_transpose_y_0, x = transpose_128, y = transpose_129)[name = tensor("matrix_ac_33")]; tensor matrix_bd_67_begin_0 = const()[name = tensor("matrix_bd_67_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_67_end_0 = const()[name = tensor("matrix_bd_67_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_67_end_mask_0 = const()[name = tensor("matrix_bd_67_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_67 = slice_by_index(begin = matrix_bd_67_begin_0, end = matrix_bd_67_end_0, end_mask = matrix_bd_67_end_mask_0, x = matrix_bd_65)[name = tensor("matrix_bd_67")]; tensor var_3088 = add(x = matrix_ac_33, y = matrix_bd_67)[name = tensor("op_3088")]; tensor _inversed_scores_65_y_0 = const()[name = tensor("_inversed_scores_65_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_65 = mul(x = var_3088, y = _inversed_scores_65_y_0)[name = tensor("_inversed_scores_65")]; tensor scores_67 = select(a = var_14, b = _inversed_scores_65, cond = mask_3)[name = tensor("scores_67")]; tensor var_3094 = softmax(axis = var_32, x = scores_67)[name = tensor("op_3094")]; tensor input_865 = select(a = var_13, b = var_3094, cond = mask_3)[name = tensor("input_865")]; tensor x_367_transpose_x_0 = const()[name = tensor("x_367_transpose_x_0"), val = tensor(false)]; tensor x_367_transpose_y_0 = const()[name = tensor("x_367_transpose_y_0"), val = tensor(false)]; tensor value_35 = transpose(perm = value_35_perm_0, x = v_33)[name = tensor("transpose_196")]; tensor x_367 = matmul(transpose_x = x_367_transpose_x_0, transpose_y = x_367_transpose_y_0, x = input_865, y = value_35)[name = tensor("x_367")]; tensor var_3098_perm_0 = const()[name = tensor("op_3098_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3099 = const()[name = tensor("op_3099"), val = tensor([1, -1, 1024])]; tensor var_3098 = transpose(perm = var_3098_perm_0, x = x_367)[name = tensor("transpose_195")]; tensor input_867 = reshape(shape = var_3099, x = var_3098)[name = tensor("input_867")]; tensor input_869 = linear(bias = encoder_layers_16_self_attn_linear_out_bias, weight = encoder_layers_16_self_attn_linear_out_weight_quantized, x = input_867)[name = tensor("linear_151")]; tensor input_871 = add(x = input_863, y = input_869)[name = tensor("input_871")]; tensor x_371_axes_0 = const()[name = tensor("x_371_axes_0"), val = tensor([-1])]; tensor x_371 = layer_norm(axes = x_371_axes_0, beta = encoder_layers_16_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_16_norm_conv_weight, x = input_871)[name = tensor("x_371")]; tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; tensor input_873 = transpose(perm = input_873_perm_0, x = x_371)[name = tensor("transpose_194")]; tensor input_875 = conv(bias = encoder_layers_16_conv_pointwise_conv1_bias, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = encoder_layers_16_conv_pointwise_conv1_weight_quantized, x = input_873)[name = tensor("input_875")]; tensor x_373_split_num_splits_0 = const()[name = tensor("x_373_split_num_splits_0"), val = tensor(2)]; tensor x_373_split_axis_0 = const()[name = tensor("x_373_split_axis_0"), val = tensor(1)]; tensor x_373_split_0, tensor x_373_split_1 = split(axis = x_373_split_axis_0, num_splits = x_373_split_num_splits_0, x = input_875)[name = tensor("x_373_split")]; tensor x_373_split_1_sigmoid = sigmoid(x = x_373_split_1)[name = tensor("x_373_split_1_sigmoid")]; tensor x_373 = mul(x = x_373_split_0, y = x_373_split_1_sigmoid)[name = tensor("x_373")]; tensor input_877 = select(a = var_13, b = x_373, cond = var_339)[name = tensor("input_877")]; tensor const_177 = const()[name = tensor("const_177"), val = tensor(0x0p+0)]; tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; tensor input_879 = pad(constant_val = const_177, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877)[name = tensor("input_879")]; tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(1024)]; tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; tensor const_280_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_280_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594537536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594547904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594546816)))]; tensor const_281 = const()[name = tensor("const_281"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594552064)))]; tensor input_883 = conv(bias = const_281, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_280_quantized, x = input_879)[name = tensor("input_883")]; tensor input_885 = silu(x = input_883)[name = tensor("input_885")]; tensor x_375_pad_type_0 = const()[name = tensor("x_375_pad_type_0"), val = tensor("valid")]; tensor x_375_strides_0 = const()[name = tensor("x_375_strides_0"), val = tensor([1])]; tensor x_375_pad_0 = const()[name = tensor("x_375_pad_0"), val = tensor([0, 0])]; tensor x_375_dilations_0 = const()[name = tensor("x_375_dilations_0"), val = tensor([1])]; tensor x_375_groups_0 = const()[name = tensor("x_375_groups_0"), val = tensor(1)]; tensor x_375 = conv(bias = encoder_layers_16_conv_pointwise_conv2_bias, dilations = x_375_dilations_0, groups = x_375_groups_0, pad = x_375_pad_0, pad_type = x_375_pad_type_0, strides = x_375_strides_0, weight = encoder_layers_16_conv_pointwise_conv2_weight_quantized, x = input_885)[name = tensor("x_375")]; tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; tensor input_887 = transpose(perm = input_887_perm_0, x = x_375)[name = tensor("transpose_193")]; tensor input_889 = add(x = input_871, y = input_887)[name = tensor("input_889")]; tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; tensor input_891 = layer_norm(axes = input_891_axes_0, beta = encoder_layers_16_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_16_norm_feed_forward2_weight, x = input_889)[name = tensor("input_891")]; tensor input_893 = linear(bias = encoder_layers_16_feed_forward2_linear1_bias, weight = encoder_layers_16_feed_forward2_linear1_weight_quantized, x = input_891)[name = tensor("linear_152")]; tensor input_895 = silu(x = input_893)[name = tensor("input_895")]; tensor input_899 = linear(bias = encoder_layers_16_feed_forward2_linear2_bias, weight = encoder_layers_16_feed_forward2_linear2_weight_quantized, x = input_895)[name = tensor("linear_153")]; tensor var_3165 = const()[name = tensor("op_3165"), val = tensor(0x1p-1)]; tensor var_3166 = mul(x = input_899, y = var_3165)[name = tensor("op_3166")]; tensor input_901 = add(x = input_889, y = var_3166)[name = tensor("input_901")]; tensor input_903_axes_0 = const()[name = tensor("input_903_axes_0"), val = tensor([-1])]; tensor input_903 = layer_norm(axes = input_903_axes_0, beta = encoder_layers_16_norm_out_bias, epsilon = var_11, gamma = encoder_layers_16_norm_out_weight, x = input_901)[name = tensor("input_903")]; tensor input_905_axes_0 = const()[name = tensor("input_905_axes_0"), val = tensor([-1])]; tensor input_905 = layer_norm(axes = input_905_axes_0, beta = encoder_layers_17_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_17_norm_feed_forward1_weight, x = input_903)[name = tensor("input_905")]; tensor input_907 = linear(bias = encoder_layers_17_feed_forward1_linear1_bias, weight = encoder_layers_17_feed_forward1_linear1_weight_quantized, x = input_905)[name = tensor("linear_154")]; tensor input_909 = silu(x = input_907)[name = tensor("input_909")]; tensor input_913 = linear(bias = encoder_layers_17_feed_forward1_linear2_bias, weight = encoder_layers_17_feed_forward1_linear2_weight_quantized, x = input_909)[name = tensor("linear_155")]; tensor var_3196 = const()[name = tensor("op_3196"), val = tensor(0x1p-1)]; tensor var_3197 = mul(x = input_913, y = var_3196)[name = tensor("op_3197")]; tensor input_915 = add(x = input_903, y = var_3197)[name = tensor("input_915")]; tensor query_35_axes_0 = const()[name = tensor("query_35_axes_0"), val = tensor([-1])]; tensor query_35 = layer_norm(axes = query_35_axes_0, beta = encoder_layers_17_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_17_norm_self_att_weight, x = input_915)[name = tensor("query_35")]; tensor var_3213 = linear(bias = encoder_layers_17_self_attn_linear_q_bias, weight = encoder_layers_17_self_attn_linear_q_weight_quantized, x = query_35)[name = tensor("linear_156")]; tensor var_3214 = const()[name = tensor("op_3214"), val = tensor([1, -1, 8, 128])]; tensor q_103 = reshape(shape = var_3214, x = var_3213)[name = tensor("q_103")]; tensor var_3218 = linear(bias = encoder_layers_17_self_attn_linear_k_bias, weight = encoder_layers_17_self_attn_linear_k_weight_quantized, x = query_35)[name = tensor("linear_157")]; tensor var_3219 = const()[name = tensor("op_3219"), val = tensor([1, -1, 8, 128])]; tensor k_69 = reshape(shape = var_3219, x = var_3218)[name = tensor("k_69")]; tensor var_3223 = linear(bias = encoder_layers_17_self_attn_linear_v_bias, weight = encoder_layers_17_self_attn_linear_v_weight_quantized, x = query_35)[name = tensor("linear_158")]; tensor var_3224 = const()[name = tensor("op_3224"), val = tensor([1, -1, 8, 128])]; tensor v_35 = reshape(shape = var_3224, x = var_3223)[name = tensor("v_35")]; tensor value_37_perm_0 = const()[name = tensor("value_37_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_3236 = add(x = q_103, y = encoder_layers_17_self_attn_pos_bias_u)[name = tensor("op_3236")]; tensor var_3238 = add(x = q_103, y = encoder_layers_17_self_attn_pos_bias_v)[name = tensor("op_3238")]; tensor q_with_bias_v_35_perm_0 = const()[name = tensor("q_with_bias_v_35_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_3240_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3240_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594556224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594940736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594940288)))]; tensor x_383_transpose_x_0 = const()[name = tensor("x_383_transpose_x_0"), val = tensor(false)]; tensor x_383_transpose_y_0 = const()[name = tensor("x_383_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_35 = transpose(perm = q_with_bias_v_35_perm_0, x = var_3238)[name = tensor("transpose_192")]; tensor x_383 = matmul(transpose_x = x_383_transpose_x_0, transpose_y = x_383_transpose_y_0, x = q_with_bias_v_35, y = op_3240_quantized)[name = tensor("x_383")]; tensor const_184 = const()[name = tensor("const_184"), val = tensor(0x0p+0)]; tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_385_mode_0 = const()[name = tensor("x_385_mode_0"), val = tensor("constant")]; tensor x_385 = pad(constant_val = const_184, mode = x_385_mode_0, pad = x_385_pad_0, x = x_383)[name = tensor("x_385")]; tensor var_3248 = const()[name = tensor("op_3248"), val = tensor([1, 8, -1, 188])]; tensor x_387 = reshape(shape = var_3248, x = x_385)[name = tensor("x_387")]; tensor var_3252_begin_0 = const()[name = tensor("op_3252_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3252_end_0 = const()[name = tensor("op_3252_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3252_end_mask_0 = const()[name = tensor("op_3252_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3252 = slice_by_index(begin = var_3252_begin_0, end = var_3252_end_0, end_mask = var_3252_end_mask_0, x = x_387)[name = tensor("op_3252")]; tensor var_3253 = const()[name = tensor("op_3253"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_69 = reshape(shape = var_3253, x = var_3252)[name = tensor("matrix_bd_69")]; tensor matrix_ac_35_transpose_x_0 = const()[name = tensor("matrix_ac_35_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_35_transpose_y_0 = const()[name = tensor("matrix_ac_35_transpose_y_0"), val = tensor(false)]; tensor transpose_130_perm_0 = const()[name = tensor("transpose_130_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_131_perm_0 = const()[name = tensor("transpose_131_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_131 = transpose(perm = transpose_131_perm_0, x = k_69)[name = tensor("transpose_190")]; tensor transpose_130 = transpose(perm = transpose_130_perm_0, x = var_3236)[name = tensor("transpose_191")]; tensor matrix_ac_35 = matmul(transpose_x = matrix_ac_35_transpose_x_0, transpose_y = matrix_ac_35_transpose_y_0, x = transpose_130, y = transpose_131)[name = tensor("matrix_ac_35")]; tensor matrix_bd_71_begin_0 = const()[name = tensor("matrix_bd_71_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_71_end_0 = const()[name = tensor("matrix_bd_71_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_71_end_mask_0 = const()[name = tensor("matrix_bd_71_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_71 = slice_by_index(begin = matrix_bd_71_begin_0, end = matrix_bd_71_end_0, end_mask = matrix_bd_71_end_mask_0, x = matrix_bd_69)[name = tensor("matrix_bd_71")]; tensor var_3262 = add(x = matrix_ac_35, y = matrix_bd_71)[name = tensor("op_3262")]; tensor _inversed_scores_69_y_0 = const()[name = tensor("_inversed_scores_69_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_69 = mul(x = var_3262, y = _inversed_scores_69_y_0)[name = tensor("_inversed_scores_69")]; tensor scores_71 = select(a = var_14, b = _inversed_scores_69, cond = mask_3)[name = tensor("scores_71")]; tensor var_3268 = softmax(axis = var_32, x = scores_71)[name = tensor("op_3268")]; tensor input_917 = select(a = var_13, b = var_3268, cond = mask_3)[name = tensor("input_917")]; tensor x_389_transpose_x_0 = const()[name = tensor("x_389_transpose_x_0"), val = tensor(false)]; tensor x_389_transpose_y_0 = const()[name = tensor("x_389_transpose_y_0"), val = tensor(false)]; tensor value_37 = transpose(perm = value_37_perm_0, x = v_35)[name = tensor("transpose_189")]; tensor x_389 = matmul(transpose_x = x_389_transpose_x_0, transpose_y = x_389_transpose_y_0, x = input_917, y = value_37)[name = tensor("x_389")]; tensor var_3272_perm_0 = const()[name = tensor("op_3272_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3273 = const()[name = tensor("op_3273"), val = tensor([1, -1, 1024])]; tensor var_3272 = transpose(perm = var_3272_perm_0, x = x_389)[name = tensor("transpose_188")]; tensor input_919 = reshape(shape = var_3273, x = var_3272)[name = tensor("input_919")]; tensor input_921 = linear(bias = encoder_layers_17_self_attn_linear_out_bias, weight = encoder_layers_17_self_attn_linear_out_weight_quantized, x = input_919)[name = tensor("linear_160")]; tensor input_923 = add(x = input_915, y = input_921)[name = tensor("input_923")]; tensor x_393_axes_0 = const()[name = tensor("x_393_axes_0"), val = tensor([-1])]; tensor x_393 = layer_norm(axes = x_393_axes_0, beta = encoder_layers_17_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_17_norm_conv_weight, x = input_923)[name = tensor("x_393")]; tensor input_925_perm_0 = const()[name = tensor("input_925_perm_0"), val = tensor([0, 2, 1])]; tensor input_927_pad_type_0 = const()[name = tensor("input_927_pad_type_0"), val = tensor("valid")]; tensor input_927_strides_0 = const()[name = tensor("input_927_strides_0"), val = tensor([1])]; tensor input_927_pad_0 = const()[name = tensor("input_927_pad_0"), val = tensor([0, 0])]; tensor input_927_dilations_0 = const()[name = tensor("input_927_dilations_0"), val = tensor([1])]; tensor input_927_groups_0 = const()[name = tensor("input_927_groups_0"), val = tensor(1)]; tensor input_925 = transpose(perm = input_925_perm_0, x = x_393)[name = tensor("transpose_187")]; tensor input_927 = conv(bias = encoder_layers_17_conv_pointwise_conv1_bias, dilations = input_927_dilations_0, groups = input_927_groups_0, pad = input_927_pad_0, pad_type = input_927_pad_type_0, strides = input_927_strides_0, weight = encoder_layers_17_conv_pointwise_conv1_weight_quantized, x = input_925)[name = tensor("input_927")]; tensor x_395_split_num_splits_0 = const()[name = tensor("x_395_split_num_splits_0"), val = tensor(2)]; tensor x_395_split_axis_0 = const()[name = tensor("x_395_split_axis_0"), val = tensor(1)]; tensor x_395_split_0, tensor x_395_split_1 = split(axis = x_395_split_axis_0, num_splits = x_395_split_num_splits_0, x = input_927)[name = tensor("x_395_split")]; tensor x_395_split_1_sigmoid = sigmoid(x = x_395_split_1)[name = tensor("x_395_split_1_sigmoid")]; tensor x_395 = mul(x = x_395_split_0, y = x_395_split_1_sigmoid)[name = tensor("x_395")]; tensor input_929 = select(a = var_13, b = x_395, cond = var_339)[name = tensor("input_929")]; tensor const_187 = const()[name = tensor("const_187"), val = tensor(0x0p+0)]; tensor input_931_pad_0 = const()[name = tensor("input_931_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_931_mode_0 = const()[name = tensor("input_931_mode_0"), val = tensor("constant")]; tensor input_931 = pad(constant_val = const_187, mode = input_931_mode_0, pad = input_931_pad_0, x = input_929)[name = tensor("input_931")]; tensor input_933_pad_type_0 = const()[name = tensor("input_933_pad_type_0"), val = tensor("valid")]; tensor input_933_groups_0 = const()[name = tensor("input_933_groups_0"), val = tensor(1024)]; tensor input_933_strides_0 = const()[name = tensor("input_933_strides_0"), val = tensor([1])]; tensor input_933_pad_0 = const()[name = tensor("input_933_pad_0"), val = tensor([0, 0])]; tensor input_933_dilations_0 = const()[name = tensor("input_933_dilations_0"), val = tensor([1])]; tensor const_282_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_282_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594942336))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594952704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594951616)))]; tensor const_283 = const()[name = tensor("const_283"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594956864)))]; tensor input_935 = conv(bias = const_283, dilations = input_933_dilations_0, groups = input_933_groups_0, pad = input_933_pad_0, pad_type = input_933_pad_type_0, strides = input_933_strides_0, weight = const_282_quantized, x = input_931)[name = tensor("input_935")]; tensor input_937 = silu(x = input_935)[name = tensor("input_937")]; tensor x_397_pad_type_0 = const()[name = tensor("x_397_pad_type_0"), val = tensor("valid")]; tensor x_397_strides_0 = const()[name = tensor("x_397_strides_0"), val = tensor([1])]; tensor x_397_pad_0 = const()[name = tensor("x_397_pad_0"), val = tensor([0, 0])]; tensor x_397_dilations_0 = const()[name = tensor("x_397_dilations_0"), val = tensor([1])]; tensor x_397_groups_0 = const()[name = tensor("x_397_groups_0"), val = tensor(1)]; tensor x_397 = conv(bias = encoder_layers_17_conv_pointwise_conv2_bias, dilations = x_397_dilations_0, groups = x_397_groups_0, pad = x_397_pad_0, pad_type = x_397_pad_type_0, strides = x_397_strides_0, weight = encoder_layers_17_conv_pointwise_conv2_weight_quantized, x = input_937)[name = tensor("x_397")]; tensor input_939_perm_0 = const()[name = tensor("input_939_perm_0"), val = tensor([0, 2, 1])]; tensor input_939 = transpose(perm = input_939_perm_0, x = x_397)[name = tensor("transpose_186")]; tensor input_941 = add(x = input_923, y = input_939)[name = tensor("input_941")]; tensor input_943_axes_0 = const()[name = tensor("input_943_axes_0"), val = tensor([-1])]; tensor input_943 = layer_norm(axes = input_943_axes_0, beta = encoder_layers_17_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_17_norm_feed_forward2_weight, x = input_941)[name = tensor("input_943")]; tensor input_945 = linear(bias = encoder_layers_17_feed_forward2_linear1_bias, weight = encoder_layers_17_feed_forward2_linear1_weight_quantized, x = input_943)[name = tensor("linear_161")]; tensor input_947 = silu(x = input_945)[name = tensor("input_947")]; tensor input_951 = linear(bias = encoder_layers_17_feed_forward2_linear2_bias, weight = encoder_layers_17_feed_forward2_linear2_weight_quantized, x = input_947)[name = tensor("linear_162")]; tensor var_3339 = const()[name = tensor("op_3339"), val = tensor(0x1p-1)]; tensor var_3340 = mul(x = input_951, y = var_3339)[name = tensor("op_3340")]; tensor input_953 = add(x = input_941, y = var_3340)[name = tensor("input_953")]; tensor input_955_axes_0 = const()[name = tensor("input_955_axes_0"), val = tensor([-1])]; tensor input_955 = layer_norm(axes = input_955_axes_0, beta = encoder_layers_17_norm_out_bias, epsilon = var_11, gamma = encoder_layers_17_norm_out_weight, x = input_953)[name = tensor("input_955")]; tensor input_957_axes_0 = const()[name = tensor("input_957_axes_0"), val = tensor([-1])]; tensor input_957 = layer_norm(axes = input_957_axes_0, beta = encoder_layers_18_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_18_norm_feed_forward1_weight, x = input_955)[name = tensor("input_957")]; tensor input_959 = linear(bias = encoder_layers_18_feed_forward1_linear1_bias, weight = encoder_layers_18_feed_forward1_linear1_weight_quantized, x = input_957)[name = tensor("linear_163")]; tensor input_961 = silu(x = input_959)[name = tensor("input_961")]; tensor input_965 = linear(bias = encoder_layers_18_feed_forward1_linear2_bias, weight = encoder_layers_18_feed_forward1_linear2_weight_quantized, x = input_961)[name = tensor("linear_164")]; tensor var_3370 = const()[name = tensor("op_3370"), val = tensor(0x1p-1)]; tensor var_3371 = mul(x = input_965, y = var_3370)[name = tensor("op_3371")]; tensor input_967 = add(x = input_955, y = var_3371)[name = tensor("input_967")]; tensor query_37_axes_0 = const()[name = tensor("query_37_axes_0"), val = tensor([-1])]; tensor query_37 = layer_norm(axes = query_37_axes_0, beta = encoder_layers_18_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_18_norm_self_att_weight, x = input_967)[name = tensor("query_37")]; tensor var_3387 = linear(bias = encoder_layers_18_self_attn_linear_q_bias, weight = encoder_layers_18_self_attn_linear_q_weight_quantized, x = query_37)[name = tensor("linear_165")]; tensor var_3388 = const()[name = tensor("op_3388"), val = tensor([1, -1, 8, 128])]; tensor q_109 = reshape(shape = var_3388, x = var_3387)[name = tensor("q_109")]; tensor var_3392 = linear(bias = encoder_layers_18_self_attn_linear_k_bias, weight = encoder_layers_18_self_attn_linear_k_weight_quantized, x = query_37)[name = tensor("linear_166")]; tensor var_3393 = const()[name = tensor("op_3393"), val = tensor([1, -1, 8, 128])]; tensor k_73 = reshape(shape = var_3393, x = var_3392)[name = tensor("k_73")]; tensor var_3397 = linear(bias = encoder_layers_18_self_attn_linear_v_bias, weight = encoder_layers_18_self_attn_linear_v_weight_quantized, x = query_37)[name = tensor("linear_167")]; tensor var_3398 = const()[name = tensor("op_3398"), val = tensor([1, -1, 8, 128])]; tensor v_37 = reshape(shape = var_3398, x = var_3397)[name = tensor("v_37")]; tensor value_39_perm_0 = const()[name = tensor("value_39_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_3410 = add(x = q_109, y = encoder_layers_18_self_attn_pos_bias_u)[name = tensor("op_3410")]; tensor var_3412 = add(x = q_109, y = encoder_layers_18_self_attn_pos_bias_v)[name = tensor("op_3412")]; tensor q_with_bias_v_37_perm_0 = const()[name = tensor("q_with_bias_v_37_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_3414_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3414_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594961024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595345536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595345088)))]; tensor x_405_transpose_x_0 = const()[name = tensor("x_405_transpose_x_0"), val = tensor(false)]; tensor x_405_transpose_y_0 = const()[name = tensor("x_405_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_37 = transpose(perm = q_with_bias_v_37_perm_0, x = var_3412)[name = tensor("transpose_185")]; tensor x_405 = matmul(transpose_x = x_405_transpose_x_0, transpose_y = x_405_transpose_y_0, x = q_with_bias_v_37, y = op_3414_quantized)[name = tensor("x_405")]; tensor const_194 = const()[name = tensor("const_194"), val = tensor(0x0p+0)]; tensor x_407_pad_0 = const()[name = tensor("x_407_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_407_mode_0 = const()[name = tensor("x_407_mode_0"), val = tensor("constant")]; tensor x_407 = pad(constant_val = const_194, mode = x_407_mode_0, pad = x_407_pad_0, x = x_405)[name = tensor("x_407")]; tensor var_3422 = const()[name = tensor("op_3422"), val = tensor([1, 8, -1, 188])]; tensor x_409 = reshape(shape = var_3422, x = x_407)[name = tensor("x_409")]; tensor var_3426_begin_0 = const()[name = tensor("op_3426_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3426_end_0 = const()[name = tensor("op_3426_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3426_end_mask_0 = const()[name = tensor("op_3426_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3426 = slice_by_index(begin = var_3426_begin_0, end = var_3426_end_0, end_mask = var_3426_end_mask_0, x = x_409)[name = tensor("op_3426")]; tensor var_3427 = const()[name = tensor("op_3427"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_73 = reshape(shape = var_3427, x = var_3426)[name = tensor("matrix_bd_73")]; tensor matrix_ac_37_transpose_x_0 = const()[name = tensor("matrix_ac_37_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_37_transpose_y_0 = const()[name = tensor("matrix_ac_37_transpose_y_0"), val = tensor(false)]; tensor transpose_132_perm_0 = const()[name = tensor("transpose_132_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_133_perm_0 = const()[name = tensor("transpose_133_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_133 = transpose(perm = transpose_133_perm_0, x = k_73)[name = tensor("transpose_183")]; tensor transpose_132 = transpose(perm = transpose_132_perm_0, x = var_3410)[name = tensor("transpose_184")]; tensor matrix_ac_37 = matmul(transpose_x = matrix_ac_37_transpose_x_0, transpose_y = matrix_ac_37_transpose_y_0, x = transpose_132, y = transpose_133)[name = tensor("matrix_ac_37")]; tensor matrix_bd_75_begin_0 = const()[name = tensor("matrix_bd_75_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_75_end_0 = const()[name = tensor("matrix_bd_75_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_75_end_mask_0 = const()[name = tensor("matrix_bd_75_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_75 = slice_by_index(begin = matrix_bd_75_begin_0, end = matrix_bd_75_end_0, end_mask = matrix_bd_75_end_mask_0, x = matrix_bd_73)[name = tensor("matrix_bd_75")]; tensor var_3436 = add(x = matrix_ac_37, y = matrix_bd_75)[name = tensor("op_3436")]; tensor _inversed_scores_73_y_0 = const()[name = tensor("_inversed_scores_73_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_73 = mul(x = var_3436, y = _inversed_scores_73_y_0)[name = tensor("_inversed_scores_73")]; tensor scores_75 = select(a = var_14, b = _inversed_scores_73, cond = mask_3)[name = tensor("scores_75")]; tensor var_3442 = softmax(axis = var_32, x = scores_75)[name = tensor("op_3442")]; tensor input_969 = select(a = var_13, b = var_3442, cond = mask_3)[name = tensor("input_969")]; tensor x_411_transpose_x_0 = const()[name = tensor("x_411_transpose_x_0"), val = tensor(false)]; tensor x_411_transpose_y_0 = const()[name = tensor("x_411_transpose_y_0"), val = tensor(false)]; tensor value_39 = transpose(perm = value_39_perm_0, x = v_37)[name = tensor("transpose_182")]; tensor x_411 = matmul(transpose_x = x_411_transpose_x_0, transpose_y = x_411_transpose_y_0, x = input_969, y = value_39)[name = tensor("x_411")]; tensor var_3446_perm_0 = const()[name = tensor("op_3446_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3447 = const()[name = tensor("op_3447"), val = tensor([1, -1, 1024])]; tensor var_3446 = transpose(perm = var_3446_perm_0, x = x_411)[name = tensor("transpose_181")]; tensor input_971 = reshape(shape = var_3447, x = var_3446)[name = tensor("input_971")]; tensor input_973 = linear(bias = encoder_layers_18_self_attn_linear_out_bias, weight = encoder_layers_18_self_attn_linear_out_weight_quantized, x = input_971)[name = tensor("linear_169")]; tensor input_975 = add(x = input_967, y = input_973)[name = tensor("input_975")]; tensor x_415_axes_0 = const()[name = tensor("x_415_axes_0"), val = tensor([-1])]; tensor x_415 = layer_norm(axes = x_415_axes_0, beta = encoder_layers_18_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_18_norm_conv_weight, x = input_975)[name = tensor("x_415")]; tensor input_977_perm_0 = const()[name = tensor("input_977_perm_0"), val = tensor([0, 2, 1])]; tensor input_979_pad_type_0 = const()[name = tensor("input_979_pad_type_0"), val = tensor("valid")]; tensor input_979_strides_0 = const()[name = tensor("input_979_strides_0"), val = tensor([1])]; tensor input_979_pad_0 = const()[name = tensor("input_979_pad_0"), val = tensor([0, 0])]; tensor input_979_dilations_0 = const()[name = tensor("input_979_dilations_0"), val = tensor([1])]; tensor input_979_groups_0 = const()[name = tensor("input_979_groups_0"), val = tensor(1)]; tensor input_977 = transpose(perm = input_977_perm_0, x = x_415)[name = tensor("transpose_180")]; tensor input_979 = conv(bias = encoder_layers_18_conv_pointwise_conv1_bias, dilations = input_979_dilations_0, groups = input_979_groups_0, pad = input_979_pad_0, pad_type = input_979_pad_type_0, strides = input_979_strides_0, weight = encoder_layers_18_conv_pointwise_conv1_weight_quantized, x = input_977)[name = tensor("input_979")]; tensor x_417_split_num_splits_0 = const()[name = tensor("x_417_split_num_splits_0"), val = tensor(2)]; tensor x_417_split_axis_0 = const()[name = tensor("x_417_split_axis_0"), val = tensor(1)]; tensor x_417_split_0, tensor x_417_split_1 = split(axis = x_417_split_axis_0, num_splits = x_417_split_num_splits_0, x = input_979)[name = tensor("x_417_split")]; tensor x_417_split_1_sigmoid = sigmoid(x = x_417_split_1)[name = tensor("x_417_split_1_sigmoid")]; tensor x_417 = mul(x = x_417_split_0, y = x_417_split_1_sigmoid)[name = tensor("x_417")]; tensor input_981 = select(a = var_13, b = x_417, cond = var_339)[name = tensor("input_981")]; tensor const_197 = const()[name = tensor("const_197"), val = tensor(0x0p+0)]; tensor input_983_pad_0 = const()[name = tensor("input_983_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_983_mode_0 = const()[name = tensor("input_983_mode_0"), val = tensor("constant")]; tensor input_983 = pad(constant_val = const_197, mode = input_983_mode_0, pad = input_983_pad_0, x = input_981)[name = tensor("input_983")]; tensor input_985_pad_type_0 = const()[name = tensor("input_985_pad_type_0"), val = tensor("valid")]; tensor input_985_groups_0 = const()[name = tensor("input_985_groups_0"), val = tensor(1024)]; tensor input_985_strides_0 = const()[name = tensor("input_985_strides_0"), val = tensor([1])]; tensor input_985_pad_0 = const()[name = tensor("input_985_pad_0"), val = tensor([0, 0])]; tensor input_985_dilations_0 = const()[name = tensor("input_985_dilations_0"), val = tensor([1])]; tensor const_284_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_284_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595347136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595357504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595356416)))]; tensor const_285 = const()[name = tensor("const_285"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595361664)))]; tensor input_987 = conv(bias = const_285, dilations = input_985_dilations_0, groups = input_985_groups_0, pad = input_985_pad_0, pad_type = input_985_pad_type_0, strides = input_985_strides_0, weight = const_284_quantized, x = input_983)[name = tensor("input_987")]; tensor input_989 = silu(x = input_987)[name = tensor("input_989")]; tensor x_419_pad_type_0 = const()[name = tensor("x_419_pad_type_0"), val = tensor("valid")]; tensor x_419_strides_0 = const()[name = tensor("x_419_strides_0"), val = tensor([1])]; tensor x_419_pad_0 = const()[name = tensor("x_419_pad_0"), val = tensor([0, 0])]; tensor x_419_dilations_0 = const()[name = tensor("x_419_dilations_0"), val = tensor([1])]; tensor x_419_groups_0 = const()[name = tensor("x_419_groups_0"), val = tensor(1)]; tensor x_419 = conv(bias = encoder_layers_18_conv_pointwise_conv2_bias, dilations = x_419_dilations_0, groups = x_419_groups_0, pad = x_419_pad_0, pad_type = x_419_pad_type_0, strides = x_419_strides_0, weight = encoder_layers_18_conv_pointwise_conv2_weight_quantized, x = input_989)[name = tensor("x_419")]; tensor input_991_perm_0 = const()[name = tensor("input_991_perm_0"), val = tensor([0, 2, 1])]; tensor input_991 = transpose(perm = input_991_perm_0, x = x_419)[name = tensor("transpose_179")]; tensor input_993 = add(x = input_975, y = input_991)[name = tensor("input_993")]; tensor input_995_axes_0 = const()[name = tensor("input_995_axes_0"), val = tensor([-1])]; tensor input_995 = layer_norm(axes = input_995_axes_0, beta = encoder_layers_18_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_18_norm_feed_forward2_weight, x = input_993)[name = tensor("input_995")]; tensor input_997 = linear(bias = encoder_layers_18_feed_forward2_linear1_bias, weight = encoder_layers_18_feed_forward2_linear1_weight_quantized, x = input_995)[name = tensor("linear_170")]; tensor input_999 = silu(x = input_997)[name = tensor("input_999")]; tensor input_1003 = linear(bias = encoder_layers_18_feed_forward2_linear2_bias, weight = encoder_layers_18_feed_forward2_linear2_weight_quantized, x = input_999)[name = tensor("linear_171")]; tensor var_3513 = const()[name = tensor("op_3513"), val = tensor(0x1p-1)]; tensor var_3514 = mul(x = input_1003, y = var_3513)[name = tensor("op_3514")]; tensor input_1005 = add(x = input_993, y = var_3514)[name = tensor("input_1005")]; tensor input_1007_axes_0 = const()[name = tensor("input_1007_axes_0"), val = tensor([-1])]; tensor input_1007 = layer_norm(axes = input_1007_axes_0, beta = encoder_layers_18_norm_out_bias, epsilon = var_11, gamma = encoder_layers_18_norm_out_weight, x = input_1005)[name = tensor("input_1007")]; tensor input_1009_axes_0 = const()[name = tensor("input_1009_axes_0"), val = tensor([-1])]; tensor input_1009 = layer_norm(axes = input_1009_axes_0, beta = encoder_layers_19_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_19_norm_feed_forward1_weight, x = input_1007)[name = tensor("input_1009")]; tensor input_1011 = linear(bias = encoder_layers_19_feed_forward1_linear1_bias, weight = encoder_layers_19_feed_forward1_linear1_weight_quantized, x = input_1009)[name = tensor("linear_172")]; tensor input_1013 = silu(x = input_1011)[name = tensor("input_1013")]; tensor input_1017 = linear(bias = encoder_layers_19_feed_forward1_linear2_bias, weight = encoder_layers_19_feed_forward1_linear2_weight_quantized, x = input_1013)[name = tensor("linear_173")]; tensor var_3544 = const()[name = tensor("op_3544"), val = tensor(0x1p-1)]; tensor var_3545 = mul(x = input_1017, y = var_3544)[name = tensor("op_3545")]; tensor input_1019 = add(x = input_1007, y = var_3545)[name = tensor("input_1019")]; tensor query_39_axes_0 = const()[name = tensor("query_39_axes_0"), val = tensor([-1])]; tensor query_39 = layer_norm(axes = query_39_axes_0, beta = encoder_layers_19_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_19_norm_self_att_weight, x = input_1019)[name = tensor("query_39")]; tensor var_3561 = linear(bias = encoder_layers_19_self_attn_linear_q_bias, weight = encoder_layers_19_self_attn_linear_q_weight_quantized, x = query_39)[name = tensor("linear_174")]; tensor var_3562 = const()[name = tensor("op_3562"), val = tensor([1, -1, 8, 128])]; tensor q_115 = reshape(shape = var_3562, x = var_3561)[name = tensor("q_115")]; tensor var_3566 = linear(bias = encoder_layers_19_self_attn_linear_k_bias, weight = encoder_layers_19_self_attn_linear_k_weight_quantized, x = query_39)[name = tensor("linear_175")]; tensor var_3567 = const()[name = tensor("op_3567"), val = tensor([1, -1, 8, 128])]; tensor k_77 = reshape(shape = var_3567, x = var_3566)[name = tensor("k_77")]; tensor var_3571 = linear(bias = encoder_layers_19_self_attn_linear_v_bias, weight = encoder_layers_19_self_attn_linear_v_weight_quantized, x = query_39)[name = tensor("linear_176")]; tensor var_3572 = const()[name = tensor("op_3572"), val = tensor([1, -1, 8, 128])]; tensor v_39 = reshape(shape = var_3572, x = var_3571)[name = tensor("v_39")]; tensor value_41_perm_0 = const()[name = tensor("value_41_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_3584 = add(x = q_115, y = encoder_layers_19_self_attn_pos_bias_u)[name = tensor("op_3584")]; tensor var_3586 = add(x = q_115, y = encoder_layers_19_self_attn_pos_bias_v)[name = tensor("op_3586")]; tensor q_with_bias_v_39_perm_0 = const()[name = tensor("q_with_bias_v_39_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_3588_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3588_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595365824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595750336))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595749888)))]; tensor x_427_transpose_x_0 = const()[name = tensor("x_427_transpose_x_0"), val = tensor(false)]; tensor x_427_transpose_y_0 = const()[name = tensor("x_427_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_39 = transpose(perm = q_with_bias_v_39_perm_0, x = var_3586)[name = tensor("transpose_178")]; tensor x_427 = matmul(transpose_x = x_427_transpose_x_0, transpose_y = x_427_transpose_y_0, x = q_with_bias_v_39, y = op_3588_quantized)[name = tensor("x_427")]; tensor const_204 = const()[name = tensor("const_204"), val = tensor(0x0p+0)]; tensor x_429_pad_0 = const()[name = tensor("x_429_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_429_mode_0 = const()[name = tensor("x_429_mode_0"), val = tensor("constant")]; tensor x_429 = pad(constant_val = const_204, mode = x_429_mode_0, pad = x_429_pad_0, x = x_427)[name = tensor("x_429")]; tensor var_3596 = const()[name = tensor("op_3596"), val = tensor([1, 8, -1, 188])]; tensor x_431 = reshape(shape = var_3596, x = x_429)[name = tensor("x_431")]; tensor var_3600_begin_0 = const()[name = tensor("op_3600_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3600_end_0 = const()[name = tensor("op_3600_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3600_end_mask_0 = const()[name = tensor("op_3600_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3600 = slice_by_index(begin = var_3600_begin_0, end = var_3600_end_0, end_mask = var_3600_end_mask_0, x = x_431)[name = tensor("op_3600")]; tensor var_3601 = const()[name = tensor("op_3601"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_77 = reshape(shape = var_3601, x = var_3600)[name = tensor("matrix_bd_77")]; tensor matrix_ac_39_transpose_x_0 = const()[name = tensor("matrix_ac_39_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_39_transpose_y_0 = const()[name = tensor("matrix_ac_39_transpose_y_0"), val = tensor(false)]; tensor transpose_134_perm_0 = const()[name = tensor("transpose_134_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_135_perm_0 = const()[name = tensor("transpose_135_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_135 = transpose(perm = transpose_135_perm_0, x = k_77)[name = tensor("transpose_176")]; tensor transpose_134 = transpose(perm = transpose_134_perm_0, x = var_3584)[name = tensor("transpose_177")]; tensor matrix_ac_39 = matmul(transpose_x = matrix_ac_39_transpose_x_0, transpose_y = matrix_ac_39_transpose_y_0, x = transpose_134, y = transpose_135)[name = tensor("matrix_ac_39")]; tensor matrix_bd_79_begin_0 = const()[name = tensor("matrix_bd_79_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_79_end_0 = const()[name = tensor("matrix_bd_79_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_79_end_mask_0 = const()[name = tensor("matrix_bd_79_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_79 = slice_by_index(begin = matrix_bd_79_begin_0, end = matrix_bd_79_end_0, end_mask = matrix_bd_79_end_mask_0, x = matrix_bd_77)[name = tensor("matrix_bd_79")]; tensor var_3610 = add(x = matrix_ac_39, y = matrix_bd_79)[name = tensor("op_3610")]; tensor _inversed_scores_77_y_0 = const()[name = tensor("_inversed_scores_77_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_77 = mul(x = var_3610, y = _inversed_scores_77_y_0)[name = tensor("_inversed_scores_77")]; tensor scores_79 = select(a = var_14, b = _inversed_scores_77, cond = mask_3)[name = tensor("scores_79")]; tensor var_3616 = softmax(axis = var_32, x = scores_79)[name = tensor("op_3616")]; tensor input_1021 = select(a = var_13, b = var_3616, cond = mask_3)[name = tensor("input_1021")]; tensor x_433_transpose_x_0 = const()[name = tensor("x_433_transpose_x_0"), val = tensor(false)]; tensor x_433_transpose_y_0 = const()[name = tensor("x_433_transpose_y_0"), val = tensor(false)]; tensor value_41 = transpose(perm = value_41_perm_0, x = v_39)[name = tensor("transpose_175")]; tensor x_433 = matmul(transpose_x = x_433_transpose_x_0, transpose_y = x_433_transpose_y_0, x = input_1021, y = value_41)[name = tensor("x_433")]; tensor var_3620_perm_0 = const()[name = tensor("op_3620_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3621 = const()[name = tensor("op_3621"), val = tensor([1, -1, 1024])]; tensor var_3620 = transpose(perm = var_3620_perm_0, x = x_433)[name = tensor("transpose_174")]; tensor input_1023 = reshape(shape = var_3621, x = var_3620)[name = tensor("input_1023")]; tensor input_1025 = linear(bias = encoder_layers_19_self_attn_linear_out_bias, weight = encoder_layers_19_self_attn_linear_out_weight_quantized, x = input_1023)[name = tensor("linear_178")]; tensor input_1027 = add(x = input_1019, y = input_1025)[name = tensor("input_1027")]; tensor x_437_axes_0 = const()[name = tensor("x_437_axes_0"), val = tensor([-1])]; tensor x_437 = layer_norm(axes = x_437_axes_0, beta = encoder_layers_19_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_19_norm_conv_weight, x = input_1027)[name = tensor("x_437")]; tensor input_1029_perm_0 = const()[name = tensor("input_1029_perm_0"), val = tensor([0, 2, 1])]; tensor input_1031_pad_type_0 = const()[name = tensor("input_1031_pad_type_0"), val = tensor("valid")]; tensor input_1031_strides_0 = const()[name = tensor("input_1031_strides_0"), val = tensor([1])]; tensor input_1031_pad_0 = const()[name = tensor("input_1031_pad_0"), val = tensor([0, 0])]; tensor input_1031_dilations_0 = const()[name = tensor("input_1031_dilations_0"), val = tensor([1])]; tensor input_1031_groups_0 = const()[name = tensor("input_1031_groups_0"), val = tensor(1)]; tensor input_1029 = transpose(perm = input_1029_perm_0, x = x_437)[name = tensor("transpose_173")]; tensor input_1031 = conv(bias = encoder_layers_19_conv_pointwise_conv1_bias, dilations = input_1031_dilations_0, groups = input_1031_groups_0, pad = input_1031_pad_0, pad_type = input_1031_pad_type_0, strides = input_1031_strides_0, weight = encoder_layers_19_conv_pointwise_conv1_weight_quantized, x = input_1029)[name = tensor("input_1031")]; tensor x_439_split_num_splits_0 = const()[name = tensor("x_439_split_num_splits_0"), val = tensor(2)]; tensor x_439_split_axis_0 = const()[name = tensor("x_439_split_axis_0"), val = tensor(1)]; tensor x_439_split_0, tensor x_439_split_1 = split(axis = x_439_split_axis_0, num_splits = x_439_split_num_splits_0, x = input_1031)[name = tensor("x_439_split")]; tensor x_439_split_1_sigmoid = sigmoid(x = x_439_split_1)[name = tensor("x_439_split_1_sigmoid")]; tensor x_439 = mul(x = x_439_split_0, y = x_439_split_1_sigmoid)[name = tensor("x_439")]; tensor input_1033 = select(a = var_13, b = x_439, cond = var_339)[name = tensor("input_1033")]; tensor const_207 = const()[name = tensor("const_207"), val = tensor(0x0p+0)]; tensor input_1035_pad_0 = const()[name = tensor("input_1035_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1035_mode_0 = const()[name = tensor("input_1035_mode_0"), val = tensor("constant")]; tensor input_1035 = pad(constant_val = const_207, mode = input_1035_mode_0, pad = input_1035_pad_0, x = input_1033)[name = tensor("input_1035")]; tensor input_1037_pad_type_0 = const()[name = tensor("input_1037_pad_type_0"), val = tensor("valid")]; tensor input_1037_groups_0 = const()[name = tensor("input_1037_groups_0"), val = tensor(1024)]; tensor input_1037_strides_0 = const()[name = tensor("input_1037_strides_0"), val = tensor([1])]; tensor input_1037_pad_0 = const()[name = tensor("input_1037_pad_0"), val = tensor([0, 0])]; tensor input_1037_dilations_0 = const()[name = tensor("input_1037_dilations_0"), val = tensor([1])]; tensor const_286_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_286_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595751936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595762304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595761216)))]; tensor const_287 = const()[name = tensor("const_287"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595766464)))]; tensor input_1039 = conv(bias = const_287, dilations = input_1037_dilations_0, groups = input_1037_groups_0, pad = input_1037_pad_0, pad_type = input_1037_pad_type_0, strides = input_1037_strides_0, weight = const_286_quantized, x = input_1035)[name = tensor("input_1039")]; tensor input_1041 = silu(x = input_1039)[name = tensor("input_1041")]; tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; tensor x_441 = conv(bias = encoder_layers_19_conv_pointwise_conv2_bias, dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = encoder_layers_19_conv_pointwise_conv2_weight_quantized, x = input_1041)[name = tensor("x_441")]; tensor input_1043_perm_0 = const()[name = tensor("input_1043_perm_0"), val = tensor([0, 2, 1])]; tensor input_1043 = transpose(perm = input_1043_perm_0, x = x_441)[name = tensor("transpose_172")]; tensor input_1045 = add(x = input_1027, y = input_1043)[name = tensor("input_1045")]; tensor input_1047_axes_0 = const()[name = tensor("input_1047_axes_0"), val = tensor([-1])]; tensor input_1047 = layer_norm(axes = input_1047_axes_0, beta = encoder_layers_19_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_19_norm_feed_forward2_weight, x = input_1045)[name = tensor("input_1047")]; tensor input_1049 = linear(bias = encoder_layers_19_feed_forward2_linear1_bias, weight = encoder_layers_19_feed_forward2_linear1_weight_quantized, x = input_1047)[name = tensor("linear_179")]; tensor input_1051 = silu(x = input_1049)[name = tensor("input_1051")]; tensor input_1055 = linear(bias = encoder_layers_19_feed_forward2_linear2_bias, weight = encoder_layers_19_feed_forward2_linear2_weight_quantized, x = input_1051)[name = tensor("linear_180")]; tensor var_3687 = const()[name = tensor("op_3687"), val = tensor(0x1p-1)]; tensor var_3688 = mul(x = input_1055, y = var_3687)[name = tensor("op_3688")]; tensor input_1057 = add(x = input_1045, y = var_3688)[name = tensor("input_1057")]; tensor input_1059_axes_0 = const()[name = tensor("input_1059_axes_0"), val = tensor([-1])]; tensor input_1059 = layer_norm(axes = input_1059_axes_0, beta = encoder_layers_19_norm_out_bias, epsilon = var_11, gamma = encoder_layers_19_norm_out_weight, x = input_1057)[name = tensor("input_1059")]; tensor input_1061_axes_0 = const()[name = tensor("input_1061_axes_0"), val = tensor([-1])]; tensor input_1061 = layer_norm(axes = input_1061_axes_0, beta = encoder_layers_20_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_20_norm_feed_forward1_weight, x = input_1059)[name = tensor("input_1061")]; tensor input_1063 = linear(bias = encoder_layers_20_feed_forward1_linear1_bias, weight = encoder_layers_20_feed_forward1_linear1_weight_quantized, x = input_1061)[name = tensor("linear_181")]; tensor input_1065 = silu(x = input_1063)[name = tensor("input_1065")]; tensor input_1069 = linear(bias = encoder_layers_20_feed_forward1_linear2_bias, weight = encoder_layers_20_feed_forward1_linear2_weight_quantized, x = input_1065)[name = tensor("linear_182")]; tensor var_3718 = const()[name = tensor("op_3718"), val = tensor(0x1p-1)]; tensor var_3719 = mul(x = input_1069, y = var_3718)[name = tensor("op_3719")]; tensor input_1071 = add(x = input_1059, y = var_3719)[name = tensor("input_1071")]; tensor query_41_axes_0 = const()[name = tensor("query_41_axes_0"), val = tensor([-1])]; tensor query_41 = layer_norm(axes = query_41_axes_0, beta = encoder_layers_20_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_20_norm_self_att_weight, x = input_1071)[name = tensor("query_41")]; tensor var_3735 = linear(bias = encoder_layers_20_self_attn_linear_q_bias, weight = encoder_layers_20_self_attn_linear_q_weight_quantized, x = query_41)[name = tensor("linear_183")]; tensor var_3736 = const()[name = tensor("op_3736"), val = tensor([1, -1, 8, 128])]; tensor q_121 = reshape(shape = var_3736, x = var_3735)[name = tensor("q_121")]; tensor var_3740 = linear(bias = encoder_layers_20_self_attn_linear_k_bias, weight = encoder_layers_20_self_attn_linear_k_weight_quantized, x = query_41)[name = tensor("linear_184")]; tensor var_3741 = const()[name = tensor("op_3741"), val = tensor([1, -1, 8, 128])]; tensor k_81 = reshape(shape = var_3741, x = var_3740)[name = tensor("k_81")]; tensor var_3745 = linear(bias = encoder_layers_20_self_attn_linear_v_bias, weight = encoder_layers_20_self_attn_linear_v_weight_quantized, x = query_41)[name = tensor("linear_185")]; tensor var_3746 = const()[name = tensor("op_3746"), val = tensor([1, -1, 8, 128])]; tensor v_41 = reshape(shape = var_3746, x = var_3745)[name = tensor("v_41")]; tensor value_43_perm_0 = const()[name = tensor("value_43_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_3758 = add(x = q_121, y = encoder_layers_20_self_attn_pos_bias_u)[name = tensor("op_3758")]; tensor var_3760 = add(x = q_121, y = encoder_layers_20_self_attn_pos_bias_v)[name = tensor("op_3760")]; tensor q_with_bias_v_41_perm_0 = const()[name = tensor("q_with_bias_v_41_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_3762_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3762_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(595770624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596155136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596154688)))]; tensor x_449_transpose_x_0 = const()[name = tensor("x_449_transpose_x_0"), val = tensor(false)]; tensor x_449_transpose_y_0 = const()[name = tensor("x_449_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_41 = transpose(perm = q_with_bias_v_41_perm_0, x = var_3760)[name = tensor("transpose_171")]; tensor x_449 = matmul(transpose_x = x_449_transpose_x_0, transpose_y = x_449_transpose_y_0, x = q_with_bias_v_41, y = op_3762_quantized)[name = tensor("x_449")]; tensor const_214 = const()[name = tensor("const_214"), val = tensor(0x0p+0)]; tensor x_451_pad_0 = const()[name = tensor("x_451_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_451_mode_0 = const()[name = tensor("x_451_mode_0"), val = tensor("constant")]; tensor x_451 = pad(constant_val = const_214, mode = x_451_mode_0, pad = x_451_pad_0, x = x_449)[name = tensor("x_451")]; tensor var_3770 = const()[name = tensor("op_3770"), val = tensor([1, 8, -1, 188])]; tensor x_453 = reshape(shape = var_3770, x = x_451)[name = tensor("x_453")]; tensor var_3774_begin_0 = const()[name = tensor("op_3774_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3774_end_0 = const()[name = tensor("op_3774_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3774_end_mask_0 = const()[name = tensor("op_3774_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3774 = slice_by_index(begin = var_3774_begin_0, end = var_3774_end_0, end_mask = var_3774_end_mask_0, x = x_453)[name = tensor("op_3774")]; tensor var_3775 = const()[name = tensor("op_3775"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_81 = reshape(shape = var_3775, x = var_3774)[name = tensor("matrix_bd_81")]; tensor matrix_ac_41_transpose_x_0 = const()[name = tensor("matrix_ac_41_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_41_transpose_y_0 = const()[name = tensor("matrix_ac_41_transpose_y_0"), val = tensor(false)]; tensor transpose_136_perm_0 = const()[name = tensor("transpose_136_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_137_perm_0 = const()[name = tensor("transpose_137_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_137 = transpose(perm = transpose_137_perm_0, x = k_81)[name = tensor("transpose_169")]; tensor transpose_136 = transpose(perm = transpose_136_perm_0, x = var_3758)[name = tensor("transpose_170")]; tensor matrix_ac_41 = matmul(transpose_x = matrix_ac_41_transpose_x_0, transpose_y = matrix_ac_41_transpose_y_0, x = transpose_136, y = transpose_137)[name = tensor("matrix_ac_41")]; tensor matrix_bd_83_begin_0 = const()[name = tensor("matrix_bd_83_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_83_end_0 = const()[name = tensor("matrix_bd_83_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_83_end_mask_0 = const()[name = tensor("matrix_bd_83_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_83 = slice_by_index(begin = matrix_bd_83_begin_0, end = matrix_bd_83_end_0, end_mask = matrix_bd_83_end_mask_0, x = matrix_bd_81)[name = tensor("matrix_bd_83")]; tensor var_3784 = add(x = matrix_ac_41, y = matrix_bd_83)[name = tensor("op_3784")]; tensor _inversed_scores_81_y_0 = const()[name = tensor("_inversed_scores_81_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_81 = mul(x = var_3784, y = _inversed_scores_81_y_0)[name = tensor("_inversed_scores_81")]; tensor scores_83 = select(a = var_14, b = _inversed_scores_81, cond = mask_3)[name = tensor("scores_83")]; tensor var_3790 = softmax(axis = var_32, x = scores_83)[name = tensor("op_3790")]; tensor input_1073 = select(a = var_13, b = var_3790, cond = mask_3)[name = tensor("input_1073")]; tensor x_455_transpose_x_0 = const()[name = tensor("x_455_transpose_x_0"), val = tensor(false)]; tensor x_455_transpose_y_0 = const()[name = tensor("x_455_transpose_y_0"), val = tensor(false)]; tensor value_43 = transpose(perm = value_43_perm_0, x = v_41)[name = tensor("transpose_168")]; tensor x_455 = matmul(transpose_x = x_455_transpose_x_0, transpose_y = x_455_transpose_y_0, x = input_1073, y = value_43)[name = tensor("x_455")]; tensor var_3794_perm_0 = const()[name = tensor("op_3794_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3795 = const()[name = tensor("op_3795"), val = tensor([1, -1, 1024])]; tensor var_3794 = transpose(perm = var_3794_perm_0, x = x_455)[name = tensor("transpose_167")]; tensor input_1075 = reshape(shape = var_3795, x = var_3794)[name = tensor("input_1075")]; tensor input_1077 = linear(bias = encoder_layers_20_self_attn_linear_out_bias, weight = encoder_layers_20_self_attn_linear_out_weight_quantized, x = input_1075)[name = tensor("linear_187")]; tensor input_1079 = add(x = input_1071, y = input_1077)[name = tensor("input_1079")]; tensor x_459_axes_0 = const()[name = tensor("x_459_axes_0"), val = tensor([-1])]; tensor x_459 = layer_norm(axes = x_459_axes_0, beta = encoder_layers_20_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_20_norm_conv_weight, x = input_1079)[name = tensor("x_459")]; tensor input_1081_perm_0 = const()[name = tensor("input_1081_perm_0"), val = tensor([0, 2, 1])]; tensor input_1083_pad_type_0 = const()[name = tensor("input_1083_pad_type_0"), val = tensor("valid")]; tensor input_1083_strides_0 = const()[name = tensor("input_1083_strides_0"), val = tensor([1])]; tensor input_1083_pad_0 = const()[name = tensor("input_1083_pad_0"), val = tensor([0, 0])]; tensor input_1083_dilations_0 = const()[name = tensor("input_1083_dilations_0"), val = tensor([1])]; tensor input_1083_groups_0 = const()[name = tensor("input_1083_groups_0"), val = tensor(1)]; tensor input_1081 = transpose(perm = input_1081_perm_0, x = x_459)[name = tensor("transpose_166")]; tensor input_1083 = conv(bias = encoder_layers_20_conv_pointwise_conv1_bias, dilations = input_1083_dilations_0, groups = input_1083_groups_0, pad = input_1083_pad_0, pad_type = input_1083_pad_type_0, strides = input_1083_strides_0, weight = encoder_layers_20_conv_pointwise_conv1_weight_quantized, x = input_1081)[name = tensor("input_1083")]; tensor x_461_split_num_splits_0 = const()[name = tensor("x_461_split_num_splits_0"), val = tensor(2)]; tensor x_461_split_axis_0 = const()[name = tensor("x_461_split_axis_0"), val = tensor(1)]; tensor x_461_split_0, tensor x_461_split_1 = split(axis = x_461_split_axis_0, num_splits = x_461_split_num_splits_0, x = input_1083)[name = tensor("x_461_split")]; tensor x_461_split_1_sigmoid = sigmoid(x = x_461_split_1)[name = tensor("x_461_split_1_sigmoid")]; tensor x_461 = mul(x = x_461_split_0, y = x_461_split_1_sigmoid)[name = tensor("x_461")]; tensor input_1085 = select(a = var_13, b = x_461, cond = var_339)[name = tensor("input_1085")]; tensor const_217 = const()[name = tensor("const_217"), val = tensor(0x0p+0)]; tensor input_1087_pad_0 = const()[name = tensor("input_1087_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1087_mode_0 = const()[name = tensor("input_1087_mode_0"), val = tensor("constant")]; tensor input_1087 = pad(constant_val = const_217, mode = input_1087_mode_0, pad = input_1087_pad_0, x = input_1085)[name = tensor("input_1087")]; tensor input_1089_pad_type_0 = const()[name = tensor("input_1089_pad_type_0"), val = tensor("valid")]; tensor input_1089_groups_0 = const()[name = tensor("input_1089_groups_0"), val = tensor(1024)]; tensor input_1089_strides_0 = const()[name = tensor("input_1089_strides_0"), val = tensor([1])]; tensor input_1089_pad_0 = const()[name = tensor("input_1089_pad_0"), val = tensor([0, 0])]; tensor input_1089_dilations_0 = const()[name = tensor("input_1089_dilations_0"), val = tensor([1])]; tensor const_288_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_288_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596156736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596167104))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596166016)))]; tensor const_289 = const()[name = tensor("const_289"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596171264)))]; tensor input_1091 = conv(bias = const_289, dilations = input_1089_dilations_0, groups = input_1089_groups_0, pad = input_1089_pad_0, pad_type = input_1089_pad_type_0, strides = input_1089_strides_0, weight = const_288_quantized, x = input_1087)[name = tensor("input_1091")]; tensor input_1093 = silu(x = input_1091)[name = tensor("input_1093")]; tensor x_463_pad_type_0 = const()[name = tensor("x_463_pad_type_0"), val = tensor("valid")]; tensor x_463_strides_0 = const()[name = tensor("x_463_strides_0"), val = tensor([1])]; tensor x_463_pad_0 = const()[name = tensor("x_463_pad_0"), val = tensor([0, 0])]; tensor x_463_dilations_0 = const()[name = tensor("x_463_dilations_0"), val = tensor([1])]; tensor x_463_groups_0 = const()[name = tensor("x_463_groups_0"), val = tensor(1)]; tensor x_463 = conv(bias = encoder_layers_20_conv_pointwise_conv2_bias, dilations = x_463_dilations_0, groups = x_463_groups_0, pad = x_463_pad_0, pad_type = x_463_pad_type_0, strides = x_463_strides_0, weight = encoder_layers_20_conv_pointwise_conv2_weight_quantized, x = input_1093)[name = tensor("x_463")]; tensor input_1095_perm_0 = const()[name = tensor("input_1095_perm_0"), val = tensor([0, 2, 1])]; tensor input_1095 = transpose(perm = input_1095_perm_0, x = x_463)[name = tensor("transpose_165")]; tensor input_1097 = add(x = input_1079, y = input_1095)[name = tensor("input_1097")]; tensor input_1099_axes_0 = const()[name = tensor("input_1099_axes_0"), val = tensor([-1])]; tensor input_1099 = layer_norm(axes = input_1099_axes_0, beta = encoder_layers_20_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_20_norm_feed_forward2_weight, x = input_1097)[name = tensor("input_1099")]; tensor input_1101 = linear(bias = encoder_layers_20_feed_forward2_linear1_bias, weight = encoder_layers_20_feed_forward2_linear1_weight_quantized, x = input_1099)[name = tensor("linear_188")]; tensor input_1103 = silu(x = input_1101)[name = tensor("input_1103")]; tensor input_1107 = linear(bias = encoder_layers_20_feed_forward2_linear2_bias, weight = encoder_layers_20_feed_forward2_linear2_weight_quantized, x = input_1103)[name = tensor("linear_189")]; tensor var_3861 = const()[name = tensor("op_3861"), val = tensor(0x1p-1)]; tensor var_3862 = mul(x = input_1107, y = var_3861)[name = tensor("op_3862")]; tensor input_1109 = add(x = input_1097, y = var_3862)[name = tensor("input_1109")]; tensor input_1111_axes_0 = const()[name = tensor("input_1111_axes_0"), val = tensor([-1])]; tensor input_1111 = layer_norm(axes = input_1111_axes_0, beta = encoder_layers_20_norm_out_bias, epsilon = var_11, gamma = encoder_layers_20_norm_out_weight, x = input_1109)[name = tensor("input_1111")]; tensor input_1113_axes_0 = const()[name = tensor("input_1113_axes_0"), val = tensor([-1])]; tensor input_1113 = layer_norm(axes = input_1113_axes_0, beta = encoder_layers_21_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_21_norm_feed_forward1_weight, x = input_1111)[name = tensor("input_1113")]; tensor input_1115 = linear(bias = encoder_layers_21_feed_forward1_linear1_bias, weight = encoder_layers_21_feed_forward1_linear1_weight_quantized, x = input_1113)[name = tensor("linear_190")]; tensor input_1117 = silu(x = input_1115)[name = tensor("input_1117")]; tensor input_1121 = linear(bias = encoder_layers_21_feed_forward1_linear2_bias, weight = encoder_layers_21_feed_forward1_linear2_weight_quantized, x = input_1117)[name = tensor("linear_191")]; tensor var_3892 = const()[name = tensor("op_3892"), val = tensor(0x1p-1)]; tensor var_3893 = mul(x = input_1121, y = var_3892)[name = tensor("op_3893")]; tensor input_1123 = add(x = input_1111, y = var_3893)[name = tensor("input_1123")]; tensor query_43_axes_0 = const()[name = tensor("query_43_axes_0"), val = tensor([-1])]; tensor query_43 = layer_norm(axes = query_43_axes_0, beta = encoder_layers_21_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_21_norm_self_att_weight, x = input_1123)[name = tensor("query_43")]; tensor var_3909 = linear(bias = encoder_layers_21_self_attn_linear_q_bias, weight = encoder_layers_21_self_attn_linear_q_weight_quantized, x = query_43)[name = tensor("linear_192")]; tensor var_3910 = const()[name = tensor("op_3910"), val = tensor([1, -1, 8, 128])]; tensor q_127 = reshape(shape = var_3910, x = var_3909)[name = tensor("q_127")]; tensor var_3914 = linear(bias = encoder_layers_21_self_attn_linear_k_bias, weight = encoder_layers_21_self_attn_linear_k_weight_quantized, x = query_43)[name = tensor("linear_193")]; tensor var_3915 = const()[name = tensor("op_3915"), val = tensor([1, -1, 8, 128])]; tensor k_85 = reshape(shape = var_3915, x = var_3914)[name = tensor("k_85")]; tensor var_3919 = linear(bias = encoder_layers_21_self_attn_linear_v_bias, weight = encoder_layers_21_self_attn_linear_v_weight_quantized, x = query_43)[name = tensor("linear_194")]; tensor var_3920 = const()[name = tensor("op_3920"), val = tensor([1, -1, 8, 128])]; tensor v_43 = reshape(shape = var_3920, x = var_3919)[name = tensor("v_43")]; tensor value_45_perm_0 = const()[name = tensor("value_45_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_3932 = add(x = q_127, y = encoder_layers_21_self_attn_pos_bias_u)[name = tensor("op_3932")]; tensor var_3934 = add(x = q_127, y = encoder_layers_21_self_attn_pos_bias_v)[name = tensor("op_3934")]; tensor q_with_bias_v_43_perm_0 = const()[name = tensor("q_with_bias_v_43_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_3936_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3936_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596175424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596559936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596559488)))]; tensor x_471_transpose_x_0 = const()[name = tensor("x_471_transpose_x_0"), val = tensor(false)]; tensor x_471_transpose_y_0 = const()[name = tensor("x_471_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_43 = transpose(perm = q_with_bias_v_43_perm_0, x = var_3934)[name = tensor("transpose_164")]; tensor x_471 = matmul(transpose_x = x_471_transpose_x_0, transpose_y = x_471_transpose_y_0, x = q_with_bias_v_43, y = op_3936_quantized)[name = tensor("x_471")]; tensor const_224 = const()[name = tensor("const_224"), val = tensor(0x0p+0)]; tensor x_473_pad_0 = const()[name = tensor("x_473_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_473_mode_0 = const()[name = tensor("x_473_mode_0"), val = tensor("constant")]; tensor x_473 = pad(constant_val = const_224, mode = x_473_mode_0, pad = x_473_pad_0, x = x_471)[name = tensor("x_473")]; tensor var_3944 = const()[name = tensor("op_3944"), val = tensor([1, 8, -1, 188])]; tensor x_475 = reshape(shape = var_3944, x = x_473)[name = tensor("x_475")]; tensor var_3948_begin_0 = const()[name = tensor("op_3948_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3948_end_0 = const()[name = tensor("op_3948_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3948_end_mask_0 = const()[name = tensor("op_3948_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3948 = slice_by_index(begin = var_3948_begin_0, end = var_3948_end_0, end_mask = var_3948_end_mask_0, x = x_475)[name = tensor("op_3948")]; tensor var_3949 = const()[name = tensor("op_3949"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_85 = reshape(shape = var_3949, x = var_3948)[name = tensor("matrix_bd_85")]; tensor matrix_ac_43_transpose_x_0 = const()[name = tensor("matrix_ac_43_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_43_transpose_y_0 = const()[name = tensor("matrix_ac_43_transpose_y_0"), val = tensor(false)]; tensor transpose_138_perm_0 = const()[name = tensor("transpose_138_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_139_perm_0 = const()[name = tensor("transpose_139_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_139 = transpose(perm = transpose_139_perm_0, x = k_85)[name = tensor("transpose_162")]; tensor transpose_138 = transpose(perm = transpose_138_perm_0, x = var_3932)[name = tensor("transpose_163")]; tensor matrix_ac_43 = matmul(transpose_x = matrix_ac_43_transpose_x_0, transpose_y = matrix_ac_43_transpose_y_0, x = transpose_138, y = transpose_139)[name = tensor("matrix_ac_43")]; tensor matrix_bd_87_begin_0 = const()[name = tensor("matrix_bd_87_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_87_end_0 = const()[name = tensor("matrix_bd_87_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_87_end_mask_0 = const()[name = tensor("matrix_bd_87_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_87 = slice_by_index(begin = matrix_bd_87_begin_0, end = matrix_bd_87_end_0, end_mask = matrix_bd_87_end_mask_0, x = matrix_bd_85)[name = tensor("matrix_bd_87")]; tensor var_3958 = add(x = matrix_ac_43, y = matrix_bd_87)[name = tensor("op_3958")]; tensor _inversed_scores_85_y_0 = const()[name = tensor("_inversed_scores_85_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_85 = mul(x = var_3958, y = _inversed_scores_85_y_0)[name = tensor("_inversed_scores_85")]; tensor scores_87 = select(a = var_14, b = _inversed_scores_85, cond = mask_3)[name = tensor("scores_87")]; tensor var_3964 = softmax(axis = var_32, x = scores_87)[name = tensor("op_3964")]; tensor input_1125 = select(a = var_13, b = var_3964, cond = mask_3)[name = tensor("input_1125")]; tensor x_477_transpose_x_0 = const()[name = tensor("x_477_transpose_x_0"), val = tensor(false)]; tensor x_477_transpose_y_0 = const()[name = tensor("x_477_transpose_y_0"), val = tensor(false)]; tensor value_45 = transpose(perm = value_45_perm_0, x = v_43)[name = tensor("transpose_161")]; tensor x_477 = matmul(transpose_x = x_477_transpose_x_0, transpose_y = x_477_transpose_y_0, x = input_1125, y = value_45)[name = tensor("x_477")]; tensor var_3968_perm_0 = const()[name = tensor("op_3968_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3969 = const()[name = tensor("op_3969"), val = tensor([1, -1, 1024])]; tensor var_3968 = transpose(perm = var_3968_perm_0, x = x_477)[name = tensor("transpose_160")]; tensor input_1127 = reshape(shape = var_3969, x = var_3968)[name = tensor("input_1127")]; tensor input_1129 = linear(bias = encoder_layers_21_self_attn_linear_out_bias, weight = encoder_layers_21_self_attn_linear_out_weight_quantized, x = input_1127)[name = tensor("linear_196")]; tensor input_1131 = add(x = input_1123, y = input_1129)[name = tensor("input_1131")]; tensor x_481_axes_0 = const()[name = tensor("x_481_axes_0"), val = tensor([-1])]; tensor x_481 = layer_norm(axes = x_481_axes_0, beta = encoder_layers_21_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_21_norm_conv_weight, x = input_1131)[name = tensor("x_481")]; tensor input_1133_perm_0 = const()[name = tensor("input_1133_perm_0"), val = tensor([0, 2, 1])]; tensor input_1135_pad_type_0 = const()[name = tensor("input_1135_pad_type_0"), val = tensor("valid")]; tensor input_1135_strides_0 = const()[name = tensor("input_1135_strides_0"), val = tensor([1])]; tensor input_1135_pad_0 = const()[name = tensor("input_1135_pad_0"), val = tensor([0, 0])]; tensor input_1135_dilations_0 = const()[name = tensor("input_1135_dilations_0"), val = tensor([1])]; tensor input_1135_groups_0 = const()[name = tensor("input_1135_groups_0"), val = tensor(1)]; tensor input_1133 = transpose(perm = input_1133_perm_0, x = x_481)[name = tensor("transpose_159")]; tensor input_1135 = conv(bias = encoder_layers_21_conv_pointwise_conv1_bias, dilations = input_1135_dilations_0, groups = input_1135_groups_0, pad = input_1135_pad_0, pad_type = input_1135_pad_type_0, strides = input_1135_strides_0, weight = encoder_layers_21_conv_pointwise_conv1_weight_quantized, x = input_1133)[name = tensor("input_1135")]; tensor x_483_split_num_splits_0 = const()[name = tensor("x_483_split_num_splits_0"), val = tensor(2)]; tensor x_483_split_axis_0 = const()[name = tensor("x_483_split_axis_0"), val = tensor(1)]; tensor x_483_split_0, tensor x_483_split_1 = split(axis = x_483_split_axis_0, num_splits = x_483_split_num_splits_0, x = input_1135)[name = tensor("x_483_split")]; tensor x_483_split_1_sigmoid = sigmoid(x = x_483_split_1)[name = tensor("x_483_split_1_sigmoid")]; tensor x_483 = mul(x = x_483_split_0, y = x_483_split_1_sigmoid)[name = tensor("x_483")]; tensor input_1137 = select(a = var_13, b = x_483, cond = var_339)[name = tensor("input_1137")]; tensor const_227 = const()[name = tensor("const_227"), val = tensor(0x0p+0)]; tensor input_1139_pad_0 = const()[name = tensor("input_1139_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1139_mode_0 = const()[name = tensor("input_1139_mode_0"), val = tensor("constant")]; tensor input_1139 = pad(constant_val = const_227, mode = input_1139_mode_0, pad = input_1139_pad_0, x = input_1137)[name = tensor("input_1139")]; tensor input_1141_pad_type_0 = const()[name = tensor("input_1141_pad_type_0"), val = tensor("valid")]; tensor input_1141_groups_0 = const()[name = tensor("input_1141_groups_0"), val = tensor(1024)]; tensor input_1141_strides_0 = const()[name = tensor("input_1141_strides_0"), val = tensor([1])]; tensor input_1141_pad_0 = const()[name = tensor("input_1141_pad_0"), val = tensor([0, 0])]; tensor input_1141_dilations_0 = const()[name = tensor("input_1141_dilations_0"), val = tensor([1])]; tensor const_290_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_290_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596561536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596571904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596570816)))]; tensor const_291 = const()[name = tensor("const_291"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596576064)))]; tensor input_1143 = conv(bias = const_291, dilations = input_1141_dilations_0, groups = input_1141_groups_0, pad = input_1141_pad_0, pad_type = input_1141_pad_type_0, strides = input_1141_strides_0, weight = const_290_quantized, x = input_1139)[name = tensor("input_1143")]; tensor input_1145 = silu(x = input_1143)[name = tensor("input_1145")]; tensor x_485_pad_type_0 = const()[name = tensor("x_485_pad_type_0"), val = tensor("valid")]; tensor x_485_strides_0 = const()[name = tensor("x_485_strides_0"), val = tensor([1])]; tensor x_485_pad_0 = const()[name = tensor("x_485_pad_0"), val = tensor([0, 0])]; tensor x_485_dilations_0 = const()[name = tensor("x_485_dilations_0"), val = tensor([1])]; tensor x_485_groups_0 = const()[name = tensor("x_485_groups_0"), val = tensor(1)]; tensor x_485 = conv(bias = encoder_layers_21_conv_pointwise_conv2_bias, dilations = x_485_dilations_0, groups = x_485_groups_0, pad = x_485_pad_0, pad_type = x_485_pad_type_0, strides = x_485_strides_0, weight = encoder_layers_21_conv_pointwise_conv2_weight_quantized, x = input_1145)[name = tensor("x_485")]; tensor input_1147_perm_0 = const()[name = tensor("input_1147_perm_0"), val = tensor([0, 2, 1])]; tensor input_1147 = transpose(perm = input_1147_perm_0, x = x_485)[name = tensor("transpose_158")]; tensor input_1149 = add(x = input_1131, y = input_1147)[name = tensor("input_1149")]; tensor input_1151_axes_0 = const()[name = tensor("input_1151_axes_0"), val = tensor([-1])]; tensor input_1151 = layer_norm(axes = input_1151_axes_0, beta = encoder_layers_21_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_21_norm_feed_forward2_weight, x = input_1149)[name = tensor("input_1151")]; tensor input_1153 = linear(bias = encoder_layers_21_feed_forward2_linear1_bias, weight = encoder_layers_21_feed_forward2_linear1_weight_quantized, x = input_1151)[name = tensor("linear_197")]; tensor input_1155 = silu(x = input_1153)[name = tensor("input_1155")]; tensor input_1159 = linear(bias = encoder_layers_21_feed_forward2_linear2_bias, weight = encoder_layers_21_feed_forward2_linear2_weight_quantized, x = input_1155)[name = tensor("linear_198")]; tensor var_4035 = const()[name = tensor("op_4035"), val = tensor(0x1p-1)]; tensor var_4036 = mul(x = input_1159, y = var_4035)[name = tensor("op_4036")]; tensor input_1161 = add(x = input_1149, y = var_4036)[name = tensor("input_1161")]; tensor input_1163_axes_0 = const()[name = tensor("input_1163_axes_0"), val = tensor([-1])]; tensor input_1163 = layer_norm(axes = input_1163_axes_0, beta = encoder_layers_21_norm_out_bias, epsilon = var_11, gamma = encoder_layers_21_norm_out_weight, x = input_1161)[name = tensor("input_1163")]; tensor input_1165_axes_0 = const()[name = tensor("input_1165_axes_0"), val = tensor([-1])]; tensor input_1165 = layer_norm(axes = input_1165_axes_0, beta = encoder_layers_22_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_22_norm_feed_forward1_weight, x = input_1163)[name = tensor("input_1165")]; tensor input_1167 = linear(bias = encoder_layers_22_feed_forward1_linear1_bias, weight = encoder_layers_22_feed_forward1_linear1_weight_quantized, x = input_1165)[name = tensor("linear_199")]; tensor input_1169 = silu(x = input_1167)[name = tensor("input_1169")]; tensor input_1173 = linear(bias = encoder_layers_22_feed_forward1_linear2_bias, weight = encoder_layers_22_feed_forward1_linear2_weight_quantized, x = input_1169)[name = tensor("linear_200")]; tensor var_4066 = const()[name = tensor("op_4066"), val = tensor(0x1p-1)]; tensor var_4067 = mul(x = input_1173, y = var_4066)[name = tensor("op_4067")]; tensor input_1175 = add(x = input_1163, y = var_4067)[name = tensor("input_1175")]; tensor query_45_axes_0 = const()[name = tensor("query_45_axes_0"), val = tensor([-1])]; tensor query_45 = layer_norm(axes = query_45_axes_0, beta = encoder_layers_22_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_22_norm_self_att_weight, x = input_1175)[name = tensor("query_45")]; tensor var_4083 = linear(bias = encoder_layers_22_self_attn_linear_q_bias, weight = encoder_layers_22_self_attn_linear_q_weight_quantized, x = query_45)[name = tensor("linear_201")]; tensor var_4084 = const()[name = tensor("op_4084"), val = tensor([1, -1, 8, 128])]; tensor q_133 = reshape(shape = var_4084, x = var_4083)[name = tensor("q_133")]; tensor var_4088 = linear(bias = encoder_layers_22_self_attn_linear_k_bias, weight = encoder_layers_22_self_attn_linear_k_weight_quantized, x = query_45)[name = tensor("linear_202")]; tensor var_4089 = const()[name = tensor("op_4089"), val = tensor([1, -1, 8, 128])]; tensor k_89 = reshape(shape = var_4089, x = var_4088)[name = tensor("k_89")]; tensor var_4093 = linear(bias = encoder_layers_22_self_attn_linear_v_bias, weight = encoder_layers_22_self_attn_linear_v_weight_quantized, x = query_45)[name = tensor("linear_203")]; tensor var_4094 = const()[name = tensor("op_4094"), val = tensor([1, -1, 8, 128])]; tensor v_45 = reshape(shape = var_4094, x = var_4093)[name = tensor("v_45")]; tensor value_47_perm_0 = const()[name = tensor("value_47_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_4106 = add(x = q_133, y = encoder_layers_22_self_attn_pos_bias_u)[name = tensor("op_4106")]; tensor var_4108 = add(x = q_133, y = encoder_layers_22_self_attn_pos_bias_v)[name = tensor("op_4108")]; tensor q_with_bias_v_45_perm_0 = const()[name = tensor("q_with_bias_v_45_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_4110_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4110_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596580224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596964736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596964288)))]; tensor x_493_transpose_x_0 = const()[name = tensor("x_493_transpose_x_0"), val = tensor(false)]; tensor x_493_transpose_y_0 = const()[name = tensor("x_493_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_45 = transpose(perm = q_with_bias_v_45_perm_0, x = var_4108)[name = tensor("transpose_157")]; tensor x_493 = matmul(transpose_x = x_493_transpose_x_0, transpose_y = x_493_transpose_y_0, x = q_with_bias_v_45, y = op_4110_quantized)[name = tensor("x_493")]; tensor const_234 = const()[name = tensor("const_234"), val = tensor(0x0p+0)]; tensor x_495_pad_0 = const()[name = tensor("x_495_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_495_mode_0 = const()[name = tensor("x_495_mode_0"), val = tensor("constant")]; tensor x_495 = pad(constant_val = const_234, mode = x_495_mode_0, pad = x_495_pad_0, x = x_493)[name = tensor("x_495")]; tensor var_4118 = const()[name = tensor("op_4118"), val = tensor([1, 8, -1, 188])]; tensor x_497 = reshape(shape = var_4118, x = x_495)[name = tensor("x_497")]; tensor var_4122_begin_0 = const()[name = tensor("op_4122_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_4122_end_0 = const()[name = tensor("op_4122_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_4122_end_mask_0 = const()[name = tensor("op_4122_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_4122 = slice_by_index(begin = var_4122_begin_0, end = var_4122_end_0, end_mask = var_4122_end_mask_0, x = x_497)[name = tensor("op_4122")]; tensor var_4123 = const()[name = tensor("op_4123"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_89 = reshape(shape = var_4123, x = var_4122)[name = tensor("matrix_bd_89")]; tensor matrix_ac_45_transpose_x_0 = const()[name = tensor("matrix_ac_45_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_45_transpose_y_0 = const()[name = tensor("matrix_ac_45_transpose_y_0"), val = tensor(false)]; tensor transpose_140_perm_0 = const()[name = tensor("transpose_140_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_141_perm_0 = const()[name = tensor("transpose_141_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_141 = transpose(perm = transpose_141_perm_0, x = k_89)[name = tensor("transpose_155")]; tensor transpose_140 = transpose(perm = transpose_140_perm_0, x = var_4106)[name = tensor("transpose_156")]; tensor matrix_ac_45 = matmul(transpose_x = matrix_ac_45_transpose_x_0, transpose_y = matrix_ac_45_transpose_y_0, x = transpose_140, y = transpose_141)[name = tensor("matrix_ac_45")]; tensor matrix_bd_91_begin_0 = const()[name = tensor("matrix_bd_91_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_91_end_0 = const()[name = tensor("matrix_bd_91_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_91_end_mask_0 = const()[name = tensor("matrix_bd_91_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_91 = slice_by_index(begin = matrix_bd_91_begin_0, end = matrix_bd_91_end_0, end_mask = matrix_bd_91_end_mask_0, x = matrix_bd_89)[name = tensor("matrix_bd_91")]; tensor var_4132 = add(x = matrix_ac_45, y = matrix_bd_91)[name = tensor("op_4132")]; tensor _inversed_scores_89_y_0 = const()[name = tensor("_inversed_scores_89_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_89 = mul(x = var_4132, y = _inversed_scores_89_y_0)[name = tensor("_inversed_scores_89")]; tensor scores_91 = select(a = var_14, b = _inversed_scores_89, cond = mask_3)[name = tensor("scores_91")]; tensor var_4138 = softmax(axis = var_32, x = scores_91)[name = tensor("op_4138")]; tensor input_1177 = select(a = var_13, b = var_4138, cond = mask_3)[name = tensor("input_1177")]; tensor x_499_transpose_x_0 = const()[name = tensor("x_499_transpose_x_0"), val = tensor(false)]; tensor x_499_transpose_y_0 = const()[name = tensor("x_499_transpose_y_0"), val = tensor(false)]; tensor value_47 = transpose(perm = value_47_perm_0, x = v_45)[name = tensor("transpose_154")]; tensor x_499 = matmul(transpose_x = x_499_transpose_x_0, transpose_y = x_499_transpose_y_0, x = input_1177, y = value_47)[name = tensor("x_499")]; tensor var_4142_perm_0 = const()[name = tensor("op_4142_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4143 = const()[name = tensor("op_4143"), val = tensor([1, -1, 1024])]; tensor var_4142 = transpose(perm = var_4142_perm_0, x = x_499)[name = tensor("transpose_153")]; tensor input_1179 = reshape(shape = var_4143, x = var_4142)[name = tensor("input_1179")]; tensor input_1181 = linear(bias = encoder_layers_22_self_attn_linear_out_bias, weight = encoder_layers_22_self_attn_linear_out_weight_quantized, x = input_1179)[name = tensor("linear_205")]; tensor input_1183 = add(x = input_1175, y = input_1181)[name = tensor("input_1183")]; tensor x_503_axes_0 = const()[name = tensor("x_503_axes_0"), val = tensor([-1])]; tensor x_503 = layer_norm(axes = x_503_axes_0, beta = encoder_layers_22_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_22_norm_conv_weight, x = input_1183)[name = tensor("x_503")]; tensor input_1185_perm_0 = const()[name = tensor("input_1185_perm_0"), val = tensor([0, 2, 1])]; tensor input_1187_pad_type_0 = const()[name = tensor("input_1187_pad_type_0"), val = tensor("valid")]; tensor input_1187_strides_0 = const()[name = tensor("input_1187_strides_0"), val = tensor([1])]; tensor input_1187_pad_0 = const()[name = tensor("input_1187_pad_0"), val = tensor([0, 0])]; tensor input_1187_dilations_0 = const()[name = tensor("input_1187_dilations_0"), val = tensor([1])]; tensor input_1187_groups_0 = const()[name = tensor("input_1187_groups_0"), val = tensor(1)]; tensor input_1185 = transpose(perm = input_1185_perm_0, x = x_503)[name = tensor("transpose_152")]; tensor input_1187 = conv(bias = encoder_layers_22_conv_pointwise_conv1_bias, dilations = input_1187_dilations_0, groups = input_1187_groups_0, pad = input_1187_pad_0, pad_type = input_1187_pad_type_0, strides = input_1187_strides_0, weight = encoder_layers_22_conv_pointwise_conv1_weight_quantized, x = input_1185)[name = tensor("input_1187")]; tensor x_505_split_num_splits_0 = const()[name = tensor("x_505_split_num_splits_0"), val = tensor(2)]; tensor x_505_split_axis_0 = const()[name = tensor("x_505_split_axis_0"), val = tensor(1)]; tensor x_505_split_0, tensor x_505_split_1 = split(axis = x_505_split_axis_0, num_splits = x_505_split_num_splits_0, x = input_1187)[name = tensor("x_505_split")]; tensor x_505_split_1_sigmoid = sigmoid(x = x_505_split_1)[name = tensor("x_505_split_1_sigmoid")]; tensor x_505 = mul(x = x_505_split_0, y = x_505_split_1_sigmoid)[name = tensor("x_505")]; tensor input_1189 = select(a = var_13, b = x_505, cond = var_339)[name = tensor("input_1189")]; tensor const_237 = const()[name = tensor("const_237"), val = tensor(0x0p+0)]; tensor input_1191_pad_0 = const()[name = tensor("input_1191_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1191_mode_0 = const()[name = tensor("input_1191_mode_0"), val = tensor("constant")]; tensor input_1191 = pad(constant_val = const_237, mode = input_1191_mode_0, pad = input_1191_pad_0, x = input_1189)[name = tensor("input_1191")]; tensor input_1193_pad_type_0 = const()[name = tensor("input_1193_pad_type_0"), val = tensor("valid")]; tensor input_1193_groups_0 = const()[name = tensor("input_1193_groups_0"), val = tensor(1024)]; tensor input_1193_strides_0 = const()[name = tensor("input_1193_strides_0"), val = tensor([1])]; tensor input_1193_pad_0 = const()[name = tensor("input_1193_pad_0"), val = tensor([0, 0])]; tensor input_1193_dilations_0 = const()[name = tensor("input_1193_dilations_0"), val = tensor([1])]; tensor const_292_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_292_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596966336))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596976704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596975616)))]; tensor const_293 = const()[name = tensor("const_293"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596980864)))]; tensor input_1195 = conv(bias = const_293, dilations = input_1193_dilations_0, groups = input_1193_groups_0, pad = input_1193_pad_0, pad_type = input_1193_pad_type_0, strides = input_1193_strides_0, weight = const_292_quantized, x = input_1191)[name = tensor("input_1195")]; tensor input_1197 = silu(x = input_1195)[name = tensor("input_1197")]; tensor x_507_pad_type_0 = const()[name = tensor("x_507_pad_type_0"), val = tensor("valid")]; tensor x_507_strides_0 = const()[name = tensor("x_507_strides_0"), val = tensor([1])]; tensor x_507_pad_0 = const()[name = tensor("x_507_pad_0"), val = tensor([0, 0])]; tensor x_507_dilations_0 = const()[name = tensor("x_507_dilations_0"), val = tensor([1])]; tensor x_507_groups_0 = const()[name = tensor("x_507_groups_0"), val = tensor(1)]; tensor x_507 = conv(bias = encoder_layers_22_conv_pointwise_conv2_bias, dilations = x_507_dilations_0, groups = x_507_groups_0, pad = x_507_pad_0, pad_type = x_507_pad_type_0, strides = x_507_strides_0, weight = encoder_layers_22_conv_pointwise_conv2_weight_quantized, x = input_1197)[name = tensor("x_507")]; tensor input_1199_perm_0 = const()[name = tensor("input_1199_perm_0"), val = tensor([0, 2, 1])]; tensor input_1199 = transpose(perm = input_1199_perm_0, x = x_507)[name = tensor("transpose_151")]; tensor input_1201 = add(x = input_1183, y = input_1199)[name = tensor("input_1201")]; tensor input_1203_axes_0 = const()[name = tensor("input_1203_axes_0"), val = tensor([-1])]; tensor input_1203 = layer_norm(axes = input_1203_axes_0, beta = encoder_layers_22_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_22_norm_feed_forward2_weight, x = input_1201)[name = tensor("input_1203")]; tensor input_1205 = linear(bias = encoder_layers_22_feed_forward2_linear1_bias, weight = encoder_layers_22_feed_forward2_linear1_weight_quantized, x = input_1203)[name = tensor("linear_206")]; tensor input_1207 = silu(x = input_1205)[name = tensor("input_1207")]; tensor input_1211 = linear(bias = encoder_layers_22_feed_forward2_linear2_bias, weight = encoder_layers_22_feed_forward2_linear2_weight_quantized, x = input_1207)[name = tensor("linear_207")]; tensor var_4209 = const()[name = tensor("op_4209"), val = tensor(0x1p-1)]; tensor var_4210 = mul(x = input_1211, y = var_4209)[name = tensor("op_4210")]; tensor input_1213 = add(x = input_1201, y = var_4210)[name = tensor("input_1213")]; tensor input_1215_axes_0 = const()[name = tensor("input_1215_axes_0"), val = tensor([-1])]; tensor input_1215 = layer_norm(axes = input_1215_axes_0, beta = encoder_layers_22_norm_out_bias, epsilon = var_11, gamma = encoder_layers_22_norm_out_weight, x = input_1213)[name = tensor("input_1215")]; tensor input_1217_axes_0 = const()[name = tensor("input_1217_axes_0"), val = tensor([-1])]; tensor input_1217 = layer_norm(axes = input_1217_axes_0, beta = encoder_layers_23_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_23_norm_feed_forward1_weight, x = input_1215)[name = tensor("input_1217")]; tensor input_1219 = linear(bias = encoder_layers_23_feed_forward1_linear1_bias, weight = encoder_layers_23_feed_forward1_linear1_weight_quantized, x = input_1217)[name = tensor("linear_208")]; tensor input_1221 = silu(x = input_1219)[name = tensor("input_1221")]; tensor input_1225 = linear(bias = encoder_layers_23_feed_forward1_linear2_bias, weight = encoder_layers_23_feed_forward1_linear2_weight_quantized, x = input_1221)[name = tensor("linear_209")]; tensor var_4240 = const()[name = tensor("op_4240"), val = tensor(0x1p-1)]; tensor var_4241 = mul(x = input_1225, y = var_4240)[name = tensor("op_4241")]; tensor input_1227 = add(x = input_1215, y = var_4241)[name = tensor("input_1227")]; tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; tensor query = layer_norm(axes = query_axes_0, beta = encoder_layers_23_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_23_norm_self_att_weight, x = input_1227)[name = tensor("query")]; tensor var_4257 = linear(bias = encoder_layers_23_self_attn_linear_q_bias, weight = encoder_layers_23_self_attn_linear_q_weight_quantized, x = query)[name = tensor("linear_210")]; tensor var_4258 = const()[name = tensor("op_4258"), val = tensor([1, -1, 8, 128])]; tensor q_139 = reshape(shape = var_4258, x = var_4257)[name = tensor("q_139")]; tensor var_4262 = linear(bias = encoder_layers_23_self_attn_linear_k_bias, weight = encoder_layers_23_self_attn_linear_k_weight_quantized, x = query)[name = tensor("linear_211")]; tensor var_4263 = const()[name = tensor("op_4263"), val = tensor([1, -1, 8, 128])]; tensor k_93 = reshape(shape = var_4263, x = var_4262)[name = tensor("k_93")]; tensor var_4267 = linear(bias = encoder_layers_23_self_attn_linear_v_bias, weight = encoder_layers_23_self_attn_linear_v_weight_quantized, x = query)[name = tensor("linear_212")]; tensor var_4268 = const()[name = tensor("op_4268"), val = tensor([1, -1, 8, 128])]; tensor v = reshape(shape = var_4268, x = var_4267)[name = tensor("v")]; tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, -3, -1])]; tensor var_4280 = add(x = q_139, y = encoder_layers_23_self_attn_pos_bias_u)[name = tensor("op_4280")]; tensor var_4282 = add(x = q_139, y = encoder_layers_23_self_attn_pos_bias_v)[name = tensor("op_4282")]; tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, -3, -1])]; tensor op_4284_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4284_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596985024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597369536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597369088)))]; tensor x_515_transpose_x_0 = const()[name = tensor("x_515_transpose_x_0"), val = tensor(false)]; tensor x_515_transpose_y_0 = const()[name = tensor("x_515_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v = transpose(perm = q_with_bias_v_perm_0, x = var_4282)[name = tensor("transpose_150")]; tensor x_515 = matmul(transpose_x = x_515_transpose_x_0, transpose_y = x_515_transpose_y_0, x = q_with_bias_v, y = op_4284_quantized)[name = tensor("x_515")]; tensor const_244 = const()[name = tensor("const_244"), val = tensor(0x0p+0)]; tensor x_517_pad_0 = const()[name = tensor("x_517_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_517_mode_0 = const()[name = tensor("x_517_mode_0"), val = tensor("constant")]; tensor x_517 = pad(constant_val = const_244, mode = x_517_mode_0, pad = x_517_pad_0, x = x_515)[name = tensor("x_517")]; tensor var_4292 = const()[name = tensor("op_4292"), val = tensor([1, 8, -1, 188])]; tensor x_519 = reshape(shape = var_4292, x = x_517)[name = tensor("x_519")]; tensor var_4296_begin_0 = const()[name = tensor("op_4296_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_4296_end_0 = const()[name = tensor("op_4296_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_4296_end_mask_0 = const()[name = tensor("op_4296_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_4296 = slice_by_index(begin = var_4296_begin_0, end = var_4296_end_0, end_mask = var_4296_end_mask_0, x = x_519)[name = tensor("op_4296")]; tensor var_4297 = const()[name = tensor("op_4297"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_93 = reshape(shape = var_4297, x = var_4296)[name = tensor("matrix_bd_93")]; tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; tensor transpose_142_perm_0 = const()[name = tensor("transpose_142_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_143_perm_0 = const()[name = tensor("transpose_143_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_143 = transpose(perm = transpose_143_perm_0, x = k_93)[name = tensor("transpose_148")]; tensor transpose_142 = transpose(perm = transpose_142_perm_0, x = var_4280)[name = tensor("transpose_149")]; tensor matrix_ac = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_142, y = transpose_143)[name = tensor("matrix_ac")]; tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_93)[name = tensor("matrix_bd")]; tensor var_4306 = add(x = matrix_ac, y = matrix_bd)[name = tensor("op_4306")]; tensor _inversed_scores_93_y_0 = const()[name = tensor("_inversed_scores_93_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_93 = mul(x = var_4306, y = _inversed_scores_93_y_0)[name = tensor("_inversed_scores_93")]; tensor scores = select(a = var_14, b = _inversed_scores_93, cond = mask_3)[name = tensor("scores")]; tensor var_4312 = softmax(axis = var_32, x = scores)[name = tensor("op_4312")]; tensor input_1229 = select(a = var_13, b = var_4312, cond = mask_3)[name = tensor("input_1229")]; tensor x_521_transpose_x_0 = const()[name = tensor("x_521_transpose_x_0"), val = tensor(false)]; tensor x_521_transpose_y_0 = const()[name = tensor("x_521_transpose_y_0"), val = tensor(false)]; tensor value = transpose(perm = value_perm_0, x = v)[name = tensor("transpose_147")]; tensor x_521 = matmul(transpose_x = x_521_transpose_x_0, transpose_y = x_521_transpose_y_0, x = input_1229, y = value)[name = tensor("x_521")]; tensor var_4316_perm_0 = const()[name = tensor("op_4316_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4317 = const()[name = tensor("op_4317"), val = tensor([1, -1, 1024])]; tensor var_4316 = transpose(perm = var_4316_perm_0, x = x_521)[name = tensor("transpose_146")]; tensor input_1231 = reshape(shape = var_4317, x = var_4316)[name = tensor("input_1231")]; tensor input_1233 = linear(bias = encoder_layers_23_self_attn_linear_out_bias, weight = encoder_layers_23_self_attn_linear_out_weight_quantized, x = input_1231)[name = tensor("linear_214")]; tensor input_1235 = add(x = input_1227, y = input_1233)[name = tensor("input_1235")]; tensor x_525_axes_0 = const()[name = tensor("x_525_axes_0"), val = tensor([-1])]; tensor x_525 = layer_norm(axes = x_525_axes_0, beta = encoder_layers_23_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_23_norm_conv_weight, x = input_1235)[name = tensor("x_525")]; tensor input_1237_perm_0 = const()[name = tensor("input_1237_perm_0"), val = tensor([0, 2, 1])]; tensor input_1239_pad_type_0 = const()[name = tensor("input_1239_pad_type_0"), val = tensor("valid")]; tensor input_1239_strides_0 = const()[name = tensor("input_1239_strides_0"), val = tensor([1])]; tensor input_1239_pad_0 = const()[name = tensor("input_1239_pad_0"), val = tensor([0, 0])]; tensor input_1239_dilations_0 = const()[name = tensor("input_1239_dilations_0"), val = tensor([1])]; tensor input_1239_groups_0 = const()[name = tensor("input_1239_groups_0"), val = tensor(1)]; tensor input_1237 = transpose(perm = input_1237_perm_0, x = x_525)[name = tensor("transpose_145")]; tensor input_1239 = conv(bias = encoder_layers_23_conv_pointwise_conv1_bias, dilations = input_1239_dilations_0, groups = input_1239_groups_0, pad = input_1239_pad_0, pad_type = input_1239_pad_type_0, strides = input_1239_strides_0, weight = encoder_layers_23_conv_pointwise_conv1_weight_quantized, x = input_1237)[name = tensor("input_1239")]; tensor x_527_split_num_splits_0 = const()[name = tensor("x_527_split_num_splits_0"), val = tensor(2)]; tensor x_527_split_axis_0 = const()[name = tensor("x_527_split_axis_0"), val = tensor(1)]; tensor x_527_split_0, tensor x_527_split_1 = split(axis = x_527_split_axis_0, num_splits = x_527_split_num_splits_0, x = input_1239)[name = tensor("x_527_split")]; tensor x_527_split_1_sigmoid = sigmoid(x = x_527_split_1)[name = tensor("x_527_split_1_sigmoid")]; tensor x_527 = mul(x = x_527_split_0, y = x_527_split_1_sigmoid)[name = tensor("x_527")]; tensor input_1241 = select(a = var_13, b = x_527, cond = var_339)[name = tensor("input_1241")]; tensor const_247 = const()[name = tensor("const_247"), val = tensor(0x0p+0)]; tensor input_1243_pad_0 = const()[name = tensor("input_1243_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1243_mode_0 = const()[name = tensor("input_1243_mode_0"), val = tensor("constant")]; tensor input_1243 = pad(constant_val = const_247, mode = input_1243_mode_0, pad = input_1243_pad_0, x = input_1241)[name = tensor("input_1243")]; tensor input_1245_pad_type_0 = const()[name = tensor("input_1245_pad_type_0"), val = tensor("valid")]; tensor input_1245_groups_0 = const()[name = tensor("input_1245_groups_0"), val = tensor(1024)]; tensor input_1245_strides_0 = const()[name = tensor("input_1245_strides_0"), val = tensor([1])]; tensor input_1245_pad_0 = const()[name = tensor("input_1245_pad_0"), val = tensor([0, 0])]; tensor input_1245_dilations_0 = const()[name = tensor("input_1245_dilations_0"), val = tensor([1])]; tensor const_294_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("const_294_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597371136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597381504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597380416)))]; tensor const_295 = const()[name = tensor("const_295"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(597385664)))]; tensor input_1247 = conv(bias = const_295, dilations = input_1245_dilations_0, groups = input_1245_groups_0, pad = input_1245_pad_0, pad_type = input_1245_pad_type_0, strides = input_1245_strides_0, weight = const_294_quantized, x = input_1243)[name = tensor("input_1247")]; tensor input_1249 = silu(x = input_1247)[name = tensor("input_1249")]; tensor x_529_pad_type_0 = const()[name = tensor("x_529_pad_type_0"), val = tensor("valid")]; tensor x_529_strides_0 = const()[name = tensor("x_529_strides_0"), val = tensor([1])]; tensor x_529_pad_0 = const()[name = tensor("x_529_pad_0"), val = tensor([0, 0])]; tensor x_529_dilations_0 = const()[name = tensor("x_529_dilations_0"), val = tensor([1])]; tensor x_529_groups_0 = const()[name = tensor("x_529_groups_0"), val = tensor(1)]; tensor x_529 = conv(bias = encoder_layers_23_conv_pointwise_conv2_bias, dilations = x_529_dilations_0, groups = x_529_groups_0, pad = x_529_pad_0, pad_type = x_529_pad_type_0, strides = x_529_strides_0, weight = encoder_layers_23_conv_pointwise_conv2_weight_quantized, x = input_1249)[name = tensor("x_529")]; tensor input_1251_perm_0 = const()[name = tensor("input_1251_perm_0"), val = tensor([0, 2, 1])]; tensor input_1251 = transpose(perm = input_1251_perm_0, x = x_529)[name = tensor("transpose_144")]; tensor input_1253 = add(x = input_1235, y = input_1251)[name = tensor("input_1253")]; tensor input_1255_axes_0 = const()[name = tensor("input_1255_axes_0"), val = tensor([-1])]; tensor input_1255 = layer_norm(axes = input_1255_axes_0, beta = encoder_layers_23_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_23_norm_feed_forward2_weight, x = input_1253)[name = tensor("input_1255")]; tensor input_1257 = linear(bias = encoder_layers_23_feed_forward2_linear1_bias, weight = encoder_layers_23_feed_forward2_linear1_weight_quantized, x = input_1255)[name = tensor("linear_215")]; tensor input_1259 = silu(x = input_1257)[name = tensor("input_1259")]; tensor input_1263 = linear(bias = encoder_layers_23_feed_forward2_linear2_bias, weight = encoder_layers_23_feed_forward2_linear2_weight_quantized, x = input_1259)[name = tensor("linear_216")]; tensor var_4383 = const()[name = tensor("op_4383"), val = tensor(0x1p-1)]; tensor var_4384 = mul(x = input_1263, y = var_4383)[name = tensor("op_4384")]; tensor input = add(x = input_1253, y = var_4384)[name = tensor("input")]; tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; tensor audio_signal = layer_norm(axes = audio_signal_axes_0, beta = encoder_layers_23_norm_out_bias, epsilon = var_11, gamma = encoder_layers_23_norm_out_weight, x = input)[name = tensor("audio_signal")]; tensor ctc_head_raw_output = linear(bias = proj_bias, weight = proj_weight_quantized, x = audio_signal)[name = tensor("linear_217")]; } -> (ctc_head_raw_output, encoder_length); }