program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0b1"}})] { func main(tensor mel_length, tensor melspectrogram_features) { tensor proj_bias = const()[name = tensor("proj_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor proj_weight = const()[name = tensor("proj_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4288)))]; tensor encoder_pre_encode_conv_0_bias = const()[name = tensor("encoder_pre_encode_conv_0_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4202752)))]; tensor encoder_pre_encode_conv_0_weight = const()[name = tensor("encoder_pre_encode_conv_0_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4203840)))]; tensor encoder_pre_encode_conv_2_bias = const()[name = tensor("encoder_pre_encode_conv_2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4213120)))]; tensor encoder_pre_encode_conv_2_weight = const()[name = tensor("encoder_pre_encode_conv_2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4214208)))]; tensor encoder_pre_encode_conv_3_bias = const()[name = tensor("encoder_pre_encode_conv_3_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4223488)))]; tensor encoder_pre_encode_conv_3_weight = const()[name = tensor("encoder_pre_encode_conv_3_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4224576)))]; tensor encoder_pre_encode_conv_5_bias = const()[name = tensor("encoder_pre_encode_conv_5_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4486784)))]; tensor encoder_pre_encode_conv_5_weight = const()[name = tensor("encoder_pre_encode_conv_5_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4487872)))]; tensor encoder_pre_encode_conv_6_bias = const()[name = tensor("encoder_pre_encode_conv_6_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4497152)))]; tensor encoder_pre_encode_conv_6_weight = const()[name = tensor("encoder_pre_encode_conv_6_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4498240)))]; tensor encoder_pre_encode_out_bias = const()[name = tensor("encoder_pre_encode_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4760448)))]; tensor encoder_pre_encode_out_weight = const()[name = tensor("encoder_pre_encode_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4764608)))]; tensor encoder_layers_0_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_0_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15250432)))]; tensor encoder_layers_0_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_0_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15254592)))]; tensor encoder_layers_0_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15258752)))]; tensor encoder_layers_0_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15275200)))]; tensor encoder_layers_0_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32052480)))]; tensor encoder_layers_0_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32056640)))]; tensor encoder_layers_0_norm_self_att_bias = const()[name = tensor("encoder_layers_0_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48833920)))]; tensor encoder_layers_0_norm_self_att_weight = const()[name = tensor("encoder_layers_0_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48838080)))]; tensor encoder_layers_0_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48842240)))]; tensor encoder_layers_0_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48846400)))]; tensor encoder_layers_0_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48850560)))]; tensor encoder_layers_0_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_0_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48854720)))]; tensor encoder_layers_0_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53049088)))]; tensor encoder_layers_0_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_0_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53053248)))]; tensor encoder_layers_0_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57247616)))]; tensor encoder_layers_0_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_0_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57251776)))]; tensor encoder_layers_0_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_0_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61446144)))]; tensor encoder_layers_0_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_0_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61450304)))]; tensor encoder_layers_0_norm_conv_bias = const()[name = tensor("encoder_layers_0_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65644672)))]; tensor encoder_layers_0_norm_conv_weight = const()[name = tensor("encoder_layers_0_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65648832)))]; tensor encoder_layers_0_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65652992)))]; tensor encoder_layers_0_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65661248)))]; tensor encoder_layers_0_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74049920)))]; tensor encoder_layers_0_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74054080)))]; tensor encoder_layers_0_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_0_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78248448)))]; tensor encoder_layers_0_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_0_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78252608)))]; tensor encoder_layers_0_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78256768)))]; tensor encoder_layers_0_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78273216)))]; tensor encoder_layers_0_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95050496)))]; tensor encoder_layers_0_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95054656)))]; tensor encoder_layers_0_norm_out_bias = const()[name = tensor("encoder_layers_0_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111831936)))]; tensor encoder_layers_0_norm_out_weight = const()[name = tensor("encoder_layers_0_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111836096)))]; tensor encoder_layers_1_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_1_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111840256)))]; tensor encoder_layers_1_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_1_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111844416)))]; tensor encoder_layers_1_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111848576)))]; tensor encoder_layers_1_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111865024)))]; tensor encoder_layers_1_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128642304)))]; tensor encoder_layers_1_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128646464)))]; tensor encoder_layers_1_norm_self_att_bias = const()[name = tensor("encoder_layers_1_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145423744)))]; tensor encoder_layers_1_norm_self_att_weight = const()[name = tensor("encoder_layers_1_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145427904)))]; tensor encoder_layers_1_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145432064)))]; tensor encoder_layers_1_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145436224)))]; tensor encoder_layers_1_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145440384)))]; tensor encoder_layers_1_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_1_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145444544)))]; tensor encoder_layers_1_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149638912)))]; tensor encoder_layers_1_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_1_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149643072)))]; tensor encoder_layers_1_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153837440)))]; tensor encoder_layers_1_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_1_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153841600)))]; tensor encoder_layers_1_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_1_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158035968)))]; tensor encoder_layers_1_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_1_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158040128)))]; tensor encoder_layers_1_norm_conv_bias = const()[name = tensor("encoder_layers_1_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162234496)))]; tensor encoder_layers_1_norm_conv_weight = const()[name = tensor("encoder_layers_1_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162238656)))]; tensor encoder_layers_1_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162242816)))]; tensor encoder_layers_1_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162251072)))]; tensor encoder_layers_1_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170639744)))]; tensor encoder_layers_1_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170643904)))]; tensor encoder_layers_1_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_1_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174838272)))]; tensor encoder_layers_1_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_1_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174842432)))]; tensor encoder_layers_1_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174846592)))]; tensor encoder_layers_1_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174863040)))]; tensor encoder_layers_1_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191640320)))]; tensor encoder_layers_1_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(191644480)))]; tensor encoder_layers_1_norm_out_bias = const()[name = tensor("encoder_layers_1_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208421760)))]; tensor encoder_layers_1_norm_out_weight = const()[name = tensor("encoder_layers_1_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208425920)))]; tensor encoder_layers_2_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_2_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208430080)))]; tensor encoder_layers_2_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_2_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208434240)))]; tensor encoder_layers_2_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208438400)))]; tensor encoder_layers_2_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208454848)))]; tensor encoder_layers_2_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225232128)))]; tensor encoder_layers_2_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225236288)))]; tensor encoder_layers_2_norm_self_att_bias = const()[name = tensor("encoder_layers_2_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242013568)))]; tensor encoder_layers_2_norm_self_att_weight = const()[name = tensor("encoder_layers_2_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242017728)))]; tensor encoder_layers_2_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242021888)))]; tensor encoder_layers_2_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242026048)))]; tensor encoder_layers_2_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242030208)))]; tensor encoder_layers_2_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_2_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242034368)))]; tensor encoder_layers_2_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(246228736)))]; tensor encoder_layers_2_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_2_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(246232896)))]; tensor encoder_layers_2_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(250427264)))]; tensor encoder_layers_2_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_2_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(250431424)))]; tensor encoder_layers_2_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_2_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254625792)))]; tensor encoder_layers_2_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_2_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254629952)))]; tensor encoder_layers_2_norm_conv_bias = const()[name = tensor("encoder_layers_2_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258824320)))]; tensor encoder_layers_2_norm_conv_weight = const()[name = tensor("encoder_layers_2_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258828480)))]; tensor encoder_layers_2_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258832640)))]; tensor encoder_layers_2_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258840896)))]; tensor encoder_layers_2_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267229568)))]; tensor encoder_layers_2_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267233728)))]; tensor encoder_layers_2_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_2_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271428096)))]; tensor encoder_layers_2_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_2_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271432256)))]; tensor encoder_layers_2_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271436416)))]; tensor encoder_layers_2_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271452864)))]; tensor encoder_layers_2_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288230144)))]; tensor encoder_layers_2_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288234304)))]; tensor encoder_layers_2_norm_out_bias = const()[name = tensor("encoder_layers_2_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305011584)))]; tensor encoder_layers_2_norm_out_weight = const()[name = tensor("encoder_layers_2_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305015744)))]; tensor encoder_layers_3_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_3_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305019904)))]; tensor encoder_layers_3_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_3_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305024064)))]; tensor encoder_layers_3_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305028224)))]; tensor encoder_layers_3_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305044672)))]; tensor encoder_layers_3_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321821952)))]; tensor encoder_layers_3_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321826112)))]; tensor encoder_layers_3_norm_self_att_bias = const()[name = tensor("encoder_layers_3_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338603392)))]; tensor encoder_layers_3_norm_self_att_weight = const()[name = tensor("encoder_layers_3_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338607552)))]; tensor encoder_layers_3_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338611712)))]; tensor encoder_layers_3_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338615872)))]; tensor encoder_layers_3_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338620032)))]; tensor encoder_layers_3_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_3_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338624192)))]; tensor encoder_layers_3_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(342818560)))]; tensor encoder_layers_3_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_3_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(342822720)))]; tensor encoder_layers_3_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347017088)))]; tensor encoder_layers_3_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_3_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(347021248)))]; tensor encoder_layers_3_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_3_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351215616)))]; tensor encoder_layers_3_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_3_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351219776)))]; tensor encoder_layers_3_norm_conv_bias = const()[name = tensor("encoder_layers_3_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355414144)))]; tensor encoder_layers_3_norm_conv_weight = const()[name = tensor("encoder_layers_3_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355418304)))]; tensor encoder_layers_3_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355422464)))]; tensor encoder_layers_3_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355430720)))]; tensor encoder_layers_3_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363819392)))]; tensor encoder_layers_3_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363823552)))]; tensor encoder_layers_3_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_3_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368017920)))]; tensor encoder_layers_3_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_3_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368022080)))]; tensor encoder_layers_3_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368026240)))]; tensor encoder_layers_3_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368042688)))]; tensor encoder_layers_3_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384819968)))]; tensor encoder_layers_3_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384824128)))]; tensor encoder_layers_3_norm_out_bias = const()[name = tensor("encoder_layers_3_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401601408)))]; tensor encoder_layers_3_norm_out_weight = const()[name = tensor("encoder_layers_3_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401605568)))]; tensor encoder_layers_4_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_4_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401609728)))]; tensor encoder_layers_4_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_4_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401613888)))]; tensor encoder_layers_4_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401618048)))]; tensor encoder_layers_4_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401634496)))]; tensor encoder_layers_4_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418411776)))]; tensor encoder_layers_4_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418415936)))]; tensor encoder_layers_4_norm_self_att_bias = const()[name = tensor("encoder_layers_4_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435193216)))]; tensor encoder_layers_4_norm_self_att_weight = const()[name = tensor("encoder_layers_4_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435197376)))]; tensor encoder_layers_4_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435201536)))]; tensor encoder_layers_4_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435205696)))]; tensor encoder_layers_4_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435209856)))]; tensor encoder_layers_4_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_4_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435214016)))]; tensor encoder_layers_4_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439408384)))]; tensor encoder_layers_4_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_4_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439412544)))]; tensor encoder_layers_4_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(443606912)))]; tensor encoder_layers_4_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_4_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(443611072)))]; tensor encoder_layers_4_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_4_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(447805440)))]; tensor encoder_layers_4_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_4_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(447809600)))]; tensor encoder_layers_4_norm_conv_bias = const()[name = tensor("encoder_layers_4_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452003968)))]; tensor encoder_layers_4_norm_conv_weight = const()[name = tensor("encoder_layers_4_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452008128)))]; tensor encoder_layers_4_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452012288)))]; tensor encoder_layers_4_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452020544)))]; tensor encoder_layers_4_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(460409216)))]; tensor encoder_layers_4_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(460413376)))]; tensor encoder_layers_4_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_4_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(464607744)))]; tensor encoder_layers_4_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_4_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(464611904)))]; tensor encoder_layers_4_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(464616064)))]; tensor encoder_layers_4_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(464632512)))]; tensor encoder_layers_4_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481409792)))]; tensor encoder_layers_4_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481413952)))]; tensor encoder_layers_4_norm_out_bias = const()[name = tensor("encoder_layers_4_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498191232)))]; tensor encoder_layers_4_norm_out_weight = const()[name = tensor("encoder_layers_4_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498195392)))]; tensor encoder_layers_5_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_5_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498199552)))]; tensor encoder_layers_5_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_5_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498203712)))]; tensor encoder_layers_5_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498207872)))]; tensor encoder_layers_5_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498224320)))]; tensor encoder_layers_5_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515001600)))]; tensor encoder_layers_5_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515005760)))]; tensor encoder_layers_5_norm_self_att_bias = const()[name = tensor("encoder_layers_5_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531783040)))]; tensor encoder_layers_5_norm_self_att_weight = const()[name = tensor("encoder_layers_5_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531787200)))]; tensor encoder_layers_5_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531791360)))]; tensor encoder_layers_5_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531795520)))]; tensor encoder_layers_5_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531799680)))]; tensor encoder_layers_5_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_5_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531803840)))]; tensor encoder_layers_5_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535998208)))]; tensor encoder_layers_5_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_5_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(536002368)))]; tensor encoder_layers_5_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540196736)))]; tensor encoder_layers_5_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_5_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540200896)))]; tensor encoder_layers_5_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_5_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544395264)))]; tensor encoder_layers_5_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_5_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544399424)))]; tensor encoder_layers_5_norm_conv_bias = const()[name = tensor("encoder_layers_5_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548593792)))]; tensor encoder_layers_5_norm_conv_weight = const()[name = tensor("encoder_layers_5_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548597952)))]; tensor encoder_layers_5_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548602112)))]; tensor encoder_layers_5_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548610368)))]; tensor encoder_layers_5_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556999040)))]; tensor encoder_layers_5_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557003200)))]; tensor encoder_layers_5_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_5_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(561197568)))]; tensor encoder_layers_5_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_5_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(561201728)))]; tensor encoder_layers_5_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(561205888)))]; tensor encoder_layers_5_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(561222336)))]; tensor encoder_layers_5_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577999616)))]; tensor encoder_layers_5_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578003776)))]; tensor encoder_layers_5_norm_out_bias = const()[name = tensor("encoder_layers_5_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594781056)))]; tensor encoder_layers_5_norm_out_weight = const()[name = tensor("encoder_layers_5_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594785216)))]; tensor encoder_layers_6_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_6_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594789376)))]; tensor encoder_layers_6_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_6_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594793536)))]; tensor encoder_layers_6_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594797696)))]; tensor encoder_layers_6_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594814144)))]; tensor encoder_layers_6_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(611591424)))]; tensor encoder_layers_6_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(611595584)))]; tensor encoder_layers_6_norm_self_att_bias = const()[name = tensor("encoder_layers_6_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(628372864)))]; tensor encoder_layers_6_norm_self_att_weight = const()[name = tensor("encoder_layers_6_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(628377024)))]; tensor encoder_layers_6_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(628381184)))]; tensor encoder_layers_6_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(628385344)))]; tensor encoder_layers_6_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(628389504)))]; tensor encoder_layers_6_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_6_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(628393664)))]; tensor encoder_layers_6_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(632588032)))]; tensor encoder_layers_6_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_6_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(632592192)))]; tensor encoder_layers_6_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(636786560)))]; tensor encoder_layers_6_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_6_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(636790720)))]; tensor encoder_layers_6_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_6_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(640985088)))]; tensor encoder_layers_6_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_6_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(640989248)))]; tensor encoder_layers_6_norm_conv_bias = const()[name = tensor("encoder_layers_6_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(645183616)))]; tensor encoder_layers_6_norm_conv_weight = const()[name = tensor("encoder_layers_6_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(645187776)))]; tensor encoder_layers_6_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(645191936)))]; tensor encoder_layers_6_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(645200192)))]; tensor encoder_layers_6_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(653588864)))]; tensor encoder_layers_6_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(653593024)))]; tensor encoder_layers_6_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_6_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(657787392)))]; tensor encoder_layers_6_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_6_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(657791552)))]; tensor encoder_layers_6_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(657795712)))]; tensor encoder_layers_6_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(657812160)))]; tensor encoder_layers_6_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(674589440)))]; tensor encoder_layers_6_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(674593600)))]; tensor encoder_layers_6_norm_out_bias = const()[name = tensor("encoder_layers_6_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(691370880)))]; tensor encoder_layers_6_norm_out_weight = const()[name = tensor("encoder_layers_6_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(691375040)))]; tensor encoder_layers_7_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_7_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(691379200)))]; tensor encoder_layers_7_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_7_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(691383360)))]; tensor encoder_layers_7_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(691387520)))]; tensor encoder_layers_7_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(691403968)))]; tensor encoder_layers_7_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(708181248)))]; tensor encoder_layers_7_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(708185408)))]; tensor encoder_layers_7_norm_self_att_bias = const()[name = tensor("encoder_layers_7_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(724962688)))]; tensor encoder_layers_7_norm_self_att_weight = const()[name = tensor("encoder_layers_7_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(724966848)))]; tensor encoder_layers_7_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(724971008)))]; tensor encoder_layers_7_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(724975168)))]; tensor encoder_layers_7_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(724979328)))]; tensor encoder_layers_7_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_7_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(724983488)))]; tensor encoder_layers_7_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(729177856)))]; tensor encoder_layers_7_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_7_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(729182016)))]; tensor encoder_layers_7_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(733376384)))]; tensor encoder_layers_7_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_7_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(733380544)))]; tensor encoder_layers_7_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_7_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(737574912)))]; tensor encoder_layers_7_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_7_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(737579072)))]; tensor encoder_layers_7_norm_conv_bias = const()[name = tensor("encoder_layers_7_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(741773440)))]; tensor encoder_layers_7_norm_conv_weight = const()[name = tensor("encoder_layers_7_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(741777600)))]; tensor encoder_layers_7_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(741781760)))]; tensor encoder_layers_7_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(741790016)))]; tensor encoder_layers_7_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(750178688)))]; tensor encoder_layers_7_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(750182848)))]; tensor encoder_layers_7_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_7_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(754377216)))]; tensor encoder_layers_7_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_7_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(754381376)))]; tensor encoder_layers_7_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(754385536)))]; tensor encoder_layers_7_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(754401984)))]; tensor encoder_layers_7_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(771179264)))]; tensor encoder_layers_7_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(771183424)))]; tensor encoder_layers_7_norm_out_bias = const()[name = tensor("encoder_layers_7_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(787960704)))]; tensor encoder_layers_7_norm_out_weight = const()[name = tensor("encoder_layers_7_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(787964864)))]; tensor encoder_layers_8_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_8_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(787969024)))]; tensor encoder_layers_8_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_8_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(787973184)))]; tensor encoder_layers_8_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(787977344)))]; tensor encoder_layers_8_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(787993792)))]; tensor encoder_layers_8_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(804771072)))]; tensor encoder_layers_8_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(804775232)))]; tensor encoder_layers_8_norm_self_att_bias = const()[name = tensor("encoder_layers_8_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821552512)))]; tensor encoder_layers_8_norm_self_att_weight = const()[name = tensor("encoder_layers_8_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821556672)))]; tensor encoder_layers_8_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821560832)))]; tensor encoder_layers_8_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821564992)))]; tensor encoder_layers_8_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821569152)))]; tensor encoder_layers_8_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_8_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821573312)))]; tensor encoder_layers_8_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(825767680)))]; tensor encoder_layers_8_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_8_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(825771840)))]; tensor encoder_layers_8_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(829966208)))]; tensor encoder_layers_8_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_8_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(829970368)))]; tensor encoder_layers_8_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_8_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(834164736)))]; tensor encoder_layers_8_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_8_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(834168896)))]; tensor encoder_layers_8_norm_conv_bias = const()[name = tensor("encoder_layers_8_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(838363264)))]; tensor encoder_layers_8_norm_conv_weight = const()[name = tensor("encoder_layers_8_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(838367424)))]; tensor encoder_layers_8_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(838371584)))]; tensor encoder_layers_8_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(838379840)))]; tensor encoder_layers_8_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(846768512)))]; tensor encoder_layers_8_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(846772672)))]; tensor encoder_layers_8_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_8_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(850967040)))]; tensor encoder_layers_8_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_8_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(850971200)))]; tensor encoder_layers_8_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(850975360)))]; tensor encoder_layers_8_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(850991808)))]; tensor encoder_layers_8_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(867769088)))]; tensor encoder_layers_8_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(867773248)))]; tensor encoder_layers_8_norm_out_bias = const()[name = tensor("encoder_layers_8_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884550528)))]; tensor encoder_layers_8_norm_out_weight = const()[name = tensor("encoder_layers_8_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884554688)))]; tensor encoder_layers_9_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_9_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884558848)))]; tensor encoder_layers_9_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_9_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884563008)))]; tensor encoder_layers_9_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884567168)))]; tensor encoder_layers_9_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884583616)))]; tensor encoder_layers_9_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(901360896)))]; tensor encoder_layers_9_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(901365056)))]; tensor encoder_layers_9_norm_self_att_bias = const()[name = tensor("encoder_layers_9_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(918142336)))]; tensor encoder_layers_9_norm_self_att_weight = const()[name = tensor("encoder_layers_9_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(918146496)))]; tensor encoder_layers_9_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(918150656)))]; tensor encoder_layers_9_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(918154816)))]; tensor encoder_layers_9_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(918158976)))]; tensor encoder_layers_9_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_9_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(918163136)))]; tensor encoder_layers_9_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(922357504)))]; tensor encoder_layers_9_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_9_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(922361664)))]; tensor encoder_layers_9_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(926556032)))]; tensor encoder_layers_9_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_9_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(926560192)))]; tensor encoder_layers_9_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_9_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(930754560)))]; tensor encoder_layers_9_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_9_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(930758720)))]; tensor encoder_layers_9_norm_conv_bias = const()[name = tensor("encoder_layers_9_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(934953088)))]; tensor encoder_layers_9_norm_conv_weight = const()[name = tensor("encoder_layers_9_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(934957248)))]; tensor encoder_layers_9_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(934961408)))]; tensor encoder_layers_9_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(934969664)))]; tensor encoder_layers_9_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(943358336)))]; tensor encoder_layers_9_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(943362496)))]; tensor encoder_layers_9_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_9_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(947556864)))]; tensor encoder_layers_9_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_9_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(947561024)))]; tensor encoder_layers_9_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(947565184)))]; tensor encoder_layers_9_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(947581632)))]; tensor encoder_layers_9_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(964358912)))]; tensor encoder_layers_9_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(964363072)))]; tensor encoder_layers_9_norm_out_bias = const()[name = tensor("encoder_layers_9_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(981140352)))]; tensor encoder_layers_9_norm_out_weight = const()[name = tensor("encoder_layers_9_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(981144512)))]; tensor encoder_layers_10_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_10_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(981148672)))]; tensor encoder_layers_10_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_10_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(981152832)))]; tensor encoder_layers_10_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(981156992)))]; tensor encoder_layers_10_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(981173440)))]; tensor encoder_layers_10_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(997950720)))]; tensor encoder_layers_10_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(997954880)))]; tensor encoder_layers_10_norm_self_att_bias = const()[name = tensor("encoder_layers_10_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1014732160)))]; tensor encoder_layers_10_norm_self_att_weight = const()[name = tensor("encoder_layers_10_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1014736320)))]; tensor encoder_layers_10_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1014740480)))]; tensor encoder_layers_10_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1014744640)))]; tensor encoder_layers_10_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1014748800)))]; tensor encoder_layers_10_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_10_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1014752960)))]; tensor encoder_layers_10_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1018947328)))]; tensor encoder_layers_10_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_10_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1018951488)))]; tensor encoder_layers_10_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1023145856)))]; tensor encoder_layers_10_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_10_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1023150016)))]; tensor encoder_layers_10_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_10_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1027344384)))]; tensor encoder_layers_10_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_10_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1027348544)))]; tensor encoder_layers_10_norm_conv_bias = const()[name = tensor("encoder_layers_10_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031542912)))]; tensor encoder_layers_10_norm_conv_weight = const()[name = tensor("encoder_layers_10_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031547072)))]; tensor encoder_layers_10_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031551232)))]; tensor encoder_layers_10_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031559488)))]; tensor encoder_layers_10_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1039948160)))]; tensor encoder_layers_10_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1039952320)))]; tensor encoder_layers_10_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_10_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1044146688)))]; tensor encoder_layers_10_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_10_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1044150848)))]; tensor encoder_layers_10_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1044155008)))]; tensor encoder_layers_10_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1044171456)))]; tensor encoder_layers_10_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1060948736)))]; tensor encoder_layers_10_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1060952896)))]; tensor encoder_layers_10_norm_out_bias = const()[name = tensor("encoder_layers_10_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077730176)))]; tensor encoder_layers_10_norm_out_weight = const()[name = tensor("encoder_layers_10_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077734336)))]; tensor encoder_layers_11_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_11_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077738496)))]; tensor encoder_layers_11_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_11_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077742656)))]; tensor encoder_layers_11_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077746816)))]; tensor encoder_layers_11_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077763264)))]; tensor encoder_layers_11_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1094540544)))]; tensor encoder_layers_11_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1094544704)))]; tensor encoder_layers_11_norm_self_att_bias = const()[name = tensor("encoder_layers_11_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1111321984)))]; tensor encoder_layers_11_norm_self_att_weight = const()[name = tensor("encoder_layers_11_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1111326144)))]; tensor encoder_layers_11_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1111330304)))]; tensor encoder_layers_11_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1111334464)))]; tensor encoder_layers_11_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1111338624)))]; tensor encoder_layers_11_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_11_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1111342784)))]; tensor encoder_layers_11_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1115537152)))]; tensor encoder_layers_11_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_11_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1115541312)))]; tensor encoder_layers_11_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1119735680)))]; tensor encoder_layers_11_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_11_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1119739840)))]; tensor encoder_layers_11_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_11_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1123934208)))]; tensor encoder_layers_11_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_11_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1123938368)))]; tensor encoder_layers_11_norm_conv_bias = const()[name = tensor("encoder_layers_11_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1128132736)))]; tensor encoder_layers_11_norm_conv_weight = const()[name = tensor("encoder_layers_11_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1128136896)))]; tensor encoder_layers_11_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1128141056)))]; tensor encoder_layers_11_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1128149312)))]; tensor encoder_layers_11_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136537984)))]; tensor encoder_layers_11_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136542144)))]; tensor encoder_layers_11_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_11_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1140736512)))]; tensor encoder_layers_11_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_11_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1140740672)))]; tensor encoder_layers_11_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1140744832)))]; tensor encoder_layers_11_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1140761280)))]; tensor encoder_layers_11_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1157538560)))]; tensor encoder_layers_11_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1157542720)))]; tensor encoder_layers_11_norm_out_bias = const()[name = tensor("encoder_layers_11_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1174320000)))]; tensor encoder_layers_11_norm_out_weight = const()[name = tensor("encoder_layers_11_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1174324160)))]; tensor encoder_layers_12_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_12_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1174328320)))]; tensor encoder_layers_12_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_12_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1174332480)))]; tensor encoder_layers_12_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1174336640)))]; tensor encoder_layers_12_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1174353088)))]; tensor encoder_layers_12_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1191130368)))]; tensor encoder_layers_12_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1191134528)))]; tensor encoder_layers_12_norm_self_att_bias = const()[name = tensor("encoder_layers_12_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1207911808)))]; tensor encoder_layers_12_norm_self_att_weight = const()[name = tensor("encoder_layers_12_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1207915968)))]; tensor encoder_layers_12_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1207920128)))]; tensor encoder_layers_12_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1207924288)))]; tensor encoder_layers_12_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1207928448)))]; tensor encoder_layers_12_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_12_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1207932608)))]; tensor encoder_layers_12_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1212126976)))]; tensor encoder_layers_12_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_12_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1212131136)))]; tensor encoder_layers_12_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1216325504)))]; tensor encoder_layers_12_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_12_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1216329664)))]; tensor encoder_layers_12_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_12_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1220524032)))]; tensor encoder_layers_12_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_12_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1220528192)))]; tensor encoder_layers_12_norm_conv_bias = const()[name = tensor("encoder_layers_12_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1224722560)))]; tensor encoder_layers_12_norm_conv_weight = const()[name = tensor("encoder_layers_12_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1224726720)))]; tensor encoder_layers_12_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1224730880)))]; tensor encoder_layers_12_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1224739136)))]; tensor encoder_layers_12_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1233127808)))]; tensor encoder_layers_12_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1233131968)))]; tensor encoder_layers_12_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_12_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1237326336)))]; tensor encoder_layers_12_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_12_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1237330496)))]; tensor encoder_layers_12_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1237334656)))]; tensor encoder_layers_12_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1237351104)))]; tensor encoder_layers_12_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1254128384)))]; tensor encoder_layers_12_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1254132544)))]; tensor encoder_layers_12_norm_out_bias = const()[name = tensor("encoder_layers_12_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1270909824)))]; tensor encoder_layers_12_norm_out_weight = const()[name = tensor("encoder_layers_12_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1270913984)))]; tensor encoder_layers_13_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_13_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1270918144)))]; tensor encoder_layers_13_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_13_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1270922304)))]; tensor encoder_layers_13_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1270926464)))]; tensor encoder_layers_13_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1270942912)))]; tensor encoder_layers_13_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1287720192)))]; tensor encoder_layers_13_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1287724352)))]; tensor encoder_layers_13_norm_self_att_bias = const()[name = tensor("encoder_layers_13_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1304501632)))]; tensor encoder_layers_13_norm_self_att_weight = const()[name = tensor("encoder_layers_13_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1304505792)))]; tensor encoder_layers_13_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1304509952)))]; tensor encoder_layers_13_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1304514112)))]; tensor encoder_layers_13_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1304518272)))]; tensor encoder_layers_13_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_13_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1304522432)))]; tensor encoder_layers_13_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1308716800)))]; tensor encoder_layers_13_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_13_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1308720960)))]; tensor encoder_layers_13_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312915328)))]; tensor encoder_layers_13_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_13_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312919488)))]; tensor encoder_layers_13_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_13_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1317113856)))]; tensor encoder_layers_13_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_13_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1317118016)))]; tensor encoder_layers_13_norm_conv_bias = const()[name = tensor("encoder_layers_13_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1321312384)))]; tensor encoder_layers_13_norm_conv_weight = const()[name = tensor("encoder_layers_13_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1321316544)))]; tensor encoder_layers_13_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1321320704)))]; tensor encoder_layers_13_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1321328960)))]; tensor encoder_layers_13_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1329717632)))]; tensor encoder_layers_13_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1329721792)))]; tensor encoder_layers_13_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_13_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1333916160)))]; tensor encoder_layers_13_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_13_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1333920320)))]; tensor encoder_layers_13_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1333924480)))]; tensor encoder_layers_13_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1333940928)))]; tensor encoder_layers_13_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1350718208)))]; tensor encoder_layers_13_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1350722368)))]; tensor encoder_layers_13_norm_out_bias = const()[name = tensor("encoder_layers_13_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1367499648)))]; tensor encoder_layers_13_norm_out_weight = const()[name = tensor("encoder_layers_13_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1367503808)))]; tensor encoder_layers_14_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_14_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1367507968)))]; tensor encoder_layers_14_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_14_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1367512128)))]; tensor encoder_layers_14_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1367516288)))]; tensor encoder_layers_14_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1367532736)))]; tensor encoder_layers_14_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1384310016)))]; tensor encoder_layers_14_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1384314176)))]; tensor encoder_layers_14_norm_self_att_bias = const()[name = tensor("encoder_layers_14_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1401091456)))]; tensor encoder_layers_14_norm_self_att_weight = const()[name = tensor("encoder_layers_14_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1401095616)))]; tensor encoder_layers_14_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1401099776)))]; tensor encoder_layers_14_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1401103936)))]; tensor encoder_layers_14_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1401108096)))]; tensor encoder_layers_14_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_14_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1401112256)))]; tensor encoder_layers_14_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1405306624)))]; tensor encoder_layers_14_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_14_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1405310784)))]; tensor encoder_layers_14_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1409505152)))]; tensor encoder_layers_14_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_14_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1409509312)))]; tensor encoder_layers_14_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_14_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1413703680)))]; tensor encoder_layers_14_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_14_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1413707840)))]; tensor encoder_layers_14_norm_conv_bias = const()[name = tensor("encoder_layers_14_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1417902208)))]; tensor encoder_layers_14_norm_conv_weight = const()[name = tensor("encoder_layers_14_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1417906368)))]; tensor encoder_layers_14_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1417910528)))]; tensor encoder_layers_14_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1417918784)))]; tensor encoder_layers_14_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1426307456)))]; tensor encoder_layers_14_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1426311616)))]; tensor encoder_layers_14_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_14_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1430505984)))]; tensor encoder_layers_14_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_14_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1430510144)))]; tensor encoder_layers_14_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1430514304)))]; tensor encoder_layers_14_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1430530752)))]; tensor encoder_layers_14_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1447308032)))]; tensor encoder_layers_14_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1447312192)))]; tensor encoder_layers_14_norm_out_bias = const()[name = tensor("encoder_layers_14_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1464089472)))]; tensor encoder_layers_14_norm_out_weight = const()[name = tensor("encoder_layers_14_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1464093632)))]; tensor encoder_layers_15_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_15_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1464097792)))]; tensor encoder_layers_15_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_15_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1464101952)))]; tensor encoder_layers_15_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1464106112)))]; tensor encoder_layers_15_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1464122560)))]; tensor encoder_layers_15_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1480899840)))]; tensor encoder_layers_15_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1480904000)))]; tensor encoder_layers_15_norm_self_att_bias = const()[name = tensor("encoder_layers_15_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1497681280)))]; tensor encoder_layers_15_norm_self_att_weight = const()[name = tensor("encoder_layers_15_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1497685440)))]; tensor encoder_layers_15_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1497689600)))]; tensor encoder_layers_15_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1497693760)))]; tensor encoder_layers_15_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1497697920)))]; tensor encoder_layers_15_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_15_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1497702080)))]; tensor encoder_layers_15_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1501896448)))]; tensor encoder_layers_15_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_15_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1501900608)))]; tensor encoder_layers_15_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1506094976)))]; tensor encoder_layers_15_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_15_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1506099136)))]; tensor encoder_layers_15_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_15_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1510293504)))]; tensor encoder_layers_15_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_15_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1510297664)))]; tensor encoder_layers_15_norm_conv_bias = const()[name = tensor("encoder_layers_15_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1514492032)))]; tensor encoder_layers_15_norm_conv_weight = const()[name = tensor("encoder_layers_15_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1514496192)))]; tensor encoder_layers_15_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1514500352)))]; tensor encoder_layers_15_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1514508608)))]; tensor encoder_layers_15_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1522897280)))]; tensor encoder_layers_15_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1522901440)))]; tensor encoder_layers_15_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_15_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1527095808)))]; tensor encoder_layers_15_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_15_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1527099968)))]; tensor encoder_layers_15_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1527104128)))]; tensor encoder_layers_15_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1527120576)))]; tensor encoder_layers_15_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1543897856)))]; tensor encoder_layers_15_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1543902016)))]; tensor encoder_layers_15_norm_out_bias = const()[name = tensor("encoder_layers_15_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560679296)))]; tensor encoder_layers_15_norm_out_weight = const()[name = tensor("encoder_layers_15_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560683456)))]; tensor encoder_layers_16_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_16_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560687616)))]; tensor encoder_layers_16_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_16_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560691776)))]; tensor encoder_layers_16_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560695936)))]; tensor encoder_layers_16_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560712384)))]; tensor encoder_layers_16_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1577489664)))]; tensor encoder_layers_16_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1577493824)))]; tensor encoder_layers_16_norm_self_att_bias = const()[name = tensor("encoder_layers_16_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1594271104)))]; tensor encoder_layers_16_norm_self_att_weight = const()[name = tensor("encoder_layers_16_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1594275264)))]; tensor encoder_layers_16_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1594279424)))]; tensor encoder_layers_16_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1594283584)))]; tensor encoder_layers_16_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1594287744)))]; tensor encoder_layers_16_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_16_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1594291904)))]; tensor encoder_layers_16_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1598486272)))]; tensor encoder_layers_16_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_16_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1598490432)))]; tensor encoder_layers_16_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1602684800)))]; tensor encoder_layers_16_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_16_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1602688960)))]; tensor encoder_layers_16_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_16_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1606883328)))]; tensor encoder_layers_16_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_16_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1606887488)))]; tensor encoder_layers_16_norm_conv_bias = const()[name = tensor("encoder_layers_16_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1611081856)))]; tensor encoder_layers_16_norm_conv_weight = const()[name = tensor("encoder_layers_16_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1611086016)))]; tensor encoder_layers_16_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1611090176)))]; tensor encoder_layers_16_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1611098432)))]; tensor encoder_layers_16_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1619487104)))]; tensor encoder_layers_16_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1619491264)))]; tensor encoder_layers_16_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_16_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1623685632)))]; tensor encoder_layers_16_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_16_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1623689792)))]; tensor encoder_layers_16_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1623693952)))]; tensor encoder_layers_16_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1623710400)))]; tensor encoder_layers_16_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1640487680)))]; tensor encoder_layers_16_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1640491840)))]; tensor encoder_layers_16_norm_out_bias = const()[name = tensor("encoder_layers_16_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657269120)))]; tensor encoder_layers_16_norm_out_weight = const()[name = tensor("encoder_layers_16_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657273280)))]; tensor encoder_layers_17_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_17_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657277440)))]; tensor encoder_layers_17_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_17_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657281600)))]; tensor encoder_layers_17_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_17_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657285760)))]; tensor encoder_layers_17_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_17_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657302208)))]; tensor encoder_layers_17_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_17_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1674079488)))]; tensor encoder_layers_17_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_17_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1674083648)))]; tensor encoder_layers_17_norm_self_att_bias = const()[name = tensor("encoder_layers_17_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1690860928)))]; tensor encoder_layers_17_norm_self_att_weight = const()[name = tensor("encoder_layers_17_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1690865088)))]; tensor encoder_layers_17_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_17_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1690869248)))]; tensor encoder_layers_17_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_17_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1690873408)))]; tensor encoder_layers_17_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1690877568)))]; tensor encoder_layers_17_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_17_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1690881728)))]; tensor encoder_layers_17_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1695076096)))]; tensor encoder_layers_17_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_17_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1695080256)))]; tensor encoder_layers_17_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1699274624)))]; tensor encoder_layers_17_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_17_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1699278784)))]; tensor encoder_layers_17_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_17_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1703473152)))]; tensor encoder_layers_17_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_17_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1703477312)))]; tensor encoder_layers_17_norm_conv_bias = const()[name = tensor("encoder_layers_17_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1707671680)))]; tensor encoder_layers_17_norm_conv_weight = const()[name = tensor("encoder_layers_17_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1707675840)))]; tensor encoder_layers_17_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_17_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1707680000)))]; tensor encoder_layers_17_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_17_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1707688256)))]; tensor encoder_layers_17_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_17_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1716076928)))]; tensor encoder_layers_17_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_17_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1716081088)))]; tensor encoder_layers_17_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_17_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1720275456)))]; tensor encoder_layers_17_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_17_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1720279616)))]; tensor encoder_layers_17_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_17_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1720283776)))]; tensor encoder_layers_17_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_17_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1720300224)))]; tensor encoder_layers_17_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_17_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1737077504)))]; tensor encoder_layers_17_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_17_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1737081664)))]; tensor encoder_layers_17_norm_out_bias = const()[name = tensor("encoder_layers_17_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1753858944)))]; tensor encoder_layers_17_norm_out_weight = const()[name = tensor("encoder_layers_17_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1753863104)))]; tensor encoder_layers_18_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_18_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1753867264)))]; tensor encoder_layers_18_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_18_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1753871424)))]; tensor encoder_layers_18_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_18_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1753875584)))]; tensor encoder_layers_18_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_18_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1753892032)))]; tensor encoder_layers_18_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_18_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1770669312)))]; tensor encoder_layers_18_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_18_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1770673472)))]; tensor encoder_layers_18_norm_self_att_bias = const()[name = tensor("encoder_layers_18_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1787450752)))]; tensor encoder_layers_18_norm_self_att_weight = const()[name = tensor("encoder_layers_18_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1787454912)))]; tensor encoder_layers_18_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_18_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1787459072)))]; tensor encoder_layers_18_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_18_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1787463232)))]; tensor encoder_layers_18_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1787467392)))]; tensor encoder_layers_18_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_18_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1787471552)))]; tensor encoder_layers_18_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1791665920)))]; tensor encoder_layers_18_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_18_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1791670080)))]; tensor encoder_layers_18_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1795864448)))]; tensor encoder_layers_18_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_18_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1795868608)))]; tensor encoder_layers_18_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_18_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1800062976)))]; tensor encoder_layers_18_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_18_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1800067136)))]; tensor encoder_layers_18_norm_conv_bias = const()[name = tensor("encoder_layers_18_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1804261504)))]; tensor encoder_layers_18_norm_conv_weight = const()[name = tensor("encoder_layers_18_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1804265664)))]; tensor encoder_layers_18_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_18_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1804269824)))]; tensor encoder_layers_18_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_18_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1804278080)))]; tensor encoder_layers_18_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_18_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1812666752)))]; tensor encoder_layers_18_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_18_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1812670912)))]; tensor encoder_layers_18_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_18_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1816865280)))]; tensor encoder_layers_18_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_18_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1816869440)))]; tensor encoder_layers_18_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_18_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1816873600)))]; tensor encoder_layers_18_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_18_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1816890048)))]; tensor encoder_layers_18_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_18_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1833667328)))]; tensor encoder_layers_18_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_18_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1833671488)))]; tensor encoder_layers_18_norm_out_bias = const()[name = tensor("encoder_layers_18_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1850448768)))]; tensor encoder_layers_18_norm_out_weight = const()[name = tensor("encoder_layers_18_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1850452928)))]; tensor encoder_layers_19_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_19_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1850457088)))]; tensor encoder_layers_19_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_19_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1850461248)))]; tensor encoder_layers_19_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_19_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1850465408)))]; tensor encoder_layers_19_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_19_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1850481856)))]; tensor encoder_layers_19_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_19_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1867259136)))]; tensor encoder_layers_19_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_19_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1867263296)))]; tensor encoder_layers_19_norm_self_att_bias = const()[name = tensor("encoder_layers_19_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1884040576)))]; tensor encoder_layers_19_norm_self_att_weight = const()[name = tensor("encoder_layers_19_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1884044736)))]; tensor encoder_layers_19_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_19_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1884048896)))]; tensor encoder_layers_19_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_19_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1884053056)))]; tensor encoder_layers_19_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1884057216)))]; tensor encoder_layers_19_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_19_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1884061376)))]; tensor encoder_layers_19_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1888255744)))]; tensor encoder_layers_19_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_19_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1888259904)))]; tensor encoder_layers_19_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1892454272)))]; tensor encoder_layers_19_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_19_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1892458432)))]; tensor encoder_layers_19_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_19_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1896652800)))]; tensor encoder_layers_19_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_19_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1896656960)))]; tensor encoder_layers_19_norm_conv_bias = const()[name = tensor("encoder_layers_19_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1900851328)))]; tensor encoder_layers_19_norm_conv_weight = const()[name = tensor("encoder_layers_19_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1900855488)))]; tensor encoder_layers_19_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_19_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1900859648)))]; tensor encoder_layers_19_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_19_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1900867904)))]; tensor encoder_layers_19_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_19_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1909256576)))]; tensor encoder_layers_19_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_19_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1909260736)))]; tensor encoder_layers_19_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_19_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1913455104)))]; tensor encoder_layers_19_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_19_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1913459264)))]; tensor encoder_layers_19_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_19_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1913463424)))]; tensor encoder_layers_19_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_19_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1913479872)))]; tensor encoder_layers_19_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_19_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1930257152)))]; tensor encoder_layers_19_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_19_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1930261312)))]; tensor encoder_layers_19_norm_out_bias = const()[name = tensor("encoder_layers_19_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1947038592)))]; tensor encoder_layers_19_norm_out_weight = const()[name = tensor("encoder_layers_19_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1947042752)))]; tensor encoder_layers_20_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_20_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1947046912)))]; tensor encoder_layers_20_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_20_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1947051072)))]; tensor encoder_layers_20_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_20_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1947055232)))]; tensor encoder_layers_20_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_20_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1947071680)))]; tensor encoder_layers_20_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_20_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1963848960)))]; tensor encoder_layers_20_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_20_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1963853120)))]; tensor encoder_layers_20_norm_self_att_bias = const()[name = tensor("encoder_layers_20_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1980630400)))]; tensor encoder_layers_20_norm_self_att_weight = const()[name = tensor("encoder_layers_20_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1980634560)))]; tensor encoder_layers_20_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_20_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1980638720)))]; tensor encoder_layers_20_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_20_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1980642880)))]; tensor encoder_layers_20_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1980647040)))]; tensor encoder_layers_20_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_20_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1980651200)))]; tensor encoder_layers_20_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1984845568)))]; tensor encoder_layers_20_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_20_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1984849728)))]; tensor encoder_layers_20_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1989044096)))]; tensor encoder_layers_20_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_20_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1989048256)))]; tensor encoder_layers_20_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_20_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1993242624)))]; tensor encoder_layers_20_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_20_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1993246784)))]; tensor encoder_layers_20_norm_conv_bias = const()[name = tensor("encoder_layers_20_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1997441152)))]; tensor encoder_layers_20_norm_conv_weight = const()[name = tensor("encoder_layers_20_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1997445312)))]; tensor encoder_layers_20_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_20_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1997449472)))]; tensor encoder_layers_20_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_20_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1997457728)))]; tensor encoder_layers_20_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_20_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2005846400)))]; tensor encoder_layers_20_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_20_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2005850560)))]; tensor encoder_layers_20_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_20_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2010044928)))]; tensor encoder_layers_20_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_20_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2010049088)))]; tensor encoder_layers_20_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_20_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2010053248)))]; tensor encoder_layers_20_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_20_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2010069696)))]; tensor encoder_layers_20_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_20_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2026846976)))]; tensor encoder_layers_20_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_20_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2026851136)))]; tensor encoder_layers_20_norm_out_bias = const()[name = tensor("encoder_layers_20_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2043628416)))]; tensor encoder_layers_20_norm_out_weight = const()[name = tensor("encoder_layers_20_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2043632576)))]; tensor encoder_layers_21_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_21_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2043636736)))]; tensor encoder_layers_21_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_21_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2043640896)))]; tensor encoder_layers_21_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_21_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2043645056)))]; tensor encoder_layers_21_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_21_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2043661504)))]; tensor encoder_layers_21_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_21_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2060438784)))]; tensor encoder_layers_21_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_21_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2060442944)))]; tensor encoder_layers_21_norm_self_att_bias = const()[name = tensor("encoder_layers_21_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2077220224)))]; tensor encoder_layers_21_norm_self_att_weight = const()[name = tensor("encoder_layers_21_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2077224384)))]; tensor encoder_layers_21_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_21_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2077228544)))]; tensor encoder_layers_21_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_21_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2077232704)))]; tensor encoder_layers_21_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2077236864)))]; tensor encoder_layers_21_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_21_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2077241024)))]; tensor encoder_layers_21_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2081435392)))]; tensor encoder_layers_21_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_21_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2081439552)))]; tensor encoder_layers_21_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2085633920)))]; tensor encoder_layers_21_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_21_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2085638080)))]; tensor encoder_layers_21_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_21_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2089832448)))]; tensor encoder_layers_21_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_21_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2089836608)))]; tensor encoder_layers_21_norm_conv_bias = const()[name = tensor("encoder_layers_21_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2094030976)))]; tensor encoder_layers_21_norm_conv_weight = const()[name = tensor("encoder_layers_21_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2094035136)))]; tensor encoder_layers_21_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_21_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2094039296)))]; tensor encoder_layers_21_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_21_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2094047552)))]; tensor encoder_layers_21_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_21_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2102436224)))]; tensor encoder_layers_21_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_21_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2102440384)))]; tensor encoder_layers_21_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_21_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2106634752)))]; tensor encoder_layers_21_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_21_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2106638912)))]; tensor encoder_layers_21_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_21_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2106643072)))]; tensor encoder_layers_21_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_21_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2106659520)))]; tensor encoder_layers_21_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_21_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2123436800)))]; tensor encoder_layers_21_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_21_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2123440960)))]; tensor encoder_layers_21_norm_out_bias = const()[name = tensor("encoder_layers_21_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2140218240)))]; tensor encoder_layers_21_norm_out_weight = const()[name = tensor("encoder_layers_21_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2140222400)))]; tensor encoder_layers_22_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_22_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2140226560)))]; tensor encoder_layers_22_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_22_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2140230720)))]; tensor encoder_layers_22_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_22_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2140234880)))]; tensor encoder_layers_22_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_22_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2140251328)))]; tensor encoder_layers_22_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_22_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2157028608)))]; tensor encoder_layers_22_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_22_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2157032768)))]; tensor encoder_layers_22_norm_self_att_bias = const()[name = tensor("encoder_layers_22_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2173810048)))]; tensor encoder_layers_22_norm_self_att_weight = const()[name = tensor("encoder_layers_22_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2173814208)))]; tensor encoder_layers_22_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_22_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2173818368)))]; tensor encoder_layers_22_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_22_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2173822528)))]; tensor encoder_layers_22_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2173826688)))]; tensor encoder_layers_22_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_22_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2173830848)))]; tensor encoder_layers_22_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2178025216)))]; tensor encoder_layers_22_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_22_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2178029376)))]; tensor encoder_layers_22_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2182223744)))]; tensor encoder_layers_22_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_22_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2182227904)))]; tensor encoder_layers_22_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_22_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2186422272)))]; tensor encoder_layers_22_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_22_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2186426432)))]; tensor encoder_layers_22_norm_conv_bias = const()[name = tensor("encoder_layers_22_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2190620800)))]; tensor encoder_layers_22_norm_conv_weight = const()[name = tensor("encoder_layers_22_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2190624960)))]; tensor encoder_layers_22_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_22_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2190629120)))]; tensor encoder_layers_22_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_22_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2190637376)))]; tensor encoder_layers_22_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_22_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2199026048)))]; tensor encoder_layers_22_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_22_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2199030208)))]; tensor encoder_layers_22_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_22_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2203224576)))]; tensor encoder_layers_22_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_22_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2203228736)))]; tensor encoder_layers_22_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_22_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2203232896)))]; tensor encoder_layers_22_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_22_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2203249344)))]; tensor encoder_layers_22_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_22_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2220026624)))]; tensor encoder_layers_22_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_22_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2220030784)))]; tensor encoder_layers_22_norm_out_bias = const()[name = tensor("encoder_layers_22_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2236808064)))]; tensor encoder_layers_22_norm_out_weight = const()[name = tensor("encoder_layers_22_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2236812224)))]; tensor encoder_layers_23_norm_feed_forward1_bias = const()[name = tensor("encoder_layers_23_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2236816384)))]; tensor encoder_layers_23_norm_feed_forward1_weight = const()[name = tensor("encoder_layers_23_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2236820544)))]; tensor encoder_layers_23_feed_forward1_linear1_bias = const()[name = tensor("encoder_layers_23_feed_forward1_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2236824704)))]; tensor encoder_layers_23_feed_forward1_linear1_weight = const()[name = tensor("encoder_layers_23_feed_forward1_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2236841152)))]; tensor encoder_layers_23_feed_forward1_linear2_bias = const()[name = tensor("encoder_layers_23_feed_forward1_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2253618432)))]; tensor encoder_layers_23_feed_forward1_linear2_weight = const()[name = tensor("encoder_layers_23_feed_forward1_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2253622592)))]; tensor encoder_layers_23_norm_self_att_bias = const()[name = tensor("encoder_layers_23_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2270399872)))]; tensor encoder_layers_23_norm_self_att_weight = const()[name = tensor("encoder_layers_23_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2270404032)))]; tensor encoder_layers_23_self_attn_pos_bias_v = const()[name = tensor("encoder_layers_23_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2270408192)))]; tensor encoder_layers_23_self_attn_pos_bias_u = const()[name = tensor("encoder_layers_23_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2270412352)))]; tensor encoder_layers_23_self_attn_linear_q_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_q_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2270416512)))]; tensor encoder_layers_23_self_attn_linear_q_weight = const()[name = tensor("encoder_layers_23_self_attn_linear_q_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2270420672)))]; tensor encoder_layers_23_self_attn_linear_k_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_k_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2274615040)))]; tensor encoder_layers_23_self_attn_linear_k_weight = const()[name = tensor("encoder_layers_23_self_attn_linear_k_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2274619200)))]; tensor encoder_layers_23_self_attn_linear_v_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_v_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2278813568)))]; tensor encoder_layers_23_self_attn_linear_v_weight = const()[name = tensor("encoder_layers_23_self_attn_linear_v_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2278817728)))]; tensor encoder_layers_23_self_attn_linear_out_bias = const()[name = tensor("encoder_layers_23_self_attn_linear_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2283012096)))]; tensor encoder_layers_23_self_attn_linear_out_weight = const()[name = tensor("encoder_layers_23_self_attn_linear_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2283016256)))]; tensor encoder_layers_23_norm_conv_bias = const()[name = tensor("encoder_layers_23_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2287210624)))]; tensor encoder_layers_23_norm_conv_weight = const()[name = tensor("encoder_layers_23_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2287214784)))]; tensor encoder_layers_23_conv_pointwise_conv1_bias = const()[name = tensor("encoder_layers_23_conv_pointwise_conv1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2287218944)))]; tensor encoder_layers_23_conv_pointwise_conv1_weight = const()[name = tensor("encoder_layers_23_conv_pointwise_conv1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2287227200)))]; tensor encoder_layers_23_conv_pointwise_conv2_bias = const()[name = tensor("encoder_layers_23_conv_pointwise_conv2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2295615872)))]; tensor encoder_layers_23_conv_pointwise_conv2_weight = const()[name = tensor("encoder_layers_23_conv_pointwise_conv2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2295620032)))]; tensor encoder_layers_23_norm_feed_forward2_bias = const()[name = tensor("encoder_layers_23_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2299814400)))]; tensor encoder_layers_23_norm_feed_forward2_weight = const()[name = tensor("encoder_layers_23_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2299818560)))]; tensor encoder_layers_23_feed_forward2_linear1_bias = const()[name = tensor("encoder_layers_23_feed_forward2_linear1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2299822720)))]; tensor encoder_layers_23_feed_forward2_linear1_weight = const()[name = tensor("encoder_layers_23_feed_forward2_linear1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2299839168)))]; tensor encoder_layers_23_feed_forward2_linear2_bias = const()[name = tensor("encoder_layers_23_feed_forward2_linear2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2316616448)))]; tensor encoder_layers_23_feed_forward2_linear2_weight = const()[name = tensor("encoder_layers_23_feed_forward2_linear2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2316620608)))]; tensor encoder_layers_23_norm_out_bias = const()[name = tensor("encoder_layers_23_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2333397888)))]; tensor encoder_layers_23_norm_out_weight = const()[name = tensor("encoder_layers_23_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2333402048)))]; tensor var_11 = const()[name = tensor("op_11"), val = tensor(0x1.4f8b58p-17)]; tensor var_13 = const()[name = tensor("op_13"), val = tensor(0x0p+0)]; tensor var_14 = const()[name = tensor("op_14"), val = tensor(-0x1.388p+13)]; tensor var_32 = const()[name = tensor("op_32"), val = tensor(-1)]; tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; tensor var_88_dtype_0 = const()[name = tensor("op_88_dtype_0"), val = tensor("fp32")]; tensor var_89_promoted = const()[name = tensor("op_89_promoted"), val = tensor(-0x1p+0)]; tensor var_88 = cast(dtype = var_88_dtype_0, x = mel_length)[name = tensor("cast_228")]; tensor var_90 = add(x = var_88, y = var_89_promoted)[name = tensor("op_90")]; tensor _inversed_92_y_0 = const()[name = tensor("_inversed_92_y_0"), val = tensor(0x1p-1)]; tensor _inversed_92 = mul(x = var_90, y = _inversed_92_y_0)[name = tensor("_inversed_92")]; tensor var_93 = const()[name = tensor("op_93"), val = tensor(0x1p+0)]; tensor lengths_1 = add(x = _inversed_92, y = var_93)[name = tensor("lengths_1")]; tensor lengths_3 = floor(x = lengths_1)[name = tensor("lengths_3")]; tensor var_97_promoted = const()[name = tensor("op_97_promoted"), val = tensor(-0x1p+0)]; tensor var_98 = add(x = lengths_3, y = var_97_promoted)[name = tensor("op_98")]; tensor _inversed_100_y_0 = const()[name = tensor("_inversed_100_y_0"), val = tensor(0x1p-1)]; tensor _inversed_100 = mul(x = var_98, y = _inversed_100_y_0)[name = tensor("_inversed_100")]; tensor var_101 = const()[name = tensor("op_101"), val = tensor(0x1p+0)]; tensor lengths_7 = add(x = _inversed_100, y = var_101)[name = tensor("lengths_7")]; tensor lengths_9 = floor(x = lengths_7)[name = tensor("lengths_9")]; tensor var_105_promoted = const()[name = tensor("op_105_promoted"), val = tensor(-0x1p+0)]; tensor var_106 = add(x = lengths_9, y = var_105_promoted)[name = tensor("op_106")]; tensor _inversed_108_y_0 = const()[name = tensor("_inversed_108_y_0"), val = tensor(0x1p-1)]; tensor _inversed_108 = mul(x = var_106, y = _inversed_108_y_0)[name = tensor("_inversed_108")]; tensor var_109 = const()[name = tensor("op_109"), val = tensor(0x1p+0)]; tensor lengths_13 = add(x = _inversed_108, y = var_109)[name = tensor("lengths_13")]; tensor lengths = floor(x = lengths_13)[name = tensor("lengths")]; tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; tensor x_1 = transpose(perm = x_1_perm_0, x = melspectrogram_features)[name = tensor("transpose_290")]; tensor input_1 = expand_dims(axes = input_1_axes_0, x = x_1)[name = tensor("input_1")]; tensor input_3_pad_type_0 = const()[name = tensor("input_3_pad_type_0"), val = tensor("custom")]; tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_3_strides_0 = const()[name = tensor("input_3_strides_0"), val = tensor([2, 2])]; tensor input_3_dilations_0 = const()[name = tensor("input_3_dilations_0"), val = tensor([1, 1])]; tensor input_3_groups_0 = const()[name = tensor("input_3_groups_0"), val = tensor(1)]; tensor input_3 = conv(bias = encoder_pre_encode_conv_0_bias, dilations = input_3_dilations_0, groups = input_3_groups_0, pad = input_3_pad_0, pad_type = input_3_pad_type_0, strides = input_3_strides_0, weight = encoder_pre_encode_conv_0_weight, x = input_1)[name = tensor("input_3")]; tensor input_5 = relu(x = input_3)[name = tensor("input_5")]; tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("custom")]; tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_7_strides_0 = const()[name = tensor("input_7_strides_0"), val = tensor([2, 2])]; tensor input_7_groups_0 = const()[name = tensor("input_7_groups_0"), val = tensor(256)]; tensor input_7_dilations_0 = const()[name = tensor("input_7_dilations_0"), val = tensor([1, 1])]; tensor input_7 = conv(bias = encoder_pre_encode_conv_2_bias, dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = encoder_pre_encode_conv_2_weight, x = input_5)[name = tensor("input_7")]; tensor input_9_pad_type_0 = const()[name = tensor("input_9_pad_type_0"), val = tensor("valid")]; tensor input_9_strides_0 = const()[name = tensor("input_9_strides_0"), val = tensor([1, 1])]; tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_9_dilations_0 = const()[name = tensor("input_9_dilations_0"), val = tensor([1, 1])]; tensor input_9_groups_0 = const()[name = tensor("input_9_groups_0"), val = tensor(1)]; tensor input_9 = conv(bias = encoder_pre_encode_conv_3_bias, dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = encoder_pre_encode_conv_3_weight, x = input_7)[name = tensor("input_9")]; tensor input_11 = relu(x = input_9)[name = tensor("input_11")]; tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("custom")]; tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([2, 2])]; tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(256)]; tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; tensor input_13 = conv(bias = encoder_pre_encode_conv_5_bias, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = encoder_pre_encode_conv_5_weight, x = input_11)[name = tensor("input_13")]; tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("valid")]; tensor input_15_strides_0 = const()[name = tensor("input_15_strides_0"), val = tensor([1, 1])]; tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_15_dilations_0 = const()[name = tensor("input_15_dilations_0"), val = tensor([1, 1])]; tensor input_15_groups_0 = const()[name = tensor("input_15_groups_0"), val = tensor(1)]; tensor input_15 = conv(bias = encoder_pre_encode_conv_6_bias, dilations = input_15_dilations_0, groups = input_15_groups_0, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = input_15_strides_0, weight = encoder_pre_encode_conv_6_weight, x = input_13)[name = tensor("input_15")]; tensor x_3 = relu(x = input_15)[name = tensor("x_3")]; tensor var_159_perm_0 = const()[name = tensor("op_159_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_160 = const()[name = tensor("op_160"), val = tensor([1, 188, -1])]; tensor var_159 = transpose(perm = var_159_perm_0, x = x_3)[name = tensor("transpose_289")]; tensor input_17 = reshape(shape = var_160, x = var_159)[name = tensor("input_17")]; tensor audio_signal_1 = linear(bias = encoder_pre_encode_out_bias, weight = encoder_pre_encode_out_weight, x = input_17)[name = tensor("linear_0")]; tensor padding_length_dtype_0 = const()[name = tensor("padding_length_dtype_0"), val = tensor("int32")]; tensor var_171 = const()[name = tensor("op_171"), val = tensor(0x1p+5)]; tensor x_5 = mul(x = audio_signal_1, y = var_171)[name = tensor("x_5")]; tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; tensor var_200_axes_0 = const()[name = tensor("op_200_axes_0"), val = tensor([-1])]; tensor encoder_length = cast(dtype = padding_length_dtype_0, x = lengths)[name = tensor("cast_227")]; tensor var_200 = expand_dims(axes = var_200_axes_0, x = encoder_length)[name = tensor("op_200")]; tensor pad_mask_1 = less(x = expand_dims_0, y = var_200)[name = tensor("pad_mask_1")]; tensor var_202_axes_0 = const()[name = tensor("op_202_axes_0"), val = tensor([1])]; tensor var_202 = expand_dims(axes = var_202_axes_0, x = pad_mask_1)[name = tensor("op_202")]; tensor var_203 = const()[name = tensor("op_203"), val = tensor([1, 188, 1])]; tensor pad_mask_for_att_mask_1 = tile(reps = var_203, x = var_202)[name = tensor("pad_mask_for_att_mask_1")]; tensor var_205_perm_0 = const()[name = tensor("op_205_perm_0"), val = tensor([0, 2, 1])]; tensor var_205 = transpose(perm = var_205_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_288")]; tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_205)[name = tensor("pad_mask_for_att_mask")]; tensor const_7 = const()[name = tensor("const_7"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_7)[name = tensor("att_mask")]; tensor mask_1 = logical_not(x = att_mask)[name = tensor("mask_1")]; tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; tensor input_21 = layer_norm(axes = input_21_axes_0, beta = encoder_layers_0_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_0_norm_feed_forward1_weight, x = x_5)[name = tensor("input_21")]; tensor input_23 = linear(bias = encoder_layers_0_feed_forward1_linear1_bias, weight = encoder_layers_0_feed_forward1_linear1_weight, x = input_21)[name = tensor("linear_1")]; tensor input_25 = silu(x = input_23)[name = tensor("input_25")]; tensor input_29 = linear(bias = encoder_layers_0_feed_forward1_linear2_bias, weight = encoder_layers_0_feed_forward1_linear2_weight, x = input_25)[name = tensor("linear_2")]; tensor var_238 = const()[name = tensor("op_238"), val = tensor(0x1p-1)]; tensor var_239 = mul(x = input_29, y = var_238)[name = tensor("op_239")]; tensor input_31 = add(x = x_5, y = var_239)[name = tensor("input_31")]; tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; tensor query_1 = layer_norm(axes = query_1_axes_0, beta = encoder_layers_0_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_0_norm_self_att_weight, x = input_31)[name = tensor("query_1")]; tensor var_255 = linear(bias = encoder_layers_0_self_attn_linear_q_bias, weight = encoder_layers_0_self_attn_linear_q_weight, x = query_1)[name = tensor("linear_3")]; tensor var_256 = const()[name = tensor("op_256"), val = tensor([1, -1, 8, 128])]; tensor q_1 = reshape(shape = var_256, x = var_255)[name = tensor("q_1")]; tensor var_260 = linear(bias = encoder_layers_0_self_attn_linear_k_bias, weight = encoder_layers_0_self_attn_linear_k_weight, x = query_1)[name = tensor("linear_4")]; tensor var_261 = const()[name = tensor("op_261"), val = tensor([1, -1, 8, 128])]; tensor k_1 = reshape(shape = var_261, x = var_260)[name = tensor("k_1")]; tensor var_265 = linear(bias = encoder_layers_0_self_attn_linear_v_bias, weight = encoder_layers_0_self_attn_linear_v_weight, x = query_1)[name = tensor("linear_5")]; tensor var_266 = const()[name = tensor("op_266"), val = tensor([1, -1, 8, 128])]; tensor v_1 = reshape(shape = var_266, x = var_265)[name = tensor("v_1")]; tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_278 = add(x = q_1, y = encoder_layers_0_self_attn_pos_bias_u)[name = tensor("op_278")]; tensor var_280 = add(x = q_1, y = encoder_layers_0_self_attn_pos_bias_v)[name = tensor("op_280")]; tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_282 = const()[name = tensor("op_282"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2333406208)))]; tensor x_9_transpose_x_0 = const()[name = tensor("x_9_transpose_x_0"), val = tensor(false)]; tensor x_9_transpose_y_0 = const()[name = tensor("x_9_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_1 = transpose(perm = q_with_bias_v_1_perm_0, x = var_280)[name = tensor("transpose_286")]; tensor x_9 = matmul(transpose_x = x_9_transpose_x_0, transpose_y = x_9_transpose_y_0, x = q_with_bias_v_1, y = var_282)[name = tensor("x_9")]; tensor const_14 = const()[name = tensor("const_14"), val = tensor(0x0p+0)]; tensor x_11_pad_0 = const()[name = tensor("x_11_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_11_mode_0 = const()[name = tensor("x_11_mode_0"), val = tensor("constant")]; tensor x_11 = pad(constant_val = const_14, mode = x_11_mode_0, pad = x_11_pad_0, x = x_9)[name = tensor("x_11")]; tensor var_290 = const()[name = tensor("op_290"), val = tensor([1, 8, -1, 188])]; tensor x_13 = reshape(shape = var_290, x = x_11)[name = tensor("x_13")]; tensor var_294_begin_0 = const()[name = tensor("op_294_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_294_end_0 = const()[name = tensor("op_294_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_294_end_mask_0 = const()[name = tensor("op_294_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_294 = slice_by_index(begin = var_294_begin_0, end = var_294_end_0, end_mask = var_294_end_mask_0, x = x_13)[name = tensor("op_294")]; tensor var_295 = const()[name = tensor("op_295"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_1 = reshape(shape = var_295, x = var_294)[name = tensor("matrix_bd_1")]; tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = k_1)[name = tensor("transpose_284")]; tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = var_278)[name = tensor("transpose_285")]; tensor matrix_ac_1 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_72, y = transpose_73)[name = tensor("matrix_ac_1")]; tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_3 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1)[name = tensor("matrix_bd_3")]; tensor var_304 = add(x = matrix_ac_1, y = matrix_bd_3)[name = tensor("op_304")]; tensor _inversed_scores_1_y_0 = const()[name = tensor("_inversed_scores_1_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_1 = mul(x = var_304, y = _inversed_scores_1_y_0)[name = tensor("_inversed_scores_1")]; tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; tensor scores_3 = select(a = var_14, b = _inversed_scores_1, cond = mask_3)[name = tensor("scores_3")]; tensor var_310 = softmax(axis = var_32, x = scores_3)[name = tensor("op_310")]; tensor input_33 = select(a = var_13, b = var_310, cond = mask_3)[name = tensor("input_33")]; tensor x_15_transpose_x_0 = const()[name = tensor("x_15_transpose_x_0"), val = tensor(false)]; tensor x_15_transpose_y_0 = const()[name = tensor("x_15_transpose_y_0"), val = tensor(false)]; tensor value_3 = transpose(perm = value_3_perm_0, x = v_1)[name = tensor("transpose_287")]; tensor x_15 = matmul(transpose_x = x_15_transpose_x_0, transpose_y = x_15_transpose_y_0, x = input_33, y = value_3)[name = tensor("x_15")]; tensor var_314_perm_0 = const()[name = tensor("op_314_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_315 = const()[name = tensor("op_315"), val = tensor([1, -1, 1024])]; tensor var_314 = transpose(perm = var_314_perm_0, x = x_15)[name = tensor("transpose_283")]; tensor input_35 = reshape(shape = var_315, x = var_314)[name = tensor("input_35")]; tensor input_37 = linear(bias = encoder_layers_0_self_attn_linear_out_bias, weight = encoder_layers_0_self_attn_linear_out_weight, x = input_35)[name = tensor("linear_7")]; tensor input_39 = add(x = input_31, y = input_37)[name = tensor("input_39")]; tensor x_19_axes_0 = const()[name = tensor("x_19_axes_0"), val = tensor([-1])]; tensor x_19 = layer_norm(axes = x_19_axes_0, beta = encoder_layers_0_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_0_norm_conv_weight, x = input_39)[name = tensor("x_19")]; tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; tensor input_41 = transpose(perm = input_41_perm_0, x = x_19)[name = tensor("transpose_282")]; tensor input_43 = conv(bias = encoder_layers_0_conv_pointwise_conv1_bias, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = encoder_layers_0_conv_pointwise_conv1_weight, x = input_41)[name = tensor("input_43")]; tensor x_21_split_num_splits_0 = const()[name = tensor("x_21_split_num_splits_0"), val = tensor(2)]; tensor x_21_split_axis_0 = const()[name = tensor("x_21_split_axis_0"), val = tensor(1)]; tensor x_21_split_0, tensor x_21_split_1 = split(axis = x_21_split_axis_0, num_splits = x_21_split_num_splits_0, x = input_43)[name = tensor("x_21_split")]; tensor x_21_split_1_sigmoid = sigmoid(x = x_21_split_1)[name = tensor("x_21_split_1_sigmoid")]; tensor x_21 = mul(x = x_21_split_0, y = x_21_split_1_sigmoid)[name = tensor("x_21")]; tensor var_339_axes_0 = const()[name = tensor("op_339_axes_0"), val = tensor([1])]; tensor var_339 = expand_dims(axes = var_339_axes_0, x = pad_mask)[name = tensor("op_339")]; tensor input_45 = select(a = var_13, b = x_21, cond = var_339)[name = tensor("input_45")]; tensor const_17 = const()[name = tensor("const_17"), val = tensor(0x0p+0)]; tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; tensor input_47 = pad(constant_val = const_17, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45)[name = tensor("input_47")]; tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(1024)]; tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; tensor const_248 = const()[name = tensor("const_248"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2334942272)))]; tensor const_249 = const()[name = tensor("const_249"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2334979200)))]; tensor input_51 = conv(bias = const_249, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_248, x = input_47)[name = tensor("input_51")]; tensor input_53 = silu(x = input_51)[name = tensor("input_53")]; tensor x_23_pad_type_0 = const()[name = tensor("x_23_pad_type_0"), val = tensor("valid")]; tensor x_23_strides_0 = const()[name = tensor("x_23_strides_0"), val = tensor([1])]; tensor x_23_pad_0 = const()[name = tensor("x_23_pad_0"), val = tensor([0, 0])]; tensor x_23_dilations_0 = const()[name = tensor("x_23_dilations_0"), val = tensor([1])]; tensor x_23_groups_0 = const()[name = tensor("x_23_groups_0"), val = tensor(1)]; tensor x_23 = conv(bias = encoder_layers_0_conv_pointwise_conv2_bias, dilations = x_23_dilations_0, groups = x_23_groups_0, pad = x_23_pad_0, pad_type = x_23_pad_type_0, strides = x_23_strides_0, weight = encoder_layers_0_conv_pointwise_conv2_weight, x = input_53)[name = tensor("x_23")]; tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; tensor input_55 = transpose(perm = input_55_perm_0, x = x_23)[name = tensor("transpose_281")]; tensor input_57 = add(x = input_39, y = input_55)[name = tensor("input_57")]; tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; tensor input_59 = layer_norm(axes = input_59_axes_0, beta = encoder_layers_0_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_0_norm_feed_forward2_weight, x = input_57)[name = tensor("input_59")]; tensor input_61 = linear(bias = encoder_layers_0_feed_forward2_linear1_bias, weight = encoder_layers_0_feed_forward2_linear1_weight, x = input_59)[name = tensor("linear_8")]; tensor input_63 = silu(x = input_61)[name = tensor("input_63")]; tensor input_67 = linear(bias = encoder_layers_0_feed_forward2_linear2_bias, weight = encoder_layers_0_feed_forward2_linear2_weight, x = input_63)[name = tensor("linear_9")]; tensor var_381 = const()[name = tensor("op_381"), val = tensor(0x1p-1)]; tensor var_382 = mul(x = input_67, y = var_381)[name = tensor("op_382")]; tensor input_69 = add(x = input_57, y = var_382)[name = tensor("input_69")]; tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; tensor input_71 = layer_norm(axes = input_71_axes_0, beta = encoder_layers_0_norm_out_bias, epsilon = var_11, gamma = encoder_layers_0_norm_out_weight, x = input_69)[name = tensor("input_71")]; tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; tensor input_73 = layer_norm(axes = input_73_axes_0, beta = encoder_layers_1_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_1_norm_feed_forward1_weight, x = input_71)[name = tensor("input_73")]; tensor input_75 = linear(bias = encoder_layers_1_feed_forward1_linear1_bias, weight = encoder_layers_1_feed_forward1_linear1_weight, x = input_73)[name = tensor("linear_10")]; tensor input_77 = silu(x = input_75)[name = tensor("input_77")]; tensor input_81 = linear(bias = encoder_layers_1_feed_forward1_linear2_bias, weight = encoder_layers_1_feed_forward1_linear2_weight, x = input_77)[name = tensor("linear_11")]; tensor var_412 = const()[name = tensor("op_412"), val = tensor(0x1p-1)]; tensor var_413 = mul(x = input_81, y = var_412)[name = tensor("op_413")]; tensor input_83 = add(x = input_71, y = var_413)[name = tensor("input_83")]; tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; tensor query_3 = layer_norm(axes = query_3_axes_0, beta = encoder_layers_1_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_1_norm_self_att_weight, x = input_83)[name = tensor("query_3")]; tensor var_429 = linear(bias = encoder_layers_1_self_attn_linear_q_bias, weight = encoder_layers_1_self_attn_linear_q_weight, x = query_3)[name = tensor("linear_12")]; tensor var_430 = const()[name = tensor("op_430"), val = tensor([1, -1, 8, 128])]; tensor q_7 = reshape(shape = var_430, x = var_429)[name = tensor("q_7")]; tensor var_434 = linear(bias = encoder_layers_1_self_attn_linear_k_bias, weight = encoder_layers_1_self_attn_linear_k_weight, x = query_3)[name = tensor("linear_13")]; tensor var_435 = const()[name = tensor("op_435"), val = tensor([1, -1, 8, 128])]; tensor k_5 = reshape(shape = var_435, x = var_434)[name = tensor("k_5")]; tensor var_439 = linear(bias = encoder_layers_1_self_attn_linear_v_bias, weight = encoder_layers_1_self_attn_linear_v_weight, x = query_3)[name = tensor("linear_14")]; tensor var_440 = const()[name = tensor("op_440"), val = tensor([1, -1, 8, 128])]; tensor v_3 = reshape(shape = var_440, x = var_439)[name = tensor("v_3")]; tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_452 = add(x = q_7, y = encoder_layers_1_self_attn_pos_bias_u)[name = tensor("op_452")]; tensor var_454 = add(x = q_7, y = encoder_layers_1_self_attn_pos_bias_v)[name = tensor("op_454")]; tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_456 = const()[name = tensor("op_456"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2334983360)))]; tensor x_31_transpose_x_0 = const()[name = tensor("x_31_transpose_x_0"), val = tensor(false)]; tensor x_31_transpose_y_0 = const()[name = tensor("x_31_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_3 = transpose(perm = q_with_bias_v_3_perm_0, x = var_454)[name = tensor("transpose_279")]; tensor x_31 = matmul(transpose_x = x_31_transpose_x_0, transpose_y = x_31_transpose_y_0, x = q_with_bias_v_3, y = var_456)[name = tensor("x_31")]; tensor const_24 = const()[name = tensor("const_24"), val = tensor(0x0p+0)]; tensor x_33_pad_0 = const()[name = tensor("x_33_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_33_mode_0 = const()[name = tensor("x_33_mode_0"), val = tensor("constant")]; tensor x_33 = pad(constant_val = const_24, mode = x_33_mode_0, pad = x_33_pad_0, x = x_31)[name = tensor("x_33")]; tensor var_464 = const()[name = tensor("op_464"), val = tensor([1, 8, -1, 188])]; tensor x_35 = reshape(shape = var_464, x = x_33)[name = tensor("x_35")]; tensor var_468_begin_0 = const()[name = tensor("op_468_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_468_end_0 = const()[name = tensor("op_468_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_468_end_mask_0 = const()[name = tensor("op_468_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_468 = slice_by_index(begin = var_468_begin_0, end = var_468_end_0, end_mask = var_468_end_mask_0, x = x_35)[name = tensor("op_468")]; tensor var_469 = const()[name = tensor("op_469"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_5 = reshape(shape = var_469, x = var_468)[name = tensor("matrix_bd_5")]; tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = k_5)[name = tensor("transpose_277")]; tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = var_452)[name = tensor("transpose_278")]; tensor matrix_ac_3 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_74, y = transpose_75)[name = tensor("matrix_ac_3")]; tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_7 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5)[name = tensor("matrix_bd_7")]; tensor var_478 = add(x = matrix_ac_3, y = matrix_bd_7)[name = tensor("op_478")]; tensor _inversed_scores_5_y_0 = const()[name = tensor("_inversed_scores_5_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_5 = mul(x = var_478, y = _inversed_scores_5_y_0)[name = tensor("_inversed_scores_5")]; tensor scores_7 = select(a = var_14, b = _inversed_scores_5, cond = mask_3)[name = tensor("scores_7")]; tensor var_484 = softmax(axis = var_32, x = scores_7)[name = tensor("op_484")]; tensor input_85 = select(a = var_13, b = var_484, cond = mask_3)[name = tensor("input_85")]; tensor x_37_transpose_x_0 = const()[name = tensor("x_37_transpose_x_0"), val = tensor(false)]; tensor x_37_transpose_y_0 = const()[name = tensor("x_37_transpose_y_0"), val = tensor(false)]; tensor value_5 = transpose(perm = value_5_perm_0, x = v_3)[name = tensor("transpose_280")]; tensor x_37 = matmul(transpose_x = x_37_transpose_x_0, transpose_y = x_37_transpose_y_0, x = input_85, y = value_5)[name = tensor("x_37")]; tensor var_488_perm_0 = const()[name = tensor("op_488_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_489 = const()[name = tensor("op_489"), val = tensor([1, -1, 1024])]; tensor var_488 = transpose(perm = var_488_perm_0, x = x_37)[name = tensor("transpose_276")]; tensor input_87 = reshape(shape = var_489, x = var_488)[name = tensor("input_87")]; tensor input_89 = linear(bias = encoder_layers_1_self_attn_linear_out_bias, weight = encoder_layers_1_self_attn_linear_out_weight, x = input_87)[name = tensor("linear_16")]; tensor input_91 = add(x = input_83, y = input_89)[name = tensor("input_91")]; tensor x_41_axes_0 = const()[name = tensor("x_41_axes_0"), val = tensor([-1])]; tensor x_41 = layer_norm(axes = x_41_axes_0, beta = encoder_layers_1_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_1_norm_conv_weight, x = input_91)[name = tensor("x_41")]; tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; tensor input_93 = transpose(perm = input_93_perm_0, x = x_41)[name = tensor("transpose_275")]; tensor input_95 = conv(bias = encoder_layers_1_conv_pointwise_conv1_bias, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = encoder_layers_1_conv_pointwise_conv1_weight, x = input_93)[name = tensor("input_95")]; tensor x_43_split_num_splits_0 = const()[name = tensor("x_43_split_num_splits_0"), val = tensor(2)]; tensor x_43_split_axis_0 = const()[name = tensor("x_43_split_axis_0"), val = tensor(1)]; tensor x_43_split_0, tensor x_43_split_1 = split(axis = x_43_split_axis_0, num_splits = x_43_split_num_splits_0, x = input_95)[name = tensor("x_43_split")]; tensor x_43_split_1_sigmoid = sigmoid(x = x_43_split_1)[name = tensor("x_43_split_1_sigmoid")]; tensor x_43 = mul(x = x_43_split_0, y = x_43_split_1_sigmoid)[name = tensor("x_43")]; tensor input_97 = select(a = var_13, b = x_43, cond = var_339)[name = tensor("input_97")]; tensor const_27 = const()[name = tensor("const_27"), val = tensor(0x0p+0)]; tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; tensor input_99 = pad(constant_val = const_27, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97)[name = tensor("input_99")]; tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(1024)]; tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; tensor const_250 = const()[name = tensor("const_250"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2336519424)))]; tensor const_251 = const()[name = tensor("const_251"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2336556352)))]; tensor input_103 = conv(bias = const_251, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_250, x = input_99)[name = tensor("input_103")]; tensor input_105 = silu(x = input_103)[name = tensor("input_105")]; tensor x_45_pad_type_0 = const()[name = tensor("x_45_pad_type_0"), val = tensor("valid")]; tensor x_45_strides_0 = const()[name = tensor("x_45_strides_0"), val = tensor([1])]; tensor x_45_pad_0 = const()[name = tensor("x_45_pad_0"), val = tensor([0, 0])]; tensor x_45_dilations_0 = const()[name = tensor("x_45_dilations_0"), val = tensor([1])]; tensor x_45_groups_0 = const()[name = tensor("x_45_groups_0"), val = tensor(1)]; tensor x_45 = conv(bias = encoder_layers_1_conv_pointwise_conv2_bias, dilations = x_45_dilations_0, groups = x_45_groups_0, pad = x_45_pad_0, pad_type = x_45_pad_type_0, strides = x_45_strides_0, weight = encoder_layers_1_conv_pointwise_conv2_weight, x = input_105)[name = tensor("x_45")]; tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; tensor input_107 = transpose(perm = input_107_perm_0, x = x_45)[name = tensor("transpose_274")]; tensor input_109 = add(x = input_91, y = input_107)[name = tensor("input_109")]; tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; tensor input_111 = layer_norm(axes = input_111_axes_0, beta = encoder_layers_1_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_1_norm_feed_forward2_weight, x = input_109)[name = tensor("input_111")]; tensor input_113 = linear(bias = encoder_layers_1_feed_forward2_linear1_bias, weight = encoder_layers_1_feed_forward2_linear1_weight, x = input_111)[name = tensor("linear_17")]; tensor input_115 = silu(x = input_113)[name = tensor("input_115")]; tensor input_119 = linear(bias = encoder_layers_1_feed_forward2_linear2_bias, weight = encoder_layers_1_feed_forward2_linear2_weight, x = input_115)[name = tensor("linear_18")]; tensor var_555 = const()[name = tensor("op_555"), val = tensor(0x1p-1)]; tensor var_556 = mul(x = input_119, y = var_555)[name = tensor("op_556")]; tensor input_121 = add(x = input_109, y = var_556)[name = tensor("input_121")]; tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; tensor input_123 = layer_norm(axes = input_123_axes_0, beta = encoder_layers_1_norm_out_bias, epsilon = var_11, gamma = encoder_layers_1_norm_out_weight, x = input_121)[name = tensor("input_123")]; tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; tensor input_125 = layer_norm(axes = input_125_axes_0, beta = encoder_layers_2_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_2_norm_feed_forward1_weight, x = input_123)[name = tensor("input_125")]; tensor input_127 = linear(bias = encoder_layers_2_feed_forward1_linear1_bias, weight = encoder_layers_2_feed_forward1_linear1_weight, x = input_125)[name = tensor("linear_19")]; tensor input_129 = silu(x = input_127)[name = tensor("input_129")]; tensor input_133 = linear(bias = encoder_layers_2_feed_forward1_linear2_bias, weight = encoder_layers_2_feed_forward1_linear2_weight, x = input_129)[name = tensor("linear_20")]; tensor var_586 = const()[name = tensor("op_586"), val = tensor(0x1p-1)]; tensor var_587 = mul(x = input_133, y = var_586)[name = tensor("op_587")]; tensor input_135 = add(x = input_123, y = var_587)[name = tensor("input_135")]; tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; tensor query_5 = layer_norm(axes = query_5_axes_0, beta = encoder_layers_2_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_2_norm_self_att_weight, x = input_135)[name = tensor("query_5")]; tensor var_603 = linear(bias = encoder_layers_2_self_attn_linear_q_bias, weight = encoder_layers_2_self_attn_linear_q_weight, x = query_5)[name = tensor("linear_21")]; tensor var_604 = const()[name = tensor("op_604"), val = tensor([1, -1, 8, 128])]; tensor q_13 = reshape(shape = var_604, x = var_603)[name = tensor("q_13")]; tensor var_608 = linear(bias = encoder_layers_2_self_attn_linear_k_bias, weight = encoder_layers_2_self_attn_linear_k_weight, x = query_5)[name = tensor("linear_22")]; tensor var_609 = const()[name = tensor("op_609"), val = tensor([1, -1, 8, 128])]; tensor k_9 = reshape(shape = var_609, x = var_608)[name = tensor("k_9")]; tensor var_613 = linear(bias = encoder_layers_2_self_attn_linear_v_bias, weight = encoder_layers_2_self_attn_linear_v_weight, x = query_5)[name = tensor("linear_23")]; tensor var_614 = const()[name = tensor("op_614"), val = tensor([1, -1, 8, 128])]; tensor v_5 = reshape(shape = var_614, x = var_613)[name = tensor("v_5")]; tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_626 = add(x = q_13, y = encoder_layers_2_self_attn_pos_bias_u)[name = tensor("op_626")]; tensor var_628 = add(x = q_13, y = encoder_layers_2_self_attn_pos_bias_v)[name = tensor("op_628")]; tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_630 = const()[name = tensor("op_630"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2336560512)))]; tensor x_53_transpose_x_0 = const()[name = tensor("x_53_transpose_x_0"), val = tensor(false)]; tensor x_53_transpose_y_0 = const()[name = tensor("x_53_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_5 = transpose(perm = q_with_bias_v_5_perm_0, x = var_628)[name = tensor("transpose_272")]; tensor x_53 = matmul(transpose_x = x_53_transpose_x_0, transpose_y = x_53_transpose_y_0, x = q_with_bias_v_5, y = var_630)[name = tensor("x_53")]; tensor const_34 = const()[name = tensor("const_34"), val = tensor(0x0p+0)]; tensor x_55_pad_0 = const()[name = tensor("x_55_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_55_mode_0 = const()[name = tensor("x_55_mode_0"), val = tensor("constant")]; tensor x_55 = pad(constant_val = const_34, mode = x_55_mode_0, pad = x_55_pad_0, x = x_53)[name = tensor("x_55")]; tensor var_638 = const()[name = tensor("op_638"), val = tensor([1, 8, -1, 188])]; tensor x_57 = reshape(shape = var_638, x = x_55)[name = tensor("x_57")]; tensor var_642_begin_0 = const()[name = tensor("op_642_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_642_end_0 = const()[name = tensor("op_642_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_642_end_mask_0 = const()[name = tensor("op_642_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_642 = slice_by_index(begin = var_642_begin_0, end = var_642_end_0, end_mask = var_642_end_mask_0, x = x_57)[name = tensor("op_642")]; tensor var_643 = const()[name = tensor("op_643"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_9 = reshape(shape = var_643, x = var_642)[name = tensor("matrix_bd_9")]; tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = k_9)[name = tensor("transpose_270")]; tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = var_626)[name = tensor("transpose_271")]; tensor matrix_ac_5 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_76, y = transpose_77)[name = tensor("matrix_ac_5")]; tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_11 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9)[name = tensor("matrix_bd_11")]; tensor var_652 = add(x = matrix_ac_5, y = matrix_bd_11)[name = tensor("op_652")]; tensor _inversed_scores_9_y_0 = const()[name = tensor("_inversed_scores_9_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_9 = mul(x = var_652, y = _inversed_scores_9_y_0)[name = tensor("_inversed_scores_9")]; tensor scores_11 = select(a = var_14, b = _inversed_scores_9, cond = mask_3)[name = tensor("scores_11")]; tensor var_658 = softmax(axis = var_32, x = scores_11)[name = tensor("op_658")]; tensor input_137 = select(a = var_13, b = var_658, cond = mask_3)[name = tensor("input_137")]; tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; tensor value_7 = transpose(perm = value_7_perm_0, x = v_5)[name = tensor("transpose_273")]; tensor x_59 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = input_137, y = value_7)[name = tensor("x_59")]; tensor var_662_perm_0 = const()[name = tensor("op_662_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_663 = const()[name = tensor("op_663"), val = tensor([1, -1, 1024])]; tensor var_662 = transpose(perm = var_662_perm_0, x = x_59)[name = tensor("transpose_269")]; tensor input_139 = reshape(shape = var_663, x = var_662)[name = tensor("input_139")]; tensor input_141 = linear(bias = encoder_layers_2_self_attn_linear_out_bias, weight = encoder_layers_2_self_attn_linear_out_weight, x = input_139)[name = tensor("linear_25")]; tensor input_143 = add(x = input_135, y = input_141)[name = tensor("input_143")]; tensor x_63_axes_0 = const()[name = tensor("x_63_axes_0"), val = tensor([-1])]; tensor x_63 = layer_norm(axes = x_63_axes_0, beta = encoder_layers_2_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_2_norm_conv_weight, x = input_143)[name = tensor("x_63")]; tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; tensor input_145 = transpose(perm = input_145_perm_0, x = x_63)[name = tensor("transpose_268")]; tensor input_147 = conv(bias = encoder_layers_2_conv_pointwise_conv1_bias, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = encoder_layers_2_conv_pointwise_conv1_weight, x = input_145)[name = tensor("input_147")]; tensor x_65_split_num_splits_0 = const()[name = tensor("x_65_split_num_splits_0"), val = tensor(2)]; tensor x_65_split_axis_0 = const()[name = tensor("x_65_split_axis_0"), val = tensor(1)]; tensor x_65_split_0, tensor x_65_split_1 = split(axis = x_65_split_axis_0, num_splits = x_65_split_num_splits_0, x = input_147)[name = tensor("x_65_split")]; tensor x_65_split_1_sigmoid = sigmoid(x = x_65_split_1)[name = tensor("x_65_split_1_sigmoid")]; tensor x_65 = mul(x = x_65_split_0, y = x_65_split_1_sigmoid)[name = tensor("x_65")]; tensor input_149 = select(a = var_13, b = x_65, cond = var_339)[name = tensor("input_149")]; tensor const_37 = const()[name = tensor("const_37"), val = tensor(0x0p+0)]; tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; tensor input_151 = pad(constant_val = const_37, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149)[name = tensor("input_151")]; tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(1024)]; tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; tensor const_252 = const()[name = tensor("const_252"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2338096576)))]; tensor const_253 = const()[name = tensor("const_253"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2338133504)))]; tensor input_155 = conv(bias = const_253, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_252, x = input_151)[name = tensor("input_155")]; tensor input_157 = silu(x = input_155)[name = tensor("input_157")]; tensor x_67_pad_type_0 = const()[name = tensor("x_67_pad_type_0"), val = tensor("valid")]; tensor x_67_strides_0 = const()[name = tensor("x_67_strides_0"), val = tensor([1])]; tensor x_67_pad_0 = const()[name = tensor("x_67_pad_0"), val = tensor([0, 0])]; tensor x_67_dilations_0 = const()[name = tensor("x_67_dilations_0"), val = tensor([1])]; tensor x_67_groups_0 = const()[name = tensor("x_67_groups_0"), val = tensor(1)]; tensor x_67 = conv(bias = encoder_layers_2_conv_pointwise_conv2_bias, dilations = x_67_dilations_0, groups = x_67_groups_0, pad = x_67_pad_0, pad_type = x_67_pad_type_0, strides = x_67_strides_0, weight = encoder_layers_2_conv_pointwise_conv2_weight, x = input_157)[name = tensor("x_67")]; tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; tensor input_159 = transpose(perm = input_159_perm_0, x = x_67)[name = tensor("transpose_267")]; tensor input_161 = add(x = input_143, y = input_159)[name = tensor("input_161")]; tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; tensor input_163 = layer_norm(axes = input_163_axes_0, beta = encoder_layers_2_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_2_norm_feed_forward2_weight, x = input_161)[name = tensor("input_163")]; tensor input_165 = linear(bias = encoder_layers_2_feed_forward2_linear1_bias, weight = encoder_layers_2_feed_forward2_linear1_weight, x = input_163)[name = tensor("linear_26")]; tensor input_167 = silu(x = input_165)[name = tensor("input_167")]; tensor input_171 = linear(bias = encoder_layers_2_feed_forward2_linear2_bias, weight = encoder_layers_2_feed_forward2_linear2_weight, x = input_167)[name = tensor("linear_27")]; tensor var_729 = const()[name = tensor("op_729"), val = tensor(0x1p-1)]; tensor var_730 = mul(x = input_171, y = var_729)[name = tensor("op_730")]; tensor input_173 = add(x = input_161, y = var_730)[name = tensor("input_173")]; tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; tensor input_175 = layer_norm(axes = input_175_axes_0, beta = encoder_layers_2_norm_out_bias, epsilon = var_11, gamma = encoder_layers_2_norm_out_weight, x = input_173)[name = tensor("input_175")]; tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; tensor input_177 = layer_norm(axes = input_177_axes_0, beta = encoder_layers_3_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_3_norm_feed_forward1_weight, x = input_175)[name = tensor("input_177")]; tensor input_179 = linear(bias = encoder_layers_3_feed_forward1_linear1_bias, weight = encoder_layers_3_feed_forward1_linear1_weight, x = input_177)[name = tensor("linear_28")]; tensor input_181 = silu(x = input_179)[name = tensor("input_181")]; tensor input_185 = linear(bias = encoder_layers_3_feed_forward1_linear2_bias, weight = encoder_layers_3_feed_forward1_linear2_weight, x = input_181)[name = tensor("linear_29")]; tensor var_760 = const()[name = tensor("op_760"), val = tensor(0x1p-1)]; tensor var_761 = mul(x = input_185, y = var_760)[name = tensor("op_761")]; tensor input_187 = add(x = input_175, y = var_761)[name = tensor("input_187")]; tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; tensor query_7 = layer_norm(axes = query_7_axes_0, beta = encoder_layers_3_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_3_norm_self_att_weight, x = input_187)[name = tensor("query_7")]; tensor var_777 = linear(bias = encoder_layers_3_self_attn_linear_q_bias, weight = encoder_layers_3_self_attn_linear_q_weight, x = query_7)[name = tensor("linear_30")]; tensor var_778 = const()[name = tensor("op_778"), val = tensor([1, -1, 8, 128])]; tensor q_19 = reshape(shape = var_778, x = var_777)[name = tensor("q_19")]; tensor var_782 = linear(bias = encoder_layers_3_self_attn_linear_k_bias, weight = encoder_layers_3_self_attn_linear_k_weight, x = query_7)[name = tensor("linear_31")]; tensor var_783 = const()[name = tensor("op_783"), val = tensor([1, -1, 8, 128])]; tensor k_13 = reshape(shape = var_783, x = var_782)[name = tensor("k_13")]; tensor var_787 = linear(bias = encoder_layers_3_self_attn_linear_v_bias, weight = encoder_layers_3_self_attn_linear_v_weight, x = query_7)[name = tensor("linear_32")]; tensor var_788 = const()[name = tensor("op_788"), val = tensor([1, -1, 8, 128])]; tensor v_7 = reshape(shape = var_788, x = var_787)[name = tensor("v_7")]; tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_800 = add(x = q_19, y = encoder_layers_3_self_attn_pos_bias_u)[name = tensor("op_800")]; tensor var_802 = add(x = q_19, y = encoder_layers_3_self_attn_pos_bias_v)[name = tensor("op_802")]; tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_804 = const()[name = tensor("op_804"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2338137664)))]; tensor x_75_transpose_x_0 = const()[name = tensor("x_75_transpose_x_0"), val = tensor(false)]; tensor x_75_transpose_y_0 = const()[name = tensor("x_75_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_7 = transpose(perm = q_with_bias_v_7_perm_0, x = var_802)[name = tensor("transpose_265")]; tensor x_75 = matmul(transpose_x = x_75_transpose_x_0, transpose_y = x_75_transpose_y_0, x = q_with_bias_v_7, y = var_804)[name = tensor("x_75")]; tensor const_44 = const()[name = tensor("const_44"), val = tensor(0x0p+0)]; tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_77_mode_0 = const()[name = tensor("x_77_mode_0"), val = tensor("constant")]; tensor x_77 = pad(constant_val = const_44, mode = x_77_mode_0, pad = x_77_pad_0, x = x_75)[name = tensor("x_77")]; tensor var_812 = const()[name = tensor("op_812"), val = tensor([1, 8, -1, 188])]; tensor x_79 = reshape(shape = var_812, x = x_77)[name = tensor("x_79")]; tensor var_816_begin_0 = const()[name = tensor("op_816_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_816_end_0 = const()[name = tensor("op_816_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_816_end_mask_0 = const()[name = tensor("op_816_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_816 = slice_by_index(begin = var_816_begin_0, end = var_816_end_0, end_mask = var_816_end_mask_0, x = x_79)[name = tensor("op_816")]; tensor var_817 = const()[name = tensor("op_817"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_13 = reshape(shape = var_817, x = var_816)[name = tensor("matrix_bd_13")]; tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = k_13)[name = tensor("transpose_263")]; tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = var_800)[name = tensor("transpose_264")]; tensor matrix_ac_7 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_78, y = transpose_79)[name = tensor("matrix_ac_7")]; tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_15 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13)[name = tensor("matrix_bd_15")]; tensor var_826 = add(x = matrix_ac_7, y = matrix_bd_15)[name = tensor("op_826")]; tensor _inversed_scores_13_y_0 = const()[name = tensor("_inversed_scores_13_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_13 = mul(x = var_826, y = _inversed_scores_13_y_0)[name = tensor("_inversed_scores_13")]; tensor scores_15 = select(a = var_14, b = _inversed_scores_13, cond = mask_3)[name = tensor("scores_15")]; tensor var_832 = softmax(axis = var_32, x = scores_15)[name = tensor("op_832")]; tensor input_189 = select(a = var_13, b = var_832, cond = mask_3)[name = tensor("input_189")]; tensor x_81_transpose_x_0 = const()[name = tensor("x_81_transpose_x_0"), val = tensor(false)]; tensor x_81_transpose_y_0 = const()[name = tensor("x_81_transpose_y_0"), val = tensor(false)]; tensor value_9 = transpose(perm = value_9_perm_0, x = v_7)[name = tensor("transpose_266")]; tensor x_81 = matmul(transpose_x = x_81_transpose_x_0, transpose_y = x_81_transpose_y_0, x = input_189, y = value_9)[name = tensor("x_81")]; tensor var_836_perm_0 = const()[name = tensor("op_836_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_837 = const()[name = tensor("op_837"), val = tensor([1, -1, 1024])]; tensor var_836 = transpose(perm = var_836_perm_0, x = x_81)[name = tensor("transpose_262")]; tensor input_191 = reshape(shape = var_837, x = var_836)[name = tensor("input_191")]; tensor input_193 = linear(bias = encoder_layers_3_self_attn_linear_out_bias, weight = encoder_layers_3_self_attn_linear_out_weight, x = input_191)[name = tensor("linear_34")]; tensor input_195 = add(x = input_187, y = input_193)[name = tensor("input_195")]; tensor x_85_axes_0 = const()[name = tensor("x_85_axes_0"), val = tensor([-1])]; tensor x_85 = layer_norm(axes = x_85_axes_0, beta = encoder_layers_3_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_3_norm_conv_weight, x = input_195)[name = tensor("x_85")]; tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; tensor input_197 = transpose(perm = input_197_perm_0, x = x_85)[name = tensor("transpose_261")]; tensor input_199 = conv(bias = encoder_layers_3_conv_pointwise_conv1_bias, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = encoder_layers_3_conv_pointwise_conv1_weight, x = input_197)[name = tensor("input_199")]; tensor x_87_split_num_splits_0 = const()[name = tensor("x_87_split_num_splits_0"), val = tensor(2)]; tensor x_87_split_axis_0 = const()[name = tensor("x_87_split_axis_0"), val = tensor(1)]; tensor x_87_split_0, tensor x_87_split_1 = split(axis = x_87_split_axis_0, num_splits = x_87_split_num_splits_0, x = input_199)[name = tensor("x_87_split")]; tensor x_87_split_1_sigmoid = sigmoid(x = x_87_split_1)[name = tensor("x_87_split_1_sigmoid")]; tensor x_87 = mul(x = x_87_split_0, y = x_87_split_1_sigmoid)[name = tensor("x_87")]; tensor input_201 = select(a = var_13, b = x_87, cond = var_339)[name = tensor("input_201")]; tensor const_47 = const()[name = tensor("const_47"), val = tensor(0x0p+0)]; tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; tensor input_203 = pad(constant_val = const_47, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201)[name = tensor("input_203")]; tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(1024)]; tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; tensor const_254 = const()[name = tensor("const_254"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2339673728)))]; tensor const_255 = const()[name = tensor("const_255"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2339710656)))]; tensor input_207 = conv(bias = const_255, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_254, x = input_203)[name = tensor("input_207")]; tensor input_209 = silu(x = input_207)[name = tensor("input_209")]; tensor x_89_pad_type_0 = const()[name = tensor("x_89_pad_type_0"), val = tensor("valid")]; tensor x_89_strides_0 = const()[name = tensor("x_89_strides_0"), val = tensor([1])]; tensor x_89_pad_0 = const()[name = tensor("x_89_pad_0"), val = tensor([0, 0])]; tensor x_89_dilations_0 = const()[name = tensor("x_89_dilations_0"), val = tensor([1])]; tensor x_89_groups_0 = const()[name = tensor("x_89_groups_0"), val = tensor(1)]; tensor x_89 = conv(bias = encoder_layers_3_conv_pointwise_conv2_bias, dilations = x_89_dilations_0, groups = x_89_groups_0, pad = x_89_pad_0, pad_type = x_89_pad_type_0, strides = x_89_strides_0, weight = encoder_layers_3_conv_pointwise_conv2_weight, x = input_209)[name = tensor("x_89")]; tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; tensor input_211 = transpose(perm = input_211_perm_0, x = x_89)[name = tensor("transpose_260")]; tensor input_213 = add(x = input_195, y = input_211)[name = tensor("input_213")]; tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; tensor input_215 = layer_norm(axes = input_215_axes_0, beta = encoder_layers_3_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_3_norm_feed_forward2_weight, x = input_213)[name = tensor("input_215")]; tensor input_217 = linear(bias = encoder_layers_3_feed_forward2_linear1_bias, weight = encoder_layers_3_feed_forward2_linear1_weight, x = input_215)[name = tensor("linear_35")]; tensor input_219 = silu(x = input_217)[name = tensor("input_219")]; tensor input_223 = linear(bias = encoder_layers_3_feed_forward2_linear2_bias, weight = encoder_layers_3_feed_forward2_linear2_weight, x = input_219)[name = tensor("linear_36")]; tensor var_903 = const()[name = tensor("op_903"), val = tensor(0x1p-1)]; tensor var_904 = mul(x = input_223, y = var_903)[name = tensor("op_904")]; tensor input_225 = add(x = input_213, y = var_904)[name = tensor("input_225")]; tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; tensor input_227 = layer_norm(axes = input_227_axes_0, beta = encoder_layers_3_norm_out_bias, epsilon = var_11, gamma = encoder_layers_3_norm_out_weight, x = input_225)[name = tensor("input_227")]; tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; tensor input_229 = layer_norm(axes = input_229_axes_0, beta = encoder_layers_4_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_4_norm_feed_forward1_weight, x = input_227)[name = tensor("input_229")]; tensor input_231 = linear(bias = encoder_layers_4_feed_forward1_linear1_bias, weight = encoder_layers_4_feed_forward1_linear1_weight, x = input_229)[name = tensor("linear_37")]; tensor input_233 = silu(x = input_231)[name = tensor("input_233")]; tensor input_237 = linear(bias = encoder_layers_4_feed_forward1_linear2_bias, weight = encoder_layers_4_feed_forward1_linear2_weight, x = input_233)[name = tensor("linear_38")]; tensor var_934 = const()[name = tensor("op_934"), val = tensor(0x1p-1)]; tensor var_935 = mul(x = input_237, y = var_934)[name = tensor("op_935")]; tensor input_239 = add(x = input_227, y = var_935)[name = tensor("input_239")]; tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; tensor query_9 = layer_norm(axes = query_9_axes_0, beta = encoder_layers_4_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_4_norm_self_att_weight, x = input_239)[name = tensor("query_9")]; tensor var_951 = linear(bias = encoder_layers_4_self_attn_linear_q_bias, weight = encoder_layers_4_self_attn_linear_q_weight, x = query_9)[name = tensor("linear_39")]; tensor var_952 = const()[name = tensor("op_952"), val = tensor([1, -1, 8, 128])]; tensor q_25 = reshape(shape = var_952, x = var_951)[name = tensor("q_25")]; tensor var_956 = linear(bias = encoder_layers_4_self_attn_linear_k_bias, weight = encoder_layers_4_self_attn_linear_k_weight, x = query_9)[name = tensor("linear_40")]; tensor var_957 = const()[name = tensor("op_957"), val = tensor([1, -1, 8, 128])]; tensor k_17 = reshape(shape = var_957, x = var_956)[name = tensor("k_17")]; tensor var_961 = linear(bias = encoder_layers_4_self_attn_linear_v_bias, weight = encoder_layers_4_self_attn_linear_v_weight, x = query_9)[name = tensor("linear_41")]; tensor var_962 = const()[name = tensor("op_962"), val = tensor([1, -1, 8, 128])]; tensor v_9 = reshape(shape = var_962, x = var_961)[name = tensor("v_9")]; tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_974 = add(x = q_25, y = encoder_layers_4_self_attn_pos_bias_u)[name = tensor("op_974")]; tensor var_976 = add(x = q_25, y = encoder_layers_4_self_attn_pos_bias_v)[name = tensor("op_976")]; tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_978 = const()[name = tensor("op_978"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2339714816)))]; tensor x_97_transpose_x_0 = const()[name = tensor("x_97_transpose_x_0"), val = tensor(false)]; tensor x_97_transpose_y_0 = const()[name = tensor("x_97_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_9 = transpose(perm = q_with_bias_v_9_perm_0, x = var_976)[name = tensor("transpose_258")]; tensor x_97 = matmul(transpose_x = x_97_transpose_x_0, transpose_y = x_97_transpose_y_0, x = q_with_bias_v_9, y = var_978)[name = tensor("x_97")]; tensor const_54 = const()[name = tensor("const_54"), val = tensor(0x0p+0)]; tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_99_mode_0 = const()[name = tensor("x_99_mode_0"), val = tensor("constant")]; tensor x_99 = pad(constant_val = const_54, mode = x_99_mode_0, pad = x_99_pad_0, x = x_97)[name = tensor("x_99")]; tensor var_986 = const()[name = tensor("op_986"), val = tensor([1, 8, -1, 188])]; tensor x_101 = reshape(shape = var_986, x = x_99)[name = tensor("x_101")]; tensor var_990_begin_0 = const()[name = tensor("op_990_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_990_end_0 = const()[name = tensor("op_990_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_990_end_mask_0 = const()[name = tensor("op_990_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_990 = slice_by_index(begin = var_990_begin_0, end = var_990_end_0, end_mask = var_990_end_mask_0, x = x_101)[name = tensor("op_990")]; tensor var_991 = const()[name = tensor("op_991"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_17 = reshape(shape = var_991, x = var_990)[name = tensor("matrix_bd_17")]; tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = k_17)[name = tensor("transpose_256")]; tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = var_974)[name = tensor("transpose_257")]; tensor matrix_ac_9 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_80, y = transpose_81)[name = tensor("matrix_ac_9")]; tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_19 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17)[name = tensor("matrix_bd_19")]; tensor var_1000 = add(x = matrix_ac_9, y = matrix_bd_19)[name = tensor("op_1000")]; tensor _inversed_scores_17_y_0 = const()[name = tensor("_inversed_scores_17_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_17 = mul(x = var_1000, y = _inversed_scores_17_y_0)[name = tensor("_inversed_scores_17")]; tensor scores_19 = select(a = var_14, b = _inversed_scores_17, cond = mask_3)[name = tensor("scores_19")]; tensor var_1006 = softmax(axis = var_32, x = scores_19)[name = tensor("op_1006")]; tensor input_241 = select(a = var_13, b = var_1006, cond = mask_3)[name = tensor("input_241")]; tensor x_103_transpose_x_0 = const()[name = tensor("x_103_transpose_x_0"), val = tensor(false)]; tensor x_103_transpose_y_0 = const()[name = tensor("x_103_transpose_y_0"), val = tensor(false)]; tensor value_11 = transpose(perm = value_11_perm_0, x = v_9)[name = tensor("transpose_259")]; tensor x_103 = matmul(transpose_x = x_103_transpose_x_0, transpose_y = x_103_transpose_y_0, x = input_241, y = value_11)[name = tensor("x_103")]; tensor var_1010_perm_0 = const()[name = tensor("op_1010_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1011 = const()[name = tensor("op_1011"), val = tensor([1, -1, 1024])]; tensor var_1010 = transpose(perm = var_1010_perm_0, x = x_103)[name = tensor("transpose_255")]; tensor input_243 = reshape(shape = var_1011, x = var_1010)[name = tensor("input_243")]; tensor input_245 = linear(bias = encoder_layers_4_self_attn_linear_out_bias, weight = encoder_layers_4_self_attn_linear_out_weight, x = input_243)[name = tensor("linear_43")]; tensor input_247 = add(x = input_239, y = input_245)[name = tensor("input_247")]; tensor x_107_axes_0 = const()[name = tensor("x_107_axes_0"), val = tensor([-1])]; tensor x_107 = layer_norm(axes = x_107_axes_0, beta = encoder_layers_4_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_4_norm_conv_weight, x = input_247)[name = tensor("x_107")]; tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; tensor input_249 = transpose(perm = input_249_perm_0, x = x_107)[name = tensor("transpose_254")]; tensor input_251 = conv(bias = encoder_layers_4_conv_pointwise_conv1_bias, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = encoder_layers_4_conv_pointwise_conv1_weight, x = input_249)[name = tensor("input_251")]; tensor x_109_split_num_splits_0 = const()[name = tensor("x_109_split_num_splits_0"), val = tensor(2)]; tensor x_109_split_axis_0 = const()[name = tensor("x_109_split_axis_0"), val = tensor(1)]; tensor x_109_split_0, tensor x_109_split_1 = split(axis = x_109_split_axis_0, num_splits = x_109_split_num_splits_0, x = input_251)[name = tensor("x_109_split")]; tensor x_109_split_1_sigmoid = sigmoid(x = x_109_split_1)[name = tensor("x_109_split_1_sigmoid")]; tensor x_109 = mul(x = x_109_split_0, y = x_109_split_1_sigmoid)[name = tensor("x_109")]; tensor input_253 = select(a = var_13, b = x_109, cond = var_339)[name = tensor("input_253")]; tensor const_57 = const()[name = tensor("const_57"), val = tensor(0x0p+0)]; tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; tensor input_255 = pad(constant_val = const_57, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253)[name = tensor("input_255")]; tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(1024)]; tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; tensor const_256 = const()[name = tensor("const_256"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2341250880)))]; tensor const_257 = const()[name = tensor("const_257"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2341287808)))]; tensor input_259 = conv(bias = const_257, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_256, x = input_255)[name = tensor("input_259")]; tensor input_261 = silu(x = input_259)[name = tensor("input_261")]; tensor x_111_pad_type_0 = const()[name = tensor("x_111_pad_type_0"), val = tensor("valid")]; tensor x_111_strides_0 = const()[name = tensor("x_111_strides_0"), val = tensor([1])]; tensor x_111_pad_0 = const()[name = tensor("x_111_pad_0"), val = tensor([0, 0])]; tensor x_111_dilations_0 = const()[name = tensor("x_111_dilations_0"), val = tensor([1])]; tensor x_111_groups_0 = const()[name = tensor("x_111_groups_0"), val = tensor(1)]; tensor x_111 = conv(bias = encoder_layers_4_conv_pointwise_conv2_bias, dilations = x_111_dilations_0, groups = x_111_groups_0, pad = x_111_pad_0, pad_type = x_111_pad_type_0, strides = x_111_strides_0, weight = encoder_layers_4_conv_pointwise_conv2_weight, x = input_261)[name = tensor("x_111")]; tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; tensor input_263 = transpose(perm = input_263_perm_0, x = x_111)[name = tensor("transpose_253")]; tensor input_265 = add(x = input_247, y = input_263)[name = tensor("input_265")]; tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; tensor input_267 = layer_norm(axes = input_267_axes_0, beta = encoder_layers_4_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_4_norm_feed_forward2_weight, x = input_265)[name = tensor("input_267")]; tensor input_269 = linear(bias = encoder_layers_4_feed_forward2_linear1_bias, weight = encoder_layers_4_feed_forward2_linear1_weight, x = input_267)[name = tensor("linear_44")]; tensor input_271 = silu(x = input_269)[name = tensor("input_271")]; tensor input_275 = linear(bias = encoder_layers_4_feed_forward2_linear2_bias, weight = encoder_layers_4_feed_forward2_linear2_weight, x = input_271)[name = tensor("linear_45")]; tensor var_1077 = const()[name = tensor("op_1077"), val = tensor(0x1p-1)]; tensor var_1078 = mul(x = input_275, y = var_1077)[name = tensor("op_1078")]; tensor input_277 = add(x = input_265, y = var_1078)[name = tensor("input_277")]; tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; tensor input_279 = layer_norm(axes = input_279_axes_0, beta = encoder_layers_4_norm_out_bias, epsilon = var_11, gamma = encoder_layers_4_norm_out_weight, x = input_277)[name = tensor("input_279")]; tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; tensor input_281 = layer_norm(axes = input_281_axes_0, beta = encoder_layers_5_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_5_norm_feed_forward1_weight, x = input_279)[name = tensor("input_281")]; tensor input_283 = linear(bias = encoder_layers_5_feed_forward1_linear1_bias, weight = encoder_layers_5_feed_forward1_linear1_weight, x = input_281)[name = tensor("linear_46")]; tensor input_285 = silu(x = input_283)[name = tensor("input_285")]; tensor input_289 = linear(bias = encoder_layers_5_feed_forward1_linear2_bias, weight = encoder_layers_5_feed_forward1_linear2_weight, x = input_285)[name = tensor("linear_47")]; tensor var_1108 = const()[name = tensor("op_1108"), val = tensor(0x1p-1)]; tensor var_1109 = mul(x = input_289, y = var_1108)[name = tensor("op_1109")]; tensor input_291 = add(x = input_279, y = var_1109)[name = tensor("input_291")]; tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; tensor query_11 = layer_norm(axes = query_11_axes_0, beta = encoder_layers_5_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_5_norm_self_att_weight, x = input_291)[name = tensor("query_11")]; tensor var_1125 = linear(bias = encoder_layers_5_self_attn_linear_q_bias, weight = encoder_layers_5_self_attn_linear_q_weight, x = query_11)[name = tensor("linear_48")]; tensor var_1126 = const()[name = tensor("op_1126"), val = tensor([1, -1, 8, 128])]; tensor q_31 = reshape(shape = var_1126, x = var_1125)[name = tensor("q_31")]; tensor var_1130 = linear(bias = encoder_layers_5_self_attn_linear_k_bias, weight = encoder_layers_5_self_attn_linear_k_weight, x = query_11)[name = tensor("linear_49")]; tensor var_1131 = const()[name = tensor("op_1131"), val = tensor([1, -1, 8, 128])]; tensor k_21 = reshape(shape = var_1131, x = var_1130)[name = tensor("k_21")]; tensor var_1135 = linear(bias = encoder_layers_5_self_attn_linear_v_bias, weight = encoder_layers_5_self_attn_linear_v_weight, x = query_11)[name = tensor("linear_50")]; tensor var_1136 = const()[name = tensor("op_1136"), val = tensor([1, -1, 8, 128])]; tensor v_11 = reshape(shape = var_1136, x = var_1135)[name = tensor("v_11")]; tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1148 = add(x = q_31, y = encoder_layers_5_self_attn_pos_bias_u)[name = tensor("op_1148")]; tensor var_1150 = add(x = q_31, y = encoder_layers_5_self_attn_pos_bias_v)[name = tensor("op_1150")]; tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1152 = const()[name = tensor("op_1152"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2341291968)))]; tensor x_119_transpose_x_0 = const()[name = tensor("x_119_transpose_x_0"), val = tensor(false)]; tensor x_119_transpose_y_0 = const()[name = tensor("x_119_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_11 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1150)[name = tensor("transpose_251")]; tensor x_119 = matmul(transpose_x = x_119_transpose_x_0, transpose_y = x_119_transpose_y_0, x = q_with_bias_v_11, y = var_1152)[name = tensor("x_119")]; tensor const_64 = const()[name = tensor("const_64"), val = tensor(0x0p+0)]; tensor x_121_pad_0 = const()[name = tensor("x_121_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_121_mode_0 = const()[name = tensor("x_121_mode_0"), val = tensor("constant")]; tensor x_121 = pad(constant_val = const_64, mode = x_121_mode_0, pad = x_121_pad_0, x = x_119)[name = tensor("x_121")]; tensor var_1160 = const()[name = tensor("op_1160"), val = tensor([1, 8, -1, 188])]; tensor x_123 = reshape(shape = var_1160, x = x_121)[name = tensor("x_123")]; tensor var_1164_begin_0 = const()[name = tensor("op_1164_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1164_end_0 = const()[name = tensor("op_1164_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1164_end_mask_0 = const()[name = tensor("op_1164_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1164 = slice_by_index(begin = var_1164_begin_0, end = var_1164_end_0, end_mask = var_1164_end_mask_0, x = x_123)[name = tensor("op_1164")]; tensor var_1165 = const()[name = tensor("op_1165"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_21 = reshape(shape = var_1165, x = var_1164)[name = tensor("matrix_bd_21")]; tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = k_21)[name = tensor("transpose_249")]; tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = var_1148)[name = tensor("transpose_250")]; tensor matrix_ac_11 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_82, y = transpose_83)[name = tensor("matrix_ac_11")]; tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_23 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21)[name = tensor("matrix_bd_23")]; tensor var_1174 = add(x = matrix_ac_11, y = matrix_bd_23)[name = tensor("op_1174")]; tensor _inversed_scores_21_y_0 = const()[name = tensor("_inversed_scores_21_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_21 = mul(x = var_1174, y = _inversed_scores_21_y_0)[name = tensor("_inversed_scores_21")]; tensor scores_23 = select(a = var_14, b = _inversed_scores_21, cond = mask_3)[name = tensor("scores_23")]; tensor var_1180 = softmax(axis = var_32, x = scores_23)[name = tensor("op_1180")]; tensor input_293 = select(a = var_13, b = var_1180, cond = mask_3)[name = tensor("input_293")]; tensor x_125_transpose_x_0 = const()[name = tensor("x_125_transpose_x_0"), val = tensor(false)]; tensor x_125_transpose_y_0 = const()[name = tensor("x_125_transpose_y_0"), val = tensor(false)]; tensor value_13 = transpose(perm = value_13_perm_0, x = v_11)[name = tensor("transpose_252")]; tensor x_125 = matmul(transpose_x = x_125_transpose_x_0, transpose_y = x_125_transpose_y_0, x = input_293, y = value_13)[name = tensor("x_125")]; tensor var_1184_perm_0 = const()[name = tensor("op_1184_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1185 = const()[name = tensor("op_1185"), val = tensor([1, -1, 1024])]; tensor var_1184 = transpose(perm = var_1184_perm_0, x = x_125)[name = tensor("transpose_248")]; tensor input_295 = reshape(shape = var_1185, x = var_1184)[name = tensor("input_295")]; tensor input_297 = linear(bias = encoder_layers_5_self_attn_linear_out_bias, weight = encoder_layers_5_self_attn_linear_out_weight, x = input_295)[name = tensor("linear_52")]; tensor input_299 = add(x = input_291, y = input_297)[name = tensor("input_299")]; tensor x_129_axes_0 = const()[name = tensor("x_129_axes_0"), val = tensor([-1])]; tensor x_129 = layer_norm(axes = x_129_axes_0, beta = encoder_layers_5_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_5_norm_conv_weight, x = input_299)[name = tensor("x_129")]; tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; tensor input_301 = transpose(perm = input_301_perm_0, x = x_129)[name = tensor("transpose_247")]; tensor input_303 = conv(bias = encoder_layers_5_conv_pointwise_conv1_bias, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = encoder_layers_5_conv_pointwise_conv1_weight, x = input_301)[name = tensor("input_303")]; tensor x_131_split_num_splits_0 = const()[name = tensor("x_131_split_num_splits_0"), val = tensor(2)]; tensor x_131_split_axis_0 = const()[name = tensor("x_131_split_axis_0"), val = tensor(1)]; tensor x_131_split_0, tensor x_131_split_1 = split(axis = x_131_split_axis_0, num_splits = x_131_split_num_splits_0, x = input_303)[name = tensor("x_131_split")]; tensor x_131_split_1_sigmoid = sigmoid(x = x_131_split_1)[name = tensor("x_131_split_1_sigmoid")]; tensor x_131 = mul(x = x_131_split_0, y = x_131_split_1_sigmoid)[name = tensor("x_131")]; tensor input_305 = select(a = var_13, b = x_131, cond = var_339)[name = tensor("input_305")]; tensor const_67 = const()[name = tensor("const_67"), val = tensor(0x0p+0)]; tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; tensor input_307 = pad(constant_val = const_67, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305)[name = tensor("input_307")]; tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(1024)]; tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; tensor const_258 = const()[name = tensor("const_258"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2342828032)))]; tensor const_259 = const()[name = tensor("const_259"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2342864960)))]; tensor input_311 = conv(bias = const_259, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_258, x = input_307)[name = tensor("input_311")]; tensor input_313 = silu(x = input_311)[name = tensor("input_313")]; tensor x_133_pad_type_0 = const()[name = tensor("x_133_pad_type_0"), val = tensor("valid")]; tensor x_133_strides_0 = const()[name = tensor("x_133_strides_0"), val = tensor([1])]; tensor x_133_pad_0 = const()[name = tensor("x_133_pad_0"), val = tensor([0, 0])]; tensor x_133_dilations_0 = const()[name = tensor("x_133_dilations_0"), val = tensor([1])]; tensor x_133_groups_0 = const()[name = tensor("x_133_groups_0"), val = tensor(1)]; tensor x_133 = conv(bias = encoder_layers_5_conv_pointwise_conv2_bias, dilations = x_133_dilations_0, groups = x_133_groups_0, pad = x_133_pad_0, pad_type = x_133_pad_type_0, strides = x_133_strides_0, weight = encoder_layers_5_conv_pointwise_conv2_weight, x = input_313)[name = tensor("x_133")]; tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; tensor input_315 = transpose(perm = input_315_perm_0, x = x_133)[name = tensor("transpose_246")]; tensor input_317 = add(x = input_299, y = input_315)[name = tensor("input_317")]; tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; tensor input_319 = layer_norm(axes = input_319_axes_0, beta = encoder_layers_5_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_5_norm_feed_forward2_weight, x = input_317)[name = tensor("input_319")]; tensor input_321 = linear(bias = encoder_layers_5_feed_forward2_linear1_bias, weight = encoder_layers_5_feed_forward2_linear1_weight, x = input_319)[name = tensor("linear_53")]; tensor input_323 = silu(x = input_321)[name = tensor("input_323")]; tensor input_327 = linear(bias = encoder_layers_5_feed_forward2_linear2_bias, weight = encoder_layers_5_feed_forward2_linear2_weight, x = input_323)[name = tensor("linear_54")]; tensor var_1251 = const()[name = tensor("op_1251"), val = tensor(0x1p-1)]; tensor var_1252 = mul(x = input_327, y = var_1251)[name = tensor("op_1252")]; tensor input_329 = add(x = input_317, y = var_1252)[name = tensor("input_329")]; tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; tensor input_331 = layer_norm(axes = input_331_axes_0, beta = encoder_layers_5_norm_out_bias, epsilon = var_11, gamma = encoder_layers_5_norm_out_weight, x = input_329)[name = tensor("input_331")]; tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; tensor input_333 = layer_norm(axes = input_333_axes_0, beta = encoder_layers_6_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_6_norm_feed_forward1_weight, x = input_331)[name = tensor("input_333")]; tensor input_335 = linear(bias = encoder_layers_6_feed_forward1_linear1_bias, weight = encoder_layers_6_feed_forward1_linear1_weight, x = input_333)[name = tensor("linear_55")]; tensor input_337 = silu(x = input_335)[name = tensor("input_337")]; tensor input_341 = linear(bias = encoder_layers_6_feed_forward1_linear2_bias, weight = encoder_layers_6_feed_forward1_linear2_weight, x = input_337)[name = tensor("linear_56")]; tensor var_1282 = const()[name = tensor("op_1282"), val = tensor(0x1p-1)]; tensor var_1283 = mul(x = input_341, y = var_1282)[name = tensor("op_1283")]; tensor input_343 = add(x = input_331, y = var_1283)[name = tensor("input_343")]; tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; tensor query_13 = layer_norm(axes = query_13_axes_0, beta = encoder_layers_6_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_6_norm_self_att_weight, x = input_343)[name = tensor("query_13")]; tensor var_1299 = linear(bias = encoder_layers_6_self_attn_linear_q_bias, weight = encoder_layers_6_self_attn_linear_q_weight, x = query_13)[name = tensor("linear_57")]; tensor var_1300 = const()[name = tensor("op_1300"), val = tensor([1, -1, 8, 128])]; tensor q_37 = reshape(shape = var_1300, x = var_1299)[name = tensor("q_37")]; tensor var_1304 = linear(bias = encoder_layers_6_self_attn_linear_k_bias, weight = encoder_layers_6_self_attn_linear_k_weight, x = query_13)[name = tensor("linear_58")]; tensor var_1305 = const()[name = tensor("op_1305"), val = tensor([1, -1, 8, 128])]; tensor k_25 = reshape(shape = var_1305, x = var_1304)[name = tensor("k_25")]; tensor var_1309 = linear(bias = encoder_layers_6_self_attn_linear_v_bias, weight = encoder_layers_6_self_attn_linear_v_weight, x = query_13)[name = tensor("linear_59")]; tensor var_1310 = const()[name = tensor("op_1310"), val = tensor([1, -1, 8, 128])]; tensor v_13 = reshape(shape = var_1310, x = var_1309)[name = tensor("v_13")]; tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1322 = add(x = q_37, y = encoder_layers_6_self_attn_pos_bias_u)[name = tensor("op_1322")]; tensor var_1324 = add(x = q_37, y = encoder_layers_6_self_attn_pos_bias_v)[name = tensor("op_1324")]; tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1326 = const()[name = tensor("op_1326"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2342869120)))]; tensor x_141_transpose_x_0 = const()[name = tensor("x_141_transpose_x_0"), val = tensor(false)]; tensor x_141_transpose_y_0 = const()[name = tensor("x_141_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_13 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1324)[name = tensor("transpose_244")]; tensor x_141 = matmul(transpose_x = x_141_transpose_x_0, transpose_y = x_141_transpose_y_0, x = q_with_bias_v_13, y = var_1326)[name = tensor("x_141")]; tensor const_74 = const()[name = tensor("const_74"), val = tensor(0x0p+0)]; tensor x_143_pad_0 = const()[name = tensor("x_143_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_143_mode_0 = const()[name = tensor("x_143_mode_0"), val = tensor("constant")]; tensor x_143 = pad(constant_val = const_74, mode = x_143_mode_0, pad = x_143_pad_0, x = x_141)[name = tensor("x_143")]; tensor var_1334 = const()[name = tensor("op_1334"), val = tensor([1, 8, -1, 188])]; tensor x_145 = reshape(shape = var_1334, x = x_143)[name = tensor("x_145")]; tensor var_1338_begin_0 = const()[name = tensor("op_1338_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1338_end_0 = const()[name = tensor("op_1338_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1338_end_mask_0 = const()[name = tensor("op_1338_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1338 = slice_by_index(begin = var_1338_begin_0, end = var_1338_end_0, end_mask = var_1338_end_mask_0, x = x_145)[name = tensor("op_1338")]; tensor var_1339 = const()[name = tensor("op_1339"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_25 = reshape(shape = var_1339, x = var_1338)[name = tensor("matrix_bd_25")]; tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_85_perm_0 = const()[name = tensor("transpose_85_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_85 = transpose(perm = transpose_85_perm_0, x = k_25)[name = tensor("transpose_242")]; tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = var_1322)[name = tensor("transpose_243")]; tensor matrix_ac_13 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_84, y = transpose_85)[name = tensor("matrix_ac_13")]; tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_27 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25)[name = tensor("matrix_bd_27")]; tensor var_1348 = add(x = matrix_ac_13, y = matrix_bd_27)[name = tensor("op_1348")]; tensor _inversed_scores_25_y_0 = const()[name = tensor("_inversed_scores_25_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_25 = mul(x = var_1348, y = _inversed_scores_25_y_0)[name = tensor("_inversed_scores_25")]; tensor scores_27 = select(a = var_14, b = _inversed_scores_25, cond = mask_3)[name = tensor("scores_27")]; tensor var_1354 = softmax(axis = var_32, x = scores_27)[name = tensor("op_1354")]; tensor input_345 = select(a = var_13, b = var_1354, cond = mask_3)[name = tensor("input_345")]; tensor x_147_transpose_x_0 = const()[name = tensor("x_147_transpose_x_0"), val = tensor(false)]; tensor x_147_transpose_y_0 = const()[name = tensor("x_147_transpose_y_0"), val = tensor(false)]; tensor value_15 = transpose(perm = value_15_perm_0, x = v_13)[name = tensor("transpose_245")]; tensor x_147 = matmul(transpose_x = x_147_transpose_x_0, transpose_y = x_147_transpose_y_0, x = input_345, y = value_15)[name = tensor("x_147")]; tensor var_1358_perm_0 = const()[name = tensor("op_1358_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1359 = const()[name = tensor("op_1359"), val = tensor([1, -1, 1024])]; tensor var_1358 = transpose(perm = var_1358_perm_0, x = x_147)[name = tensor("transpose_241")]; tensor input_347 = reshape(shape = var_1359, x = var_1358)[name = tensor("input_347")]; tensor input_349 = linear(bias = encoder_layers_6_self_attn_linear_out_bias, weight = encoder_layers_6_self_attn_linear_out_weight, x = input_347)[name = tensor("linear_61")]; tensor input_351 = add(x = input_343, y = input_349)[name = tensor("input_351")]; tensor x_151_axes_0 = const()[name = tensor("x_151_axes_0"), val = tensor([-1])]; tensor x_151 = layer_norm(axes = x_151_axes_0, beta = encoder_layers_6_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_6_norm_conv_weight, x = input_351)[name = tensor("x_151")]; tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; tensor input_353 = transpose(perm = input_353_perm_0, x = x_151)[name = tensor("transpose_240")]; tensor input_355 = conv(bias = encoder_layers_6_conv_pointwise_conv1_bias, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = encoder_layers_6_conv_pointwise_conv1_weight, x = input_353)[name = tensor("input_355")]; tensor x_153_split_num_splits_0 = const()[name = tensor("x_153_split_num_splits_0"), val = tensor(2)]; tensor x_153_split_axis_0 = const()[name = tensor("x_153_split_axis_0"), val = tensor(1)]; tensor x_153_split_0, tensor x_153_split_1 = split(axis = x_153_split_axis_0, num_splits = x_153_split_num_splits_0, x = input_355)[name = tensor("x_153_split")]; tensor x_153_split_1_sigmoid = sigmoid(x = x_153_split_1)[name = tensor("x_153_split_1_sigmoid")]; tensor x_153 = mul(x = x_153_split_0, y = x_153_split_1_sigmoid)[name = tensor("x_153")]; tensor input_357 = select(a = var_13, b = x_153, cond = var_339)[name = tensor("input_357")]; tensor const_77 = const()[name = tensor("const_77"), val = tensor(0x0p+0)]; tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; tensor input_359 = pad(constant_val = const_77, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357)[name = tensor("input_359")]; tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(1024)]; tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; tensor const_260 = const()[name = tensor("const_260"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2344405184)))]; tensor const_261 = const()[name = tensor("const_261"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2344442112)))]; tensor input_363 = conv(bias = const_261, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_260, x = input_359)[name = tensor("input_363")]; tensor input_365 = silu(x = input_363)[name = tensor("input_365")]; tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; tensor x_155 = conv(bias = encoder_layers_6_conv_pointwise_conv2_bias, dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = encoder_layers_6_conv_pointwise_conv2_weight, x = input_365)[name = tensor("x_155")]; tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; tensor input_367 = transpose(perm = input_367_perm_0, x = x_155)[name = tensor("transpose_239")]; tensor input_369 = add(x = input_351, y = input_367)[name = tensor("input_369")]; tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; tensor input_371 = layer_norm(axes = input_371_axes_0, beta = encoder_layers_6_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_6_norm_feed_forward2_weight, x = input_369)[name = tensor("input_371")]; tensor input_373 = linear(bias = encoder_layers_6_feed_forward2_linear1_bias, weight = encoder_layers_6_feed_forward2_linear1_weight, x = input_371)[name = tensor("linear_62")]; tensor input_375 = silu(x = input_373)[name = tensor("input_375")]; tensor input_379 = linear(bias = encoder_layers_6_feed_forward2_linear2_bias, weight = encoder_layers_6_feed_forward2_linear2_weight, x = input_375)[name = tensor("linear_63")]; tensor var_1425 = const()[name = tensor("op_1425"), val = tensor(0x1p-1)]; tensor var_1426 = mul(x = input_379, y = var_1425)[name = tensor("op_1426")]; tensor input_381 = add(x = input_369, y = var_1426)[name = tensor("input_381")]; tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; tensor input_383 = layer_norm(axes = input_383_axes_0, beta = encoder_layers_6_norm_out_bias, epsilon = var_11, gamma = encoder_layers_6_norm_out_weight, x = input_381)[name = tensor("input_383")]; tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; tensor input_385 = layer_norm(axes = input_385_axes_0, beta = encoder_layers_7_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_7_norm_feed_forward1_weight, x = input_383)[name = tensor("input_385")]; tensor input_387 = linear(bias = encoder_layers_7_feed_forward1_linear1_bias, weight = encoder_layers_7_feed_forward1_linear1_weight, x = input_385)[name = tensor("linear_64")]; tensor input_389 = silu(x = input_387)[name = tensor("input_389")]; tensor input_393 = linear(bias = encoder_layers_7_feed_forward1_linear2_bias, weight = encoder_layers_7_feed_forward1_linear2_weight, x = input_389)[name = tensor("linear_65")]; tensor var_1456 = const()[name = tensor("op_1456"), val = tensor(0x1p-1)]; tensor var_1457 = mul(x = input_393, y = var_1456)[name = tensor("op_1457")]; tensor input_395 = add(x = input_383, y = var_1457)[name = tensor("input_395")]; tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; tensor query_15 = layer_norm(axes = query_15_axes_0, beta = encoder_layers_7_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_7_norm_self_att_weight, x = input_395)[name = tensor("query_15")]; tensor var_1473 = linear(bias = encoder_layers_7_self_attn_linear_q_bias, weight = encoder_layers_7_self_attn_linear_q_weight, x = query_15)[name = tensor("linear_66")]; tensor var_1474 = const()[name = tensor("op_1474"), val = tensor([1, -1, 8, 128])]; tensor q_43 = reshape(shape = var_1474, x = var_1473)[name = tensor("q_43")]; tensor var_1478 = linear(bias = encoder_layers_7_self_attn_linear_k_bias, weight = encoder_layers_7_self_attn_linear_k_weight, x = query_15)[name = tensor("linear_67")]; tensor var_1479 = const()[name = tensor("op_1479"), val = tensor([1, -1, 8, 128])]; tensor k_29 = reshape(shape = var_1479, x = var_1478)[name = tensor("k_29")]; tensor var_1483 = linear(bias = encoder_layers_7_self_attn_linear_v_bias, weight = encoder_layers_7_self_attn_linear_v_weight, x = query_15)[name = tensor("linear_68")]; tensor var_1484 = const()[name = tensor("op_1484"), val = tensor([1, -1, 8, 128])]; tensor v_15 = reshape(shape = var_1484, x = var_1483)[name = tensor("v_15")]; tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1496 = add(x = q_43, y = encoder_layers_7_self_attn_pos_bias_u)[name = tensor("op_1496")]; tensor var_1498 = add(x = q_43, y = encoder_layers_7_self_attn_pos_bias_v)[name = tensor("op_1498")]; tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1500 = const()[name = tensor("op_1500"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2344446272)))]; tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_15 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1498)[name = tensor("transpose_237")]; tensor x_163 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_15, y = var_1500)[name = tensor("x_163")]; tensor const_84 = const()[name = tensor("const_84"), val = tensor(0x0p+0)]; tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; tensor x_165 = pad(constant_val = const_84, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163)[name = tensor("x_165")]; tensor var_1508 = const()[name = tensor("op_1508"), val = tensor([1, 8, -1, 188])]; tensor x_167 = reshape(shape = var_1508, x = x_165)[name = tensor("x_167")]; tensor var_1512_begin_0 = const()[name = tensor("op_1512_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1512_end_0 = const()[name = tensor("op_1512_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1512_end_mask_0 = const()[name = tensor("op_1512_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1512 = slice_by_index(begin = var_1512_begin_0, end = var_1512_end_0, end_mask = var_1512_end_mask_0, x = x_167)[name = tensor("op_1512")]; tensor var_1513 = const()[name = tensor("op_1513"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_29 = reshape(shape = var_1513, x = var_1512)[name = tensor("matrix_bd_29")]; tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; tensor transpose_86_perm_0 = const()[name = tensor("transpose_86_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_87_perm_0 = const()[name = tensor("transpose_87_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_87 = transpose(perm = transpose_87_perm_0, x = k_29)[name = tensor("transpose_235")]; tensor transpose_86 = transpose(perm = transpose_86_perm_0, x = var_1496)[name = tensor("transpose_236")]; tensor matrix_ac_15 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_86, y = transpose_87)[name = tensor("matrix_ac_15")]; tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_31 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29)[name = tensor("matrix_bd_31")]; tensor var_1522 = add(x = matrix_ac_15, y = matrix_bd_31)[name = tensor("op_1522")]; tensor _inversed_scores_29_y_0 = const()[name = tensor("_inversed_scores_29_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_29 = mul(x = var_1522, y = _inversed_scores_29_y_0)[name = tensor("_inversed_scores_29")]; tensor scores_31 = select(a = var_14, b = _inversed_scores_29, cond = mask_3)[name = tensor("scores_31")]; tensor var_1528 = softmax(axis = var_32, x = scores_31)[name = tensor("op_1528")]; tensor input_397 = select(a = var_13, b = var_1528, cond = mask_3)[name = tensor("input_397")]; tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; tensor value_17 = transpose(perm = value_17_perm_0, x = v_15)[name = tensor("transpose_238")]; tensor x_169 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_397, y = value_17)[name = tensor("x_169")]; tensor var_1532_perm_0 = const()[name = tensor("op_1532_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1533 = const()[name = tensor("op_1533"), val = tensor([1, -1, 1024])]; tensor var_1532 = transpose(perm = var_1532_perm_0, x = x_169)[name = tensor("transpose_234")]; tensor input_399 = reshape(shape = var_1533, x = var_1532)[name = tensor("input_399")]; tensor input_401 = linear(bias = encoder_layers_7_self_attn_linear_out_bias, weight = encoder_layers_7_self_attn_linear_out_weight, x = input_399)[name = tensor("linear_70")]; tensor input_403 = add(x = input_395, y = input_401)[name = tensor("input_403")]; tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; tensor x_173 = layer_norm(axes = x_173_axes_0, beta = encoder_layers_7_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_7_norm_conv_weight, x = input_403)[name = tensor("x_173")]; tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; tensor input_405 = transpose(perm = input_405_perm_0, x = x_173)[name = tensor("transpose_233")]; tensor input_407 = conv(bias = encoder_layers_7_conv_pointwise_conv1_bias, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = encoder_layers_7_conv_pointwise_conv1_weight, x = input_405)[name = tensor("input_407")]; tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; tensor x_175_split_0, tensor x_175_split_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_407)[name = tensor("x_175_split")]; tensor x_175_split_1_sigmoid = sigmoid(x = x_175_split_1)[name = tensor("x_175_split_1_sigmoid")]; tensor x_175 = mul(x = x_175_split_0, y = x_175_split_1_sigmoid)[name = tensor("x_175")]; tensor input_409 = select(a = var_13, b = x_175, cond = var_339)[name = tensor("input_409")]; tensor const_87 = const()[name = tensor("const_87"), val = tensor(0x0p+0)]; tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; tensor input_411 = pad(constant_val = const_87, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409)[name = tensor("input_411")]; tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(1024)]; tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; tensor const_262 = const()[name = tensor("const_262"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2345982336)))]; tensor const_263 = const()[name = tensor("const_263"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2346019264)))]; tensor input_415 = conv(bias = const_263, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_262, x = input_411)[name = tensor("input_415")]; tensor input_417 = silu(x = input_415)[name = tensor("input_417")]; tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(1)]; tensor x_177 = conv(bias = encoder_layers_7_conv_pointwise_conv2_bias, dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = encoder_layers_7_conv_pointwise_conv2_weight, x = input_417)[name = tensor("x_177")]; tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; tensor input_419 = transpose(perm = input_419_perm_0, x = x_177)[name = tensor("transpose_232")]; tensor input_421 = add(x = input_403, y = input_419)[name = tensor("input_421")]; tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; tensor input_423 = layer_norm(axes = input_423_axes_0, beta = encoder_layers_7_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_7_norm_feed_forward2_weight, x = input_421)[name = tensor("input_423")]; tensor input_425 = linear(bias = encoder_layers_7_feed_forward2_linear1_bias, weight = encoder_layers_7_feed_forward2_linear1_weight, x = input_423)[name = tensor("linear_71")]; tensor input_427 = silu(x = input_425)[name = tensor("input_427")]; tensor input_431 = linear(bias = encoder_layers_7_feed_forward2_linear2_bias, weight = encoder_layers_7_feed_forward2_linear2_weight, x = input_427)[name = tensor("linear_72")]; tensor var_1599 = const()[name = tensor("op_1599"), val = tensor(0x1p-1)]; tensor var_1600 = mul(x = input_431, y = var_1599)[name = tensor("op_1600")]; tensor input_433 = add(x = input_421, y = var_1600)[name = tensor("input_433")]; tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; tensor input_435 = layer_norm(axes = input_435_axes_0, beta = encoder_layers_7_norm_out_bias, epsilon = var_11, gamma = encoder_layers_7_norm_out_weight, x = input_433)[name = tensor("input_435")]; tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; tensor input_437 = layer_norm(axes = input_437_axes_0, beta = encoder_layers_8_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_8_norm_feed_forward1_weight, x = input_435)[name = tensor("input_437")]; tensor input_439 = linear(bias = encoder_layers_8_feed_forward1_linear1_bias, weight = encoder_layers_8_feed_forward1_linear1_weight, x = input_437)[name = tensor("linear_73")]; tensor input_441 = silu(x = input_439)[name = tensor("input_441")]; tensor input_445 = linear(bias = encoder_layers_8_feed_forward1_linear2_bias, weight = encoder_layers_8_feed_forward1_linear2_weight, x = input_441)[name = tensor("linear_74")]; tensor var_1630 = const()[name = tensor("op_1630"), val = tensor(0x1p-1)]; tensor var_1631 = mul(x = input_445, y = var_1630)[name = tensor("op_1631")]; tensor input_447 = add(x = input_435, y = var_1631)[name = tensor("input_447")]; tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; tensor query_17 = layer_norm(axes = query_17_axes_0, beta = encoder_layers_8_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_8_norm_self_att_weight, x = input_447)[name = tensor("query_17")]; tensor var_1647 = linear(bias = encoder_layers_8_self_attn_linear_q_bias, weight = encoder_layers_8_self_attn_linear_q_weight, x = query_17)[name = tensor("linear_75")]; tensor var_1648 = const()[name = tensor("op_1648"), val = tensor([1, -1, 8, 128])]; tensor q_49 = reshape(shape = var_1648, x = var_1647)[name = tensor("q_49")]; tensor var_1652 = linear(bias = encoder_layers_8_self_attn_linear_k_bias, weight = encoder_layers_8_self_attn_linear_k_weight, x = query_17)[name = tensor("linear_76")]; tensor var_1653 = const()[name = tensor("op_1653"), val = tensor([1, -1, 8, 128])]; tensor k_33 = reshape(shape = var_1653, x = var_1652)[name = tensor("k_33")]; tensor var_1657 = linear(bias = encoder_layers_8_self_attn_linear_v_bias, weight = encoder_layers_8_self_attn_linear_v_weight, x = query_17)[name = tensor("linear_77")]; tensor var_1658 = const()[name = tensor("op_1658"), val = tensor([1, -1, 8, 128])]; tensor v_17 = reshape(shape = var_1658, x = var_1657)[name = tensor("v_17")]; tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1670 = add(x = q_49, y = encoder_layers_8_self_attn_pos_bias_u)[name = tensor("op_1670")]; tensor var_1672 = add(x = q_49, y = encoder_layers_8_self_attn_pos_bias_v)[name = tensor("op_1672")]; tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1674 = const()[name = tensor("op_1674"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2346023424)))]; tensor x_185_transpose_x_0 = const()[name = tensor("x_185_transpose_x_0"), val = tensor(false)]; tensor x_185_transpose_y_0 = const()[name = tensor("x_185_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_17 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1672)[name = tensor("transpose_230")]; tensor x_185 = matmul(transpose_x = x_185_transpose_x_0, transpose_y = x_185_transpose_y_0, x = q_with_bias_v_17, y = var_1674)[name = tensor("x_185")]; tensor const_94 = const()[name = tensor("const_94"), val = tensor(0x0p+0)]; tensor x_187_pad_0 = const()[name = tensor("x_187_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_187_mode_0 = const()[name = tensor("x_187_mode_0"), val = tensor("constant")]; tensor x_187 = pad(constant_val = const_94, mode = x_187_mode_0, pad = x_187_pad_0, x = x_185)[name = tensor("x_187")]; tensor var_1682 = const()[name = tensor("op_1682"), val = tensor([1, 8, -1, 188])]; tensor x_189 = reshape(shape = var_1682, x = x_187)[name = tensor("x_189")]; tensor var_1686_begin_0 = const()[name = tensor("op_1686_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1686_end_0 = const()[name = tensor("op_1686_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1686_end_mask_0 = const()[name = tensor("op_1686_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1686 = slice_by_index(begin = var_1686_begin_0, end = var_1686_end_0, end_mask = var_1686_end_mask_0, x = x_189)[name = tensor("op_1686")]; tensor var_1687 = const()[name = tensor("op_1687"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_33 = reshape(shape = var_1687, x = var_1686)[name = tensor("matrix_bd_33")]; tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; tensor transpose_88_perm_0 = const()[name = tensor("transpose_88_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_89_perm_0 = const()[name = tensor("transpose_89_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_89 = transpose(perm = transpose_89_perm_0, x = k_33)[name = tensor("transpose_228")]; tensor transpose_88 = transpose(perm = transpose_88_perm_0, x = var_1670)[name = tensor("transpose_229")]; tensor matrix_ac_17 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_88, y = transpose_89)[name = tensor("matrix_ac_17")]; tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_35 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33)[name = tensor("matrix_bd_35")]; tensor var_1696 = add(x = matrix_ac_17, y = matrix_bd_35)[name = tensor("op_1696")]; tensor _inversed_scores_33_y_0 = const()[name = tensor("_inversed_scores_33_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_33 = mul(x = var_1696, y = _inversed_scores_33_y_0)[name = tensor("_inversed_scores_33")]; tensor scores_35 = select(a = var_14, b = _inversed_scores_33, cond = mask_3)[name = tensor("scores_35")]; tensor var_1702 = softmax(axis = var_32, x = scores_35)[name = tensor("op_1702")]; tensor input_449 = select(a = var_13, b = var_1702, cond = mask_3)[name = tensor("input_449")]; tensor x_191_transpose_x_0 = const()[name = tensor("x_191_transpose_x_0"), val = tensor(false)]; tensor x_191_transpose_y_0 = const()[name = tensor("x_191_transpose_y_0"), val = tensor(false)]; tensor value_19 = transpose(perm = value_19_perm_0, x = v_17)[name = tensor("transpose_231")]; tensor x_191 = matmul(transpose_x = x_191_transpose_x_0, transpose_y = x_191_transpose_y_0, x = input_449, y = value_19)[name = tensor("x_191")]; tensor var_1706_perm_0 = const()[name = tensor("op_1706_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1707 = const()[name = tensor("op_1707"), val = tensor([1, -1, 1024])]; tensor var_1706 = transpose(perm = var_1706_perm_0, x = x_191)[name = tensor("transpose_227")]; tensor input_451 = reshape(shape = var_1707, x = var_1706)[name = tensor("input_451")]; tensor input_453 = linear(bias = encoder_layers_8_self_attn_linear_out_bias, weight = encoder_layers_8_self_attn_linear_out_weight, x = input_451)[name = tensor("linear_79")]; tensor input_455 = add(x = input_447, y = input_453)[name = tensor("input_455")]; tensor x_195_axes_0 = const()[name = tensor("x_195_axes_0"), val = tensor([-1])]; tensor x_195 = layer_norm(axes = x_195_axes_0, beta = encoder_layers_8_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_8_norm_conv_weight, x = input_455)[name = tensor("x_195")]; tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; tensor input_457 = transpose(perm = input_457_perm_0, x = x_195)[name = tensor("transpose_226")]; tensor input_459 = conv(bias = encoder_layers_8_conv_pointwise_conv1_bias, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = encoder_layers_8_conv_pointwise_conv1_weight, x = input_457)[name = tensor("input_459")]; tensor x_197_split_num_splits_0 = const()[name = tensor("x_197_split_num_splits_0"), val = tensor(2)]; tensor x_197_split_axis_0 = const()[name = tensor("x_197_split_axis_0"), val = tensor(1)]; tensor x_197_split_0, tensor x_197_split_1 = split(axis = x_197_split_axis_0, num_splits = x_197_split_num_splits_0, x = input_459)[name = tensor("x_197_split")]; tensor x_197_split_1_sigmoid = sigmoid(x = x_197_split_1)[name = tensor("x_197_split_1_sigmoid")]; tensor x_197 = mul(x = x_197_split_0, y = x_197_split_1_sigmoid)[name = tensor("x_197")]; tensor input_461 = select(a = var_13, b = x_197, cond = var_339)[name = tensor("input_461")]; tensor const_97 = const()[name = tensor("const_97"), val = tensor(0x0p+0)]; tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; tensor input_463 = pad(constant_val = const_97, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461)[name = tensor("input_463")]; tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(1024)]; tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; tensor const_264 = const()[name = tensor("const_264"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2347559488)))]; tensor const_265 = const()[name = tensor("const_265"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2347596416)))]; tensor input_467 = conv(bias = const_265, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_264, x = input_463)[name = tensor("input_467")]; tensor input_469 = silu(x = input_467)[name = tensor("input_469")]; tensor x_199_pad_type_0 = const()[name = tensor("x_199_pad_type_0"), val = tensor("valid")]; tensor x_199_strides_0 = const()[name = tensor("x_199_strides_0"), val = tensor([1])]; tensor x_199_pad_0 = const()[name = tensor("x_199_pad_0"), val = tensor([0, 0])]; tensor x_199_dilations_0 = const()[name = tensor("x_199_dilations_0"), val = tensor([1])]; tensor x_199_groups_0 = const()[name = tensor("x_199_groups_0"), val = tensor(1)]; tensor x_199 = conv(bias = encoder_layers_8_conv_pointwise_conv2_bias, dilations = x_199_dilations_0, groups = x_199_groups_0, pad = x_199_pad_0, pad_type = x_199_pad_type_0, strides = x_199_strides_0, weight = encoder_layers_8_conv_pointwise_conv2_weight, x = input_469)[name = tensor("x_199")]; tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; tensor input_471 = transpose(perm = input_471_perm_0, x = x_199)[name = tensor("transpose_225")]; tensor input_473 = add(x = input_455, y = input_471)[name = tensor("input_473")]; tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; tensor input_475 = layer_norm(axes = input_475_axes_0, beta = encoder_layers_8_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_8_norm_feed_forward2_weight, x = input_473)[name = tensor("input_475")]; tensor input_477 = linear(bias = encoder_layers_8_feed_forward2_linear1_bias, weight = encoder_layers_8_feed_forward2_linear1_weight, x = input_475)[name = tensor("linear_80")]; tensor input_479 = silu(x = input_477)[name = tensor("input_479")]; tensor input_483 = linear(bias = encoder_layers_8_feed_forward2_linear2_bias, weight = encoder_layers_8_feed_forward2_linear2_weight, x = input_479)[name = tensor("linear_81")]; tensor var_1773 = const()[name = tensor("op_1773"), val = tensor(0x1p-1)]; tensor var_1774 = mul(x = input_483, y = var_1773)[name = tensor("op_1774")]; tensor input_485 = add(x = input_473, y = var_1774)[name = tensor("input_485")]; tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; tensor input_487 = layer_norm(axes = input_487_axes_0, beta = encoder_layers_8_norm_out_bias, epsilon = var_11, gamma = encoder_layers_8_norm_out_weight, x = input_485)[name = tensor("input_487")]; tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; tensor input_489 = layer_norm(axes = input_489_axes_0, beta = encoder_layers_9_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_9_norm_feed_forward1_weight, x = input_487)[name = tensor("input_489")]; tensor input_491 = linear(bias = encoder_layers_9_feed_forward1_linear1_bias, weight = encoder_layers_9_feed_forward1_linear1_weight, x = input_489)[name = tensor("linear_82")]; tensor input_493 = silu(x = input_491)[name = tensor("input_493")]; tensor input_497 = linear(bias = encoder_layers_9_feed_forward1_linear2_bias, weight = encoder_layers_9_feed_forward1_linear2_weight, x = input_493)[name = tensor("linear_83")]; tensor var_1804 = const()[name = tensor("op_1804"), val = tensor(0x1p-1)]; tensor var_1805 = mul(x = input_497, y = var_1804)[name = tensor("op_1805")]; tensor input_499 = add(x = input_487, y = var_1805)[name = tensor("input_499")]; tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; tensor query_19 = layer_norm(axes = query_19_axes_0, beta = encoder_layers_9_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_9_norm_self_att_weight, x = input_499)[name = tensor("query_19")]; tensor var_1821 = linear(bias = encoder_layers_9_self_attn_linear_q_bias, weight = encoder_layers_9_self_attn_linear_q_weight, x = query_19)[name = tensor("linear_84")]; tensor var_1822 = const()[name = tensor("op_1822"), val = tensor([1, -1, 8, 128])]; tensor q_55 = reshape(shape = var_1822, x = var_1821)[name = tensor("q_55")]; tensor var_1826 = linear(bias = encoder_layers_9_self_attn_linear_k_bias, weight = encoder_layers_9_self_attn_linear_k_weight, x = query_19)[name = tensor("linear_85")]; tensor var_1827 = const()[name = tensor("op_1827"), val = tensor([1, -1, 8, 128])]; tensor k_37 = reshape(shape = var_1827, x = var_1826)[name = tensor("k_37")]; tensor var_1831 = linear(bias = encoder_layers_9_self_attn_linear_v_bias, weight = encoder_layers_9_self_attn_linear_v_weight, x = query_19)[name = tensor("linear_86")]; tensor var_1832 = const()[name = tensor("op_1832"), val = tensor([1, -1, 8, 128])]; tensor v_19 = reshape(shape = var_1832, x = var_1831)[name = tensor("v_19")]; tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1844 = add(x = q_55, y = encoder_layers_9_self_attn_pos_bias_u)[name = tensor("op_1844")]; tensor var_1846 = add(x = q_55, y = encoder_layers_9_self_attn_pos_bias_v)[name = tensor("op_1846")]; tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1848 = const()[name = tensor("op_1848"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2347600576)))]; tensor x_207_transpose_x_0 = const()[name = tensor("x_207_transpose_x_0"), val = tensor(false)]; tensor x_207_transpose_y_0 = const()[name = tensor("x_207_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_19 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1846)[name = tensor("transpose_223")]; tensor x_207 = matmul(transpose_x = x_207_transpose_x_0, transpose_y = x_207_transpose_y_0, x = q_with_bias_v_19, y = var_1848)[name = tensor("x_207")]; tensor const_104 = const()[name = tensor("const_104"), val = tensor(0x0p+0)]; tensor x_209_pad_0 = const()[name = tensor("x_209_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_209_mode_0 = const()[name = tensor("x_209_mode_0"), val = tensor("constant")]; tensor x_209 = pad(constant_val = const_104, mode = x_209_mode_0, pad = x_209_pad_0, x = x_207)[name = tensor("x_209")]; tensor var_1856 = const()[name = tensor("op_1856"), val = tensor([1, 8, -1, 188])]; tensor x_211 = reshape(shape = var_1856, x = x_209)[name = tensor("x_211")]; tensor var_1860_begin_0 = const()[name = tensor("op_1860_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1860_end_0 = const()[name = tensor("op_1860_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1860_end_mask_0 = const()[name = tensor("op_1860_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1860 = slice_by_index(begin = var_1860_begin_0, end = var_1860_end_0, end_mask = var_1860_end_mask_0, x = x_211)[name = tensor("op_1860")]; tensor var_1861 = const()[name = tensor("op_1861"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_37 = reshape(shape = var_1861, x = var_1860)[name = tensor("matrix_bd_37")]; tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; tensor transpose_90_perm_0 = const()[name = tensor("transpose_90_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_91_perm_0 = const()[name = tensor("transpose_91_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_91 = transpose(perm = transpose_91_perm_0, x = k_37)[name = tensor("transpose_221")]; tensor transpose_90 = transpose(perm = transpose_90_perm_0, x = var_1844)[name = tensor("transpose_222")]; tensor matrix_ac_19 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_90, y = transpose_91)[name = tensor("matrix_ac_19")]; tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_39 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37)[name = tensor("matrix_bd_39")]; tensor var_1870 = add(x = matrix_ac_19, y = matrix_bd_39)[name = tensor("op_1870")]; tensor _inversed_scores_37_y_0 = const()[name = tensor("_inversed_scores_37_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_37 = mul(x = var_1870, y = _inversed_scores_37_y_0)[name = tensor("_inversed_scores_37")]; tensor scores_39 = select(a = var_14, b = _inversed_scores_37, cond = mask_3)[name = tensor("scores_39")]; tensor var_1876 = softmax(axis = var_32, x = scores_39)[name = tensor("op_1876")]; tensor input_501 = select(a = var_13, b = var_1876, cond = mask_3)[name = tensor("input_501")]; tensor x_213_transpose_x_0 = const()[name = tensor("x_213_transpose_x_0"), val = tensor(false)]; tensor x_213_transpose_y_0 = const()[name = tensor("x_213_transpose_y_0"), val = tensor(false)]; tensor value_21 = transpose(perm = value_21_perm_0, x = v_19)[name = tensor("transpose_224")]; tensor x_213 = matmul(transpose_x = x_213_transpose_x_0, transpose_y = x_213_transpose_y_0, x = input_501, y = value_21)[name = tensor("x_213")]; tensor var_1880_perm_0 = const()[name = tensor("op_1880_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1881 = const()[name = tensor("op_1881"), val = tensor([1, -1, 1024])]; tensor var_1880 = transpose(perm = var_1880_perm_0, x = x_213)[name = tensor("transpose_220")]; tensor input_503 = reshape(shape = var_1881, x = var_1880)[name = tensor("input_503")]; tensor input_505 = linear(bias = encoder_layers_9_self_attn_linear_out_bias, weight = encoder_layers_9_self_attn_linear_out_weight, x = input_503)[name = tensor("linear_88")]; tensor input_507 = add(x = input_499, y = input_505)[name = tensor("input_507")]; tensor x_217_axes_0 = const()[name = tensor("x_217_axes_0"), val = tensor([-1])]; tensor x_217 = layer_norm(axes = x_217_axes_0, beta = encoder_layers_9_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_9_norm_conv_weight, x = input_507)[name = tensor("x_217")]; tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; tensor input_509 = transpose(perm = input_509_perm_0, x = x_217)[name = tensor("transpose_219")]; tensor input_511 = conv(bias = encoder_layers_9_conv_pointwise_conv1_bias, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = encoder_layers_9_conv_pointwise_conv1_weight, x = input_509)[name = tensor("input_511")]; tensor x_219_split_num_splits_0 = const()[name = tensor("x_219_split_num_splits_0"), val = tensor(2)]; tensor x_219_split_axis_0 = const()[name = tensor("x_219_split_axis_0"), val = tensor(1)]; tensor x_219_split_0, tensor x_219_split_1 = split(axis = x_219_split_axis_0, num_splits = x_219_split_num_splits_0, x = input_511)[name = tensor("x_219_split")]; tensor x_219_split_1_sigmoid = sigmoid(x = x_219_split_1)[name = tensor("x_219_split_1_sigmoid")]; tensor x_219 = mul(x = x_219_split_0, y = x_219_split_1_sigmoid)[name = tensor("x_219")]; tensor input_513 = select(a = var_13, b = x_219, cond = var_339)[name = tensor("input_513")]; tensor const_107 = const()[name = tensor("const_107"), val = tensor(0x0p+0)]; tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; tensor input_515 = pad(constant_val = const_107, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513)[name = tensor("input_515")]; tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(1024)]; tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; tensor const_266 = const()[name = tensor("const_266"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2349136640)))]; tensor const_267 = const()[name = tensor("const_267"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2349173568)))]; tensor input_519 = conv(bias = const_267, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_266, x = input_515)[name = tensor("input_519")]; tensor input_521 = silu(x = input_519)[name = tensor("input_521")]; tensor x_221_pad_type_0 = const()[name = tensor("x_221_pad_type_0"), val = tensor("valid")]; tensor x_221_strides_0 = const()[name = tensor("x_221_strides_0"), val = tensor([1])]; tensor x_221_pad_0 = const()[name = tensor("x_221_pad_0"), val = tensor([0, 0])]; tensor x_221_dilations_0 = const()[name = tensor("x_221_dilations_0"), val = tensor([1])]; tensor x_221_groups_0 = const()[name = tensor("x_221_groups_0"), val = tensor(1)]; tensor x_221 = conv(bias = encoder_layers_9_conv_pointwise_conv2_bias, dilations = x_221_dilations_0, groups = x_221_groups_0, pad = x_221_pad_0, pad_type = x_221_pad_type_0, strides = x_221_strides_0, weight = encoder_layers_9_conv_pointwise_conv2_weight, x = input_521)[name = tensor("x_221")]; tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; tensor input_523 = transpose(perm = input_523_perm_0, x = x_221)[name = tensor("transpose_218")]; tensor input_525 = add(x = input_507, y = input_523)[name = tensor("input_525")]; tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; tensor input_527 = layer_norm(axes = input_527_axes_0, beta = encoder_layers_9_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_9_norm_feed_forward2_weight, x = input_525)[name = tensor("input_527")]; tensor input_529 = linear(bias = encoder_layers_9_feed_forward2_linear1_bias, weight = encoder_layers_9_feed_forward2_linear1_weight, x = input_527)[name = tensor("linear_89")]; tensor input_531 = silu(x = input_529)[name = tensor("input_531")]; tensor input_535 = linear(bias = encoder_layers_9_feed_forward2_linear2_bias, weight = encoder_layers_9_feed_forward2_linear2_weight, x = input_531)[name = tensor("linear_90")]; tensor var_1947 = const()[name = tensor("op_1947"), val = tensor(0x1p-1)]; tensor var_1948 = mul(x = input_535, y = var_1947)[name = tensor("op_1948")]; tensor input_537 = add(x = input_525, y = var_1948)[name = tensor("input_537")]; tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; tensor input_539 = layer_norm(axes = input_539_axes_0, beta = encoder_layers_9_norm_out_bias, epsilon = var_11, gamma = encoder_layers_9_norm_out_weight, x = input_537)[name = tensor("input_539")]; tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; tensor input_541 = layer_norm(axes = input_541_axes_0, beta = encoder_layers_10_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_10_norm_feed_forward1_weight, x = input_539)[name = tensor("input_541")]; tensor input_543 = linear(bias = encoder_layers_10_feed_forward1_linear1_bias, weight = encoder_layers_10_feed_forward1_linear1_weight, x = input_541)[name = tensor("linear_91")]; tensor input_545 = silu(x = input_543)[name = tensor("input_545")]; tensor input_549 = linear(bias = encoder_layers_10_feed_forward1_linear2_bias, weight = encoder_layers_10_feed_forward1_linear2_weight, x = input_545)[name = tensor("linear_92")]; tensor var_1978 = const()[name = tensor("op_1978"), val = tensor(0x1p-1)]; tensor var_1979 = mul(x = input_549, y = var_1978)[name = tensor("op_1979")]; tensor input_551 = add(x = input_539, y = var_1979)[name = tensor("input_551")]; tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; tensor query_21 = layer_norm(axes = query_21_axes_0, beta = encoder_layers_10_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_10_norm_self_att_weight, x = input_551)[name = tensor("query_21")]; tensor var_1995 = linear(bias = encoder_layers_10_self_attn_linear_q_bias, weight = encoder_layers_10_self_attn_linear_q_weight, x = query_21)[name = tensor("linear_93")]; tensor var_1996 = const()[name = tensor("op_1996"), val = tensor([1, -1, 8, 128])]; tensor q_61 = reshape(shape = var_1996, x = var_1995)[name = tensor("q_61")]; tensor var_2000 = linear(bias = encoder_layers_10_self_attn_linear_k_bias, weight = encoder_layers_10_self_attn_linear_k_weight, x = query_21)[name = tensor("linear_94")]; tensor var_2001 = const()[name = tensor("op_2001"), val = tensor([1, -1, 8, 128])]; tensor k_41 = reshape(shape = var_2001, x = var_2000)[name = tensor("k_41")]; tensor var_2005 = linear(bias = encoder_layers_10_self_attn_linear_v_bias, weight = encoder_layers_10_self_attn_linear_v_weight, x = query_21)[name = tensor("linear_95")]; tensor var_2006 = const()[name = tensor("op_2006"), val = tensor([1, -1, 8, 128])]; tensor v_21 = reshape(shape = var_2006, x = var_2005)[name = tensor("v_21")]; tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2018 = add(x = q_61, y = encoder_layers_10_self_attn_pos_bias_u)[name = tensor("op_2018")]; tensor var_2020 = add(x = q_61, y = encoder_layers_10_self_attn_pos_bias_v)[name = tensor("op_2020")]; tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2022 = const()[name = tensor("op_2022"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2349177728)))]; tensor x_229_transpose_x_0 = const()[name = tensor("x_229_transpose_x_0"), val = tensor(false)]; tensor x_229_transpose_y_0 = const()[name = tensor("x_229_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_21 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2020)[name = tensor("transpose_216")]; tensor x_229 = matmul(transpose_x = x_229_transpose_x_0, transpose_y = x_229_transpose_y_0, x = q_with_bias_v_21, y = var_2022)[name = tensor("x_229")]; tensor const_114 = const()[name = tensor("const_114"), val = tensor(0x0p+0)]; tensor x_231_pad_0 = const()[name = tensor("x_231_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_231_mode_0 = const()[name = tensor("x_231_mode_0"), val = tensor("constant")]; tensor x_231 = pad(constant_val = const_114, mode = x_231_mode_0, pad = x_231_pad_0, x = x_229)[name = tensor("x_231")]; tensor var_2030 = const()[name = tensor("op_2030"), val = tensor([1, 8, -1, 188])]; tensor x_233 = reshape(shape = var_2030, x = x_231)[name = tensor("x_233")]; tensor var_2034_begin_0 = const()[name = tensor("op_2034_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2034_end_0 = const()[name = tensor("op_2034_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2034_end_mask_0 = const()[name = tensor("op_2034_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2034 = slice_by_index(begin = var_2034_begin_0, end = var_2034_end_0, end_mask = var_2034_end_mask_0, x = x_233)[name = tensor("op_2034")]; tensor var_2035 = const()[name = tensor("op_2035"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_41 = reshape(shape = var_2035, x = var_2034)[name = tensor("matrix_bd_41")]; tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; tensor transpose_92_perm_0 = const()[name = tensor("transpose_92_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_93_perm_0 = const()[name = tensor("transpose_93_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_93 = transpose(perm = transpose_93_perm_0, x = k_41)[name = tensor("transpose_214")]; tensor transpose_92 = transpose(perm = transpose_92_perm_0, x = var_2018)[name = tensor("transpose_215")]; tensor matrix_ac_21 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_92, y = transpose_93)[name = tensor("matrix_ac_21")]; tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_43 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41)[name = tensor("matrix_bd_43")]; tensor var_2044 = add(x = matrix_ac_21, y = matrix_bd_43)[name = tensor("op_2044")]; tensor _inversed_scores_41_y_0 = const()[name = tensor("_inversed_scores_41_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_41 = mul(x = var_2044, y = _inversed_scores_41_y_0)[name = tensor("_inversed_scores_41")]; tensor scores_43 = select(a = var_14, b = _inversed_scores_41, cond = mask_3)[name = tensor("scores_43")]; tensor var_2050 = softmax(axis = var_32, x = scores_43)[name = tensor("op_2050")]; tensor input_553 = select(a = var_13, b = var_2050, cond = mask_3)[name = tensor("input_553")]; tensor x_235_transpose_x_0 = const()[name = tensor("x_235_transpose_x_0"), val = tensor(false)]; tensor x_235_transpose_y_0 = const()[name = tensor("x_235_transpose_y_0"), val = tensor(false)]; tensor value_23 = transpose(perm = value_23_perm_0, x = v_21)[name = tensor("transpose_217")]; tensor x_235 = matmul(transpose_x = x_235_transpose_x_0, transpose_y = x_235_transpose_y_0, x = input_553, y = value_23)[name = tensor("x_235")]; tensor var_2054_perm_0 = const()[name = tensor("op_2054_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2055 = const()[name = tensor("op_2055"), val = tensor([1, -1, 1024])]; tensor var_2054 = transpose(perm = var_2054_perm_0, x = x_235)[name = tensor("transpose_213")]; tensor input_555 = reshape(shape = var_2055, x = var_2054)[name = tensor("input_555")]; tensor input_557 = linear(bias = encoder_layers_10_self_attn_linear_out_bias, weight = encoder_layers_10_self_attn_linear_out_weight, x = input_555)[name = tensor("linear_97")]; tensor input_559 = add(x = input_551, y = input_557)[name = tensor("input_559")]; tensor x_239_axes_0 = const()[name = tensor("x_239_axes_0"), val = tensor([-1])]; tensor x_239 = layer_norm(axes = x_239_axes_0, beta = encoder_layers_10_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_10_norm_conv_weight, x = input_559)[name = tensor("x_239")]; tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; tensor input_561 = transpose(perm = input_561_perm_0, x = x_239)[name = tensor("transpose_212")]; tensor input_563 = conv(bias = encoder_layers_10_conv_pointwise_conv1_bias, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = encoder_layers_10_conv_pointwise_conv1_weight, x = input_561)[name = tensor("input_563")]; tensor x_241_split_num_splits_0 = const()[name = tensor("x_241_split_num_splits_0"), val = tensor(2)]; tensor x_241_split_axis_0 = const()[name = tensor("x_241_split_axis_0"), val = tensor(1)]; tensor x_241_split_0, tensor x_241_split_1 = split(axis = x_241_split_axis_0, num_splits = x_241_split_num_splits_0, x = input_563)[name = tensor("x_241_split")]; tensor x_241_split_1_sigmoid = sigmoid(x = x_241_split_1)[name = tensor("x_241_split_1_sigmoid")]; tensor x_241 = mul(x = x_241_split_0, y = x_241_split_1_sigmoid)[name = tensor("x_241")]; tensor input_565 = select(a = var_13, b = x_241, cond = var_339)[name = tensor("input_565")]; tensor const_117 = const()[name = tensor("const_117"), val = tensor(0x0p+0)]; tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; tensor input_567 = pad(constant_val = const_117, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565)[name = tensor("input_567")]; tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(1024)]; tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; tensor const_268 = const()[name = tensor("const_268"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2350713792)))]; tensor const_269 = const()[name = tensor("const_269"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2350750720)))]; tensor input_571 = conv(bias = const_269, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_268, x = input_567)[name = tensor("input_571")]; tensor input_573 = silu(x = input_571)[name = tensor("input_573")]; tensor x_243_pad_type_0 = const()[name = tensor("x_243_pad_type_0"), val = tensor("valid")]; tensor x_243_strides_0 = const()[name = tensor("x_243_strides_0"), val = tensor([1])]; tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0])]; tensor x_243_dilations_0 = const()[name = tensor("x_243_dilations_0"), val = tensor([1])]; tensor x_243_groups_0 = const()[name = tensor("x_243_groups_0"), val = tensor(1)]; tensor x_243 = conv(bias = encoder_layers_10_conv_pointwise_conv2_bias, dilations = x_243_dilations_0, groups = x_243_groups_0, pad = x_243_pad_0, pad_type = x_243_pad_type_0, strides = x_243_strides_0, weight = encoder_layers_10_conv_pointwise_conv2_weight, x = input_573)[name = tensor("x_243")]; tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; tensor input_575 = transpose(perm = input_575_perm_0, x = x_243)[name = tensor("transpose_211")]; tensor input_577 = add(x = input_559, y = input_575)[name = tensor("input_577")]; tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; tensor input_579 = layer_norm(axes = input_579_axes_0, beta = encoder_layers_10_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_10_norm_feed_forward2_weight, x = input_577)[name = tensor("input_579")]; tensor input_581 = linear(bias = encoder_layers_10_feed_forward2_linear1_bias, weight = encoder_layers_10_feed_forward2_linear1_weight, x = input_579)[name = tensor("linear_98")]; tensor input_583 = silu(x = input_581)[name = tensor("input_583")]; tensor input_587 = linear(bias = encoder_layers_10_feed_forward2_linear2_bias, weight = encoder_layers_10_feed_forward2_linear2_weight, x = input_583)[name = tensor("linear_99")]; tensor var_2121 = const()[name = tensor("op_2121"), val = tensor(0x1p-1)]; tensor var_2122 = mul(x = input_587, y = var_2121)[name = tensor("op_2122")]; tensor input_589 = add(x = input_577, y = var_2122)[name = tensor("input_589")]; tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; tensor input_591 = layer_norm(axes = input_591_axes_0, beta = encoder_layers_10_norm_out_bias, epsilon = var_11, gamma = encoder_layers_10_norm_out_weight, x = input_589)[name = tensor("input_591")]; tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; tensor input_593 = layer_norm(axes = input_593_axes_0, beta = encoder_layers_11_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_11_norm_feed_forward1_weight, x = input_591)[name = tensor("input_593")]; tensor input_595 = linear(bias = encoder_layers_11_feed_forward1_linear1_bias, weight = encoder_layers_11_feed_forward1_linear1_weight, x = input_593)[name = tensor("linear_100")]; tensor input_597 = silu(x = input_595)[name = tensor("input_597")]; tensor input_601 = linear(bias = encoder_layers_11_feed_forward1_linear2_bias, weight = encoder_layers_11_feed_forward1_linear2_weight, x = input_597)[name = tensor("linear_101")]; tensor var_2152 = const()[name = tensor("op_2152"), val = tensor(0x1p-1)]; tensor var_2153 = mul(x = input_601, y = var_2152)[name = tensor("op_2153")]; tensor input_603 = add(x = input_591, y = var_2153)[name = tensor("input_603")]; tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; tensor query_23 = layer_norm(axes = query_23_axes_0, beta = encoder_layers_11_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_11_norm_self_att_weight, x = input_603)[name = tensor("query_23")]; tensor var_2169 = linear(bias = encoder_layers_11_self_attn_linear_q_bias, weight = encoder_layers_11_self_attn_linear_q_weight, x = query_23)[name = tensor("linear_102")]; tensor var_2170 = const()[name = tensor("op_2170"), val = tensor([1, -1, 8, 128])]; tensor q_67 = reshape(shape = var_2170, x = var_2169)[name = tensor("q_67")]; tensor var_2174 = linear(bias = encoder_layers_11_self_attn_linear_k_bias, weight = encoder_layers_11_self_attn_linear_k_weight, x = query_23)[name = tensor("linear_103")]; tensor var_2175 = const()[name = tensor("op_2175"), val = tensor([1, -1, 8, 128])]; tensor k_45 = reshape(shape = var_2175, x = var_2174)[name = tensor("k_45")]; tensor var_2179 = linear(bias = encoder_layers_11_self_attn_linear_v_bias, weight = encoder_layers_11_self_attn_linear_v_weight, x = query_23)[name = tensor("linear_104")]; tensor var_2180 = const()[name = tensor("op_2180"), val = tensor([1, -1, 8, 128])]; tensor v_23 = reshape(shape = var_2180, x = var_2179)[name = tensor("v_23")]; tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2192 = add(x = q_67, y = encoder_layers_11_self_attn_pos_bias_u)[name = tensor("op_2192")]; tensor var_2194 = add(x = q_67, y = encoder_layers_11_self_attn_pos_bias_v)[name = tensor("op_2194")]; tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2196 = const()[name = tensor("op_2196"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2350754880)))]; tensor x_251_transpose_x_0 = const()[name = tensor("x_251_transpose_x_0"), val = tensor(false)]; tensor x_251_transpose_y_0 = const()[name = tensor("x_251_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_23 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2194)[name = tensor("transpose_209")]; tensor x_251 = matmul(transpose_x = x_251_transpose_x_0, transpose_y = x_251_transpose_y_0, x = q_with_bias_v_23, y = var_2196)[name = tensor("x_251")]; tensor const_124 = const()[name = tensor("const_124"), val = tensor(0x0p+0)]; tensor x_253_pad_0 = const()[name = tensor("x_253_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_253_mode_0 = const()[name = tensor("x_253_mode_0"), val = tensor("constant")]; tensor x_253 = pad(constant_val = const_124, mode = x_253_mode_0, pad = x_253_pad_0, x = x_251)[name = tensor("x_253")]; tensor var_2204 = const()[name = tensor("op_2204"), val = tensor([1, 8, -1, 188])]; tensor x_255 = reshape(shape = var_2204, x = x_253)[name = tensor("x_255")]; tensor var_2208_begin_0 = const()[name = tensor("op_2208_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2208_end_0 = const()[name = tensor("op_2208_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2208_end_mask_0 = const()[name = tensor("op_2208_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2208 = slice_by_index(begin = var_2208_begin_0, end = var_2208_end_0, end_mask = var_2208_end_mask_0, x = x_255)[name = tensor("op_2208")]; tensor var_2209 = const()[name = tensor("op_2209"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_45 = reshape(shape = var_2209, x = var_2208)[name = tensor("matrix_bd_45")]; tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; tensor transpose_94_perm_0 = const()[name = tensor("transpose_94_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_95_perm_0 = const()[name = tensor("transpose_95_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_95 = transpose(perm = transpose_95_perm_0, x = k_45)[name = tensor("transpose_207")]; tensor transpose_94 = transpose(perm = transpose_94_perm_0, x = var_2192)[name = tensor("transpose_208")]; tensor matrix_ac_23 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_94, y = transpose_95)[name = tensor("matrix_ac_23")]; tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_47 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45)[name = tensor("matrix_bd_47")]; tensor var_2218 = add(x = matrix_ac_23, y = matrix_bd_47)[name = tensor("op_2218")]; tensor _inversed_scores_45_y_0 = const()[name = tensor("_inversed_scores_45_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_45 = mul(x = var_2218, y = _inversed_scores_45_y_0)[name = tensor("_inversed_scores_45")]; tensor scores_47 = select(a = var_14, b = _inversed_scores_45, cond = mask_3)[name = tensor("scores_47")]; tensor var_2224 = softmax(axis = var_32, x = scores_47)[name = tensor("op_2224")]; tensor input_605 = select(a = var_13, b = var_2224, cond = mask_3)[name = tensor("input_605")]; tensor x_257_transpose_x_0 = const()[name = tensor("x_257_transpose_x_0"), val = tensor(false)]; tensor x_257_transpose_y_0 = const()[name = tensor("x_257_transpose_y_0"), val = tensor(false)]; tensor value_25 = transpose(perm = value_25_perm_0, x = v_23)[name = tensor("transpose_210")]; tensor x_257 = matmul(transpose_x = x_257_transpose_x_0, transpose_y = x_257_transpose_y_0, x = input_605, y = value_25)[name = tensor("x_257")]; tensor var_2228_perm_0 = const()[name = tensor("op_2228_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2229 = const()[name = tensor("op_2229"), val = tensor([1, -1, 1024])]; tensor var_2228 = transpose(perm = var_2228_perm_0, x = x_257)[name = tensor("transpose_206")]; tensor input_607 = reshape(shape = var_2229, x = var_2228)[name = tensor("input_607")]; tensor input_609 = linear(bias = encoder_layers_11_self_attn_linear_out_bias, weight = encoder_layers_11_self_attn_linear_out_weight, x = input_607)[name = tensor("linear_106")]; tensor input_611 = add(x = input_603, y = input_609)[name = tensor("input_611")]; tensor x_261_axes_0 = const()[name = tensor("x_261_axes_0"), val = tensor([-1])]; tensor x_261 = layer_norm(axes = x_261_axes_0, beta = encoder_layers_11_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_11_norm_conv_weight, x = input_611)[name = tensor("x_261")]; tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; tensor input_613 = transpose(perm = input_613_perm_0, x = x_261)[name = tensor("transpose_205")]; tensor input_615 = conv(bias = encoder_layers_11_conv_pointwise_conv1_bias, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = encoder_layers_11_conv_pointwise_conv1_weight, x = input_613)[name = tensor("input_615")]; tensor x_263_split_num_splits_0 = const()[name = tensor("x_263_split_num_splits_0"), val = tensor(2)]; tensor x_263_split_axis_0 = const()[name = tensor("x_263_split_axis_0"), val = tensor(1)]; tensor x_263_split_0, tensor x_263_split_1 = split(axis = x_263_split_axis_0, num_splits = x_263_split_num_splits_0, x = input_615)[name = tensor("x_263_split")]; tensor x_263_split_1_sigmoid = sigmoid(x = x_263_split_1)[name = tensor("x_263_split_1_sigmoid")]; tensor x_263 = mul(x = x_263_split_0, y = x_263_split_1_sigmoid)[name = tensor("x_263")]; tensor input_617 = select(a = var_13, b = x_263, cond = var_339)[name = tensor("input_617")]; tensor const_127 = const()[name = tensor("const_127"), val = tensor(0x0p+0)]; tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; tensor input_619 = pad(constant_val = const_127, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617)[name = tensor("input_619")]; tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(1024)]; tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; tensor const_270 = const()[name = tensor("const_270"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2352290944)))]; tensor const_271 = const()[name = tensor("const_271"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2352327872)))]; tensor input_623 = conv(bias = const_271, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_270, x = input_619)[name = tensor("input_623")]; tensor input_625 = silu(x = input_623)[name = tensor("input_625")]; tensor x_265_pad_type_0 = const()[name = tensor("x_265_pad_type_0"), val = tensor("valid")]; tensor x_265_strides_0 = const()[name = tensor("x_265_strides_0"), val = tensor([1])]; tensor x_265_pad_0 = const()[name = tensor("x_265_pad_0"), val = tensor([0, 0])]; tensor x_265_dilations_0 = const()[name = tensor("x_265_dilations_0"), val = tensor([1])]; tensor x_265_groups_0 = const()[name = tensor("x_265_groups_0"), val = tensor(1)]; tensor x_265 = conv(bias = encoder_layers_11_conv_pointwise_conv2_bias, dilations = x_265_dilations_0, groups = x_265_groups_0, pad = x_265_pad_0, pad_type = x_265_pad_type_0, strides = x_265_strides_0, weight = encoder_layers_11_conv_pointwise_conv2_weight, x = input_625)[name = tensor("x_265")]; tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; tensor input_627 = transpose(perm = input_627_perm_0, x = x_265)[name = tensor("transpose_204")]; tensor input_629 = add(x = input_611, y = input_627)[name = tensor("input_629")]; tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; tensor input_631 = layer_norm(axes = input_631_axes_0, beta = encoder_layers_11_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_11_norm_feed_forward2_weight, x = input_629)[name = tensor("input_631")]; tensor input_633 = linear(bias = encoder_layers_11_feed_forward2_linear1_bias, weight = encoder_layers_11_feed_forward2_linear1_weight, x = input_631)[name = tensor("linear_107")]; tensor input_635 = silu(x = input_633)[name = tensor("input_635")]; tensor input_639 = linear(bias = encoder_layers_11_feed_forward2_linear2_bias, weight = encoder_layers_11_feed_forward2_linear2_weight, x = input_635)[name = tensor("linear_108")]; tensor var_2295 = const()[name = tensor("op_2295"), val = tensor(0x1p-1)]; tensor var_2296 = mul(x = input_639, y = var_2295)[name = tensor("op_2296")]; tensor input_641 = add(x = input_629, y = var_2296)[name = tensor("input_641")]; tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; tensor input_643 = layer_norm(axes = input_643_axes_0, beta = encoder_layers_11_norm_out_bias, epsilon = var_11, gamma = encoder_layers_11_norm_out_weight, x = input_641)[name = tensor("input_643")]; tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; tensor input_645 = layer_norm(axes = input_645_axes_0, beta = encoder_layers_12_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_12_norm_feed_forward1_weight, x = input_643)[name = tensor("input_645")]; tensor input_647 = linear(bias = encoder_layers_12_feed_forward1_linear1_bias, weight = encoder_layers_12_feed_forward1_linear1_weight, x = input_645)[name = tensor("linear_109")]; tensor input_649 = silu(x = input_647)[name = tensor("input_649")]; tensor input_653 = linear(bias = encoder_layers_12_feed_forward1_linear2_bias, weight = encoder_layers_12_feed_forward1_linear2_weight, x = input_649)[name = tensor("linear_110")]; tensor var_2326 = const()[name = tensor("op_2326"), val = tensor(0x1p-1)]; tensor var_2327 = mul(x = input_653, y = var_2326)[name = tensor("op_2327")]; tensor input_655 = add(x = input_643, y = var_2327)[name = tensor("input_655")]; tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; tensor query_25 = layer_norm(axes = query_25_axes_0, beta = encoder_layers_12_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_12_norm_self_att_weight, x = input_655)[name = tensor("query_25")]; tensor var_2343 = linear(bias = encoder_layers_12_self_attn_linear_q_bias, weight = encoder_layers_12_self_attn_linear_q_weight, x = query_25)[name = tensor("linear_111")]; tensor var_2344 = const()[name = tensor("op_2344"), val = tensor([1, -1, 8, 128])]; tensor q_73 = reshape(shape = var_2344, x = var_2343)[name = tensor("q_73")]; tensor var_2348 = linear(bias = encoder_layers_12_self_attn_linear_k_bias, weight = encoder_layers_12_self_attn_linear_k_weight, x = query_25)[name = tensor("linear_112")]; tensor var_2349 = const()[name = tensor("op_2349"), val = tensor([1, -1, 8, 128])]; tensor k_49 = reshape(shape = var_2349, x = var_2348)[name = tensor("k_49")]; tensor var_2353 = linear(bias = encoder_layers_12_self_attn_linear_v_bias, weight = encoder_layers_12_self_attn_linear_v_weight, x = query_25)[name = tensor("linear_113")]; tensor var_2354 = const()[name = tensor("op_2354"), val = tensor([1, -1, 8, 128])]; tensor v_25 = reshape(shape = var_2354, x = var_2353)[name = tensor("v_25")]; tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2366 = add(x = q_73, y = encoder_layers_12_self_attn_pos_bias_u)[name = tensor("op_2366")]; tensor var_2368 = add(x = q_73, y = encoder_layers_12_self_attn_pos_bias_v)[name = tensor("op_2368")]; tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2370 = const()[name = tensor("op_2370"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2352332032)))]; tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_25 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2368)[name = tensor("transpose_202")]; tensor x_273 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = q_with_bias_v_25, y = var_2370)[name = tensor("x_273")]; tensor const_134 = const()[name = tensor("const_134"), val = tensor(0x0p+0)]; tensor x_275_pad_0 = const()[name = tensor("x_275_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_275_mode_0 = const()[name = tensor("x_275_mode_0"), val = tensor("constant")]; tensor x_275 = pad(constant_val = const_134, mode = x_275_mode_0, pad = x_275_pad_0, x = x_273)[name = tensor("x_275")]; tensor var_2378 = const()[name = tensor("op_2378"), val = tensor([1, 8, -1, 188])]; tensor x_277 = reshape(shape = var_2378, x = x_275)[name = tensor("x_277")]; tensor var_2382_begin_0 = const()[name = tensor("op_2382_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2382_end_0 = const()[name = tensor("op_2382_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2382_end_mask_0 = const()[name = tensor("op_2382_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2382 = slice_by_index(begin = var_2382_begin_0, end = var_2382_end_0, end_mask = var_2382_end_mask_0, x = x_277)[name = tensor("op_2382")]; tensor var_2383 = const()[name = tensor("op_2383"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_49 = reshape(shape = var_2383, x = var_2382)[name = tensor("matrix_bd_49")]; tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; tensor transpose_96_perm_0 = const()[name = tensor("transpose_96_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_97_perm_0 = const()[name = tensor("transpose_97_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_97 = transpose(perm = transpose_97_perm_0, x = k_49)[name = tensor("transpose_200")]; tensor transpose_96 = transpose(perm = transpose_96_perm_0, x = var_2366)[name = tensor("transpose_201")]; tensor matrix_ac_25 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_96, y = transpose_97)[name = tensor("matrix_ac_25")]; tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_51 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49)[name = tensor("matrix_bd_51")]; tensor var_2392 = add(x = matrix_ac_25, y = matrix_bd_51)[name = tensor("op_2392")]; tensor _inversed_scores_49_y_0 = const()[name = tensor("_inversed_scores_49_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_49 = mul(x = var_2392, y = _inversed_scores_49_y_0)[name = tensor("_inversed_scores_49")]; tensor scores_51 = select(a = var_14, b = _inversed_scores_49, cond = mask_3)[name = tensor("scores_51")]; tensor var_2398 = softmax(axis = var_32, x = scores_51)[name = tensor("op_2398")]; tensor input_657 = select(a = var_13, b = var_2398, cond = mask_3)[name = tensor("input_657")]; tensor x_279_transpose_x_0 = const()[name = tensor("x_279_transpose_x_0"), val = tensor(false)]; tensor x_279_transpose_y_0 = const()[name = tensor("x_279_transpose_y_0"), val = tensor(false)]; tensor value_27 = transpose(perm = value_27_perm_0, x = v_25)[name = tensor("transpose_203")]; tensor x_279 = matmul(transpose_x = x_279_transpose_x_0, transpose_y = x_279_transpose_y_0, x = input_657, y = value_27)[name = tensor("x_279")]; tensor var_2402_perm_0 = const()[name = tensor("op_2402_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2403 = const()[name = tensor("op_2403"), val = tensor([1, -1, 1024])]; tensor var_2402 = transpose(perm = var_2402_perm_0, x = x_279)[name = tensor("transpose_199")]; tensor input_659 = reshape(shape = var_2403, x = var_2402)[name = tensor("input_659")]; tensor input_661 = linear(bias = encoder_layers_12_self_attn_linear_out_bias, weight = encoder_layers_12_self_attn_linear_out_weight, x = input_659)[name = tensor("linear_115")]; tensor input_663 = add(x = input_655, y = input_661)[name = tensor("input_663")]; tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; tensor x_283 = layer_norm(axes = x_283_axes_0, beta = encoder_layers_12_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_12_norm_conv_weight, x = input_663)[name = tensor("x_283")]; tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; tensor input_665 = transpose(perm = input_665_perm_0, x = x_283)[name = tensor("transpose_198")]; tensor input_667 = conv(bias = encoder_layers_12_conv_pointwise_conv1_bias, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = encoder_layers_12_conv_pointwise_conv1_weight, x = input_665)[name = tensor("input_667")]; tensor x_285_split_num_splits_0 = const()[name = tensor("x_285_split_num_splits_0"), val = tensor(2)]; tensor x_285_split_axis_0 = const()[name = tensor("x_285_split_axis_0"), val = tensor(1)]; tensor x_285_split_0, tensor x_285_split_1 = split(axis = x_285_split_axis_0, num_splits = x_285_split_num_splits_0, x = input_667)[name = tensor("x_285_split")]; tensor x_285_split_1_sigmoid = sigmoid(x = x_285_split_1)[name = tensor("x_285_split_1_sigmoid")]; tensor x_285 = mul(x = x_285_split_0, y = x_285_split_1_sigmoid)[name = tensor("x_285")]; tensor input_669 = select(a = var_13, b = x_285, cond = var_339)[name = tensor("input_669")]; tensor const_137 = const()[name = tensor("const_137"), val = tensor(0x0p+0)]; tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; tensor input_671 = pad(constant_val = const_137, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669)[name = tensor("input_671")]; tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(1024)]; tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; tensor const_272 = const()[name = tensor("const_272"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2353868096)))]; tensor const_273 = const()[name = tensor("const_273"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2353905024)))]; tensor input_675 = conv(bias = const_273, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_272, x = input_671)[name = tensor("input_675")]; tensor input_677 = silu(x = input_675)[name = tensor("input_677")]; tensor x_287_pad_type_0 = const()[name = tensor("x_287_pad_type_0"), val = tensor("valid")]; tensor x_287_strides_0 = const()[name = tensor("x_287_strides_0"), val = tensor([1])]; tensor x_287_pad_0 = const()[name = tensor("x_287_pad_0"), val = tensor([0, 0])]; tensor x_287_dilations_0 = const()[name = tensor("x_287_dilations_0"), val = tensor([1])]; tensor x_287_groups_0 = const()[name = tensor("x_287_groups_0"), val = tensor(1)]; tensor x_287 = conv(bias = encoder_layers_12_conv_pointwise_conv2_bias, dilations = x_287_dilations_0, groups = x_287_groups_0, pad = x_287_pad_0, pad_type = x_287_pad_type_0, strides = x_287_strides_0, weight = encoder_layers_12_conv_pointwise_conv2_weight, x = input_677)[name = tensor("x_287")]; tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; tensor input_679 = transpose(perm = input_679_perm_0, x = x_287)[name = tensor("transpose_197")]; tensor input_681 = add(x = input_663, y = input_679)[name = tensor("input_681")]; tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; tensor input_683 = layer_norm(axes = input_683_axes_0, beta = encoder_layers_12_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_12_norm_feed_forward2_weight, x = input_681)[name = tensor("input_683")]; tensor input_685 = linear(bias = encoder_layers_12_feed_forward2_linear1_bias, weight = encoder_layers_12_feed_forward2_linear1_weight, x = input_683)[name = tensor("linear_116")]; tensor input_687 = silu(x = input_685)[name = tensor("input_687")]; tensor input_691 = linear(bias = encoder_layers_12_feed_forward2_linear2_bias, weight = encoder_layers_12_feed_forward2_linear2_weight, x = input_687)[name = tensor("linear_117")]; tensor var_2469 = const()[name = tensor("op_2469"), val = tensor(0x1p-1)]; tensor var_2470 = mul(x = input_691, y = var_2469)[name = tensor("op_2470")]; tensor input_693 = add(x = input_681, y = var_2470)[name = tensor("input_693")]; tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; tensor input_695 = layer_norm(axes = input_695_axes_0, beta = encoder_layers_12_norm_out_bias, epsilon = var_11, gamma = encoder_layers_12_norm_out_weight, x = input_693)[name = tensor("input_695")]; tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; tensor input_697 = layer_norm(axes = input_697_axes_0, beta = encoder_layers_13_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_13_norm_feed_forward1_weight, x = input_695)[name = tensor("input_697")]; tensor input_699 = linear(bias = encoder_layers_13_feed_forward1_linear1_bias, weight = encoder_layers_13_feed_forward1_linear1_weight, x = input_697)[name = tensor("linear_118")]; tensor input_701 = silu(x = input_699)[name = tensor("input_701")]; tensor input_705 = linear(bias = encoder_layers_13_feed_forward1_linear2_bias, weight = encoder_layers_13_feed_forward1_linear2_weight, x = input_701)[name = tensor("linear_119")]; tensor var_2500 = const()[name = tensor("op_2500"), val = tensor(0x1p-1)]; tensor var_2501 = mul(x = input_705, y = var_2500)[name = tensor("op_2501")]; tensor input_707 = add(x = input_695, y = var_2501)[name = tensor("input_707")]; tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; tensor query_27 = layer_norm(axes = query_27_axes_0, beta = encoder_layers_13_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_13_norm_self_att_weight, x = input_707)[name = tensor("query_27")]; tensor var_2517 = linear(bias = encoder_layers_13_self_attn_linear_q_bias, weight = encoder_layers_13_self_attn_linear_q_weight, x = query_27)[name = tensor("linear_120")]; tensor var_2518 = const()[name = tensor("op_2518"), val = tensor([1, -1, 8, 128])]; tensor q_79 = reshape(shape = var_2518, x = var_2517)[name = tensor("q_79")]; tensor var_2522 = linear(bias = encoder_layers_13_self_attn_linear_k_bias, weight = encoder_layers_13_self_attn_linear_k_weight, x = query_27)[name = tensor("linear_121")]; tensor var_2523 = const()[name = tensor("op_2523"), val = tensor([1, -1, 8, 128])]; tensor k_53 = reshape(shape = var_2523, x = var_2522)[name = tensor("k_53")]; tensor var_2527 = linear(bias = encoder_layers_13_self_attn_linear_v_bias, weight = encoder_layers_13_self_attn_linear_v_weight, x = query_27)[name = tensor("linear_122")]; tensor var_2528 = const()[name = tensor("op_2528"), val = tensor([1, -1, 8, 128])]; tensor v_27 = reshape(shape = var_2528, x = var_2527)[name = tensor("v_27")]; tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2540 = add(x = q_79, y = encoder_layers_13_self_attn_pos_bias_u)[name = tensor("op_2540")]; tensor var_2542 = add(x = q_79, y = encoder_layers_13_self_attn_pos_bias_v)[name = tensor("op_2542")]; tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2544 = const()[name = tensor("op_2544"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2353909184)))]; tensor x_295_transpose_x_0 = const()[name = tensor("x_295_transpose_x_0"), val = tensor(false)]; tensor x_295_transpose_y_0 = const()[name = tensor("x_295_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_27 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2542)[name = tensor("transpose_195")]; tensor x_295 = matmul(transpose_x = x_295_transpose_x_0, transpose_y = x_295_transpose_y_0, x = q_with_bias_v_27, y = var_2544)[name = tensor("x_295")]; tensor const_144 = const()[name = tensor("const_144"), val = tensor(0x0p+0)]; tensor x_297_pad_0 = const()[name = tensor("x_297_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_297_mode_0 = const()[name = tensor("x_297_mode_0"), val = tensor("constant")]; tensor x_297 = pad(constant_val = const_144, mode = x_297_mode_0, pad = x_297_pad_0, x = x_295)[name = tensor("x_297")]; tensor var_2552 = const()[name = tensor("op_2552"), val = tensor([1, 8, -1, 188])]; tensor x_299 = reshape(shape = var_2552, x = x_297)[name = tensor("x_299")]; tensor var_2556_begin_0 = const()[name = tensor("op_2556_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2556_end_0 = const()[name = tensor("op_2556_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2556_end_mask_0 = const()[name = tensor("op_2556_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2556 = slice_by_index(begin = var_2556_begin_0, end = var_2556_end_0, end_mask = var_2556_end_mask_0, x = x_299)[name = tensor("op_2556")]; tensor var_2557 = const()[name = tensor("op_2557"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_53 = reshape(shape = var_2557, x = var_2556)[name = tensor("matrix_bd_53")]; tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; tensor transpose_98_perm_0 = const()[name = tensor("transpose_98_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_99_perm_0 = const()[name = tensor("transpose_99_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_99 = transpose(perm = transpose_99_perm_0, x = k_53)[name = tensor("transpose_193")]; tensor transpose_98 = transpose(perm = transpose_98_perm_0, x = var_2540)[name = tensor("transpose_194")]; tensor matrix_ac_27 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_98, y = transpose_99)[name = tensor("matrix_ac_27")]; tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_55 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53)[name = tensor("matrix_bd_55")]; tensor var_2566 = add(x = matrix_ac_27, y = matrix_bd_55)[name = tensor("op_2566")]; tensor _inversed_scores_53_y_0 = const()[name = tensor("_inversed_scores_53_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_53 = mul(x = var_2566, y = _inversed_scores_53_y_0)[name = tensor("_inversed_scores_53")]; tensor scores_55 = select(a = var_14, b = _inversed_scores_53, cond = mask_3)[name = tensor("scores_55")]; tensor var_2572 = softmax(axis = var_32, x = scores_55)[name = tensor("op_2572")]; tensor input_709 = select(a = var_13, b = var_2572, cond = mask_3)[name = tensor("input_709")]; tensor x_301_transpose_x_0 = const()[name = tensor("x_301_transpose_x_0"), val = tensor(false)]; tensor x_301_transpose_y_0 = const()[name = tensor("x_301_transpose_y_0"), val = tensor(false)]; tensor value_29 = transpose(perm = value_29_perm_0, x = v_27)[name = tensor("transpose_196")]; tensor x_301 = matmul(transpose_x = x_301_transpose_x_0, transpose_y = x_301_transpose_y_0, x = input_709, y = value_29)[name = tensor("x_301")]; tensor var_2576_perm_0 = const()[name = tensor("op_2576_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2577 = const()[name = tensor("op_2577"), val = tensor([1, -1, 1024])]; tensor var_2576 = transpose(perm = var_2576_perm_0, x = x_301)[name = tensor("transpose_192")]; tensor input_711 = reshape(shape = var_2577, x = var_2576)[name = tensor("input_711")]; tensor input_713 = linear(bias = encoder_layers_13_self_attn_linear_out_bias, weight = encoder_layers_13_self_attn_linear_out_weight, x = input_711)[name = tensor("linear_124")]; tensor input_715 = add(x = input_707, y = input_713)[name = tensor("input_715")]; tensor x_305_axes_0 = const()[name = tensor("x_305_axes_0"), val = tensor([-1])]; tensor x_305 = layer_norm(axes = x_305_axes_0, beta = encoder_layers_13_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_13_norm_conv_weight, x = input_715)[name = tensor("x_305")]; tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; tensor input_717 = transpose(perm = input_717_perm_0, x = x_305)[name = tensor("transpose_191")]; tensor input_719 = conv(bias = encoder_layers_13_conv_pointwise_conv1_bias, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = encoder_layers_13_conv_pointwise_conv1_weight, x = input_717)[name = tensor("input_719")]; tensor x_307_split_num_splits_0 = const()[name = tensor("x_307_split_num_splits_0"), val = tensor(2)]; tensor x_307_split_axis_0 = const()[name = tensor("x_307_split_axis_0"), val = tensor(1)]; tensor x_307_split_0, tensor x_307_split_1 = split(axis = x_307_split_axis_0, num_splits = x_307_split_num_splits_0, x = input_719)[name = tensor("x_307_split")]; tensor x_307_split_1_sigmoid = sigmoid(x = x_307_split_1)[name = tensor("x_307_split_1_sigmoid")]; tensor x_307 = mul(x = x_307_split_0, y = x_307_split_1_sigmoid)[name = tensor("x_307")]; tensor input_721 = select(a = var_13, b = x_307, cond = var_339)[name = tensor("input_721")]; tensor const_147 = const()[name = tensor("const_147"), val = tensor(0x0p+0)]; tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; tensor input_723 = pad(constant_val = const_147, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721)[name = tensor("input_723")]; tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(1024)]; tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; tensor const_274 = const()[name = tensor("const_274"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2355445248)))]; tensor const_275 = const()[name = tensor("const_275"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2355482176)))]; tensor input_727 = conv(bias = const_275, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_274, x = input_723)[name = tensor("input_727")]; tensor input_729 = silu(x = input_727)[name = tensor("input_729")]; tensor x_309_pad_type_0 = const()[name = tensor("x_309_pad_type_0"), val = tensor("valid")]; tensor x_309_strides_0 = const()[name = tensor("x_309_strides_0"), val = tensor([1])]; tensor x_309_pad_0 = const()[name = tensor("x_309_pad_0"), val = tensor([0, 0])]; tensor x_309_dilations_0 = const()[name = tensor("x_309_dilations_0"), val = tensor([1])]; tensor x_309_groups_0 = const()[name = tensor("x_309_groups_0"), val = tensor(1)]; tensor x_309 = conv(bias = encoder_layers_13_conv_pointwise_conv2_bias, dilations = x_309_dilations_0, groups = x_309_groups_0, pad = x_309_pad_0, pad_type = x_309_pad_type_0, strides = x_309_strides_0, weight = encoder_layers_13_conv_pointwise_conv2_weight, x = input_729)[name = tensor("x_309")]; tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; tensor input_731 = transpose(perm = input_731_perm_0, x = x_309)[name = tensor("transpose_190")]; tensor input_733 = add(x = input_715, y = input_731)[name = tensor("input_733")]; tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; tensor input_735 = layer_norm(axes = input_735_axes_0, beta = encoder_layers_13_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_13_norm_feed_forward2_weight, x = input_733)[name = tensor("input_735")]; tensor input_737 = linear(bias = encoder_layers_13_feed_forward2_linear1_bias, weight = encoder_layers_13_feed_forward2_linear1_weight, x = input_735)[name = tensor("linear_125")]; tensor input_739 = silu(x = input_737)[name = tensor("input_739")]; tensor input_743 = linear(bias = encoder_layers_13_feed_forward2_linear2_bias, weight = encoder_layers_13_feed_forward2_linear2_weight, x = input_739)[name = tensor("linear_126")]; tensor var_2643 = const()[name = tensor("op_2643"), val = tensor(0x1p-1)]; tensor var_2644 = mul(x = input_743, y = var_2643)[name = tensor("op_2644")]; tensor input_745 = add(x = input_733, y = var_2644)[name = tensor("input_745")]; tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; tensor input_747 = layer_norm(axes = input_747_axes_0, beta = encoder_layers_13_norm_out_bias, epsilon = var_11, gamma = encoder_layers_13_norm_out_weight, x = input_745)[name = tensor("input_747")]; tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; tensor input_749 = layer_norm(axes = input_749_axes_0, beta = encoder_layers_14_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_14_norm_feed_forward1_weight, x = input_747)[name = tensor("input_749")]; tensor input_751 = linear(bias = encoder_layers_14_feed_forward1_linear1_bias, weight = encoder_layers_14_feed_forward1_linear1_weight, x = input_749)[name = tensor("linear_127")]; tensor input_753 = silu(x = input_751)[name = tensor("input_753")]; tensor input_757 = linear(bias = encoder_layers_14_feed_forward1_linear2_bias, weight = encoder_layers_14_feed_forward1_linear2_weight, x = input_753)[name = tensor("linear_128")]; tensor var_2674 = const()[name = tensor("op_2674"), val = tensor(0x1p-1)]; tensor var_2675 = mul(x = input_757, y = var_2674)[name = tensor("op_2675")]; tensor input_759 = add(x = input_747, y = var_2675)[name = tensor("input_759")]; tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; tensor query_29 = layer_norm(axes = query_29_axes_0, beta = encoder_layers_14_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_14_norm_self_att_weight, x = input_759)[name = tensor("query_29")]; tensor var_2691 = linear(bias = encoder_layers_14_self_attn_linear_q_bias, weight = encoder_layers_14_self_attn_linear_q_weight, x = query_29)[name = tensor("linear_129")]; tensor var_2692 = const()[name = tensor("op_2692"), val = tensor([1, -1, 8, 128])]; tensor q_85 = reshape(shape = var_2692, x = var_2691)[name = tensor("q_85")]; tensor var_2696 = linear(bias = encoder_layers_14_self_attn_linear_k_bias, weight = encoder_layers_14_self_attn_linear_k_weight, x = query_29)[name = tensor("linear_130")]; tensor var_2697 = const()[name = tensor("op_2697"), val = tensor([1, -1, 8, 128])]; tensor k_57 = reshape(shape = var_2697, x = var_2696)[name = tensor("k_57")]; tensor var_2701 = linear(bias = encoder_layers_14_self_attn_linear_v_bias, weight = encoder_layers_14_self_attn_linear_v_weight, x = query_29)[name = tensor("linear_131")]; tensor var_2702 = const()[name = tensor("op_2702"), val = tensor([1, -1, 8, 128])]; tensor v_29 = reshape(shape = var_2702, x = var_2701)[name = tensor("v_29")]; tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2714 = add(x = q_85, y = encoder_layers_14_self_attn_pos_bias_u)[name = tensor("op_2714")]; tensor var_2716 = add(x = q_85, y = encoder_layers_14_self_attn_pos_bias_v)[name = tensor("op_2716")]; tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2718 = const()[name = tensor("op_2718"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2355486336)))]; tensor x_317_transpose_x_0 = const()[name = tensor("x_317_transpose_x_0"), val = tensor(false)]; tensor x_317_transpose_y_0 = const()[name = tensor("x_317_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_29 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2716)[name = tensor("transpose_188")]; tensor x_317 = matmul(transpose_x = x_317_transpose_x_0, transpose_y = x_317_transpose_y_0, x = q_with_bias_v_29, y = var_2718)[name = tensor("x_317")]; tensor const_154 = const()[name = tensor("const_154"), val = tensor(0x0p+0)]; tensor x_319_pad_0 = const()[name = tensor("x_319_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_319_mode_0 = const()[name = tensor("x_319_mode_0"), val = tensor("constant")]; tensor x_319 = pad(constant_val = const_154, mode = x_319_mode_0, pad = x_319_pad_0, x = x_317)[name = tensor("x_319")]; tensor var_2726 = const()[name = tensor("op_2726"), val = tensor([1, 8, -1, 188])]; tensor x_321 = reshape(shape = var_2726, x = x_319)[name = tensor("x_321")]; tensor var_2730_begin_0 = const()[name = tensor("op_2730_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2730_end_0 = const()[name = tensor("op_2730_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2730_end_mask_0 = const()[name = tensor("op_2730_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2730 = slice_by_index(begin = var_2730_begin_0, end = var_2730_end_0, end_mask = var_2730_end_mask_0, x = x_321)[name = tensor("op_2730")]; tensor var_2731 = const()[name = tensor("op_2731"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_57 = reshape(shape = var_2731, x = var_2730)[name = tensor("matrix_bd_57")]; tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; tensor transpose_100_perm_0 = const()[name = tensor("transpose_100_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_101_perm_0 = const()[name = tensor("transpose_101_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_101 = transpose(perm = transpose_101_perm_0, x = k_57)[name = tensor("transpose_186")]; tensor transpose_100 = transpose(perm = transpose_100_perm_0, x = var_2714)[name = tensor("transpose_187")]; tensor matrix_ac_29 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_100, y = transpose_101)[name = tensor("matrix_ac_29")]; tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_59 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57)[name = tensor("matrix_bd_59")]; tensor var_2740 = add(x = matrix_ac_29, y = matrix_bd_59)[name = tensor("op_2740")]; tensor _inversed_scores_57_y_0 = const()[name = tensor("_inversed_scores_57_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_57 = mul(x = var_2740, y = _inversed_scores_57_y_0)[name = tensor("_inversed_scores_57")]; tensor scores_59 = select(a = var_14, b = _inversed_scores_57, cond = mask_3)[name = tensor("scores_59")]; tensor var_2746 = softmax(axis = var_32, x = scores_59)[name = tensor("op_2746")]; tensor input_761 = select(a = var_13, b = var_2746, cond = mask_3)[name = tensor("input_761")]; tensor x_323_transpose_x_0 = const()[name = tensor("x_323_transpose_x_0"), val = tensor(false)]; tensor x_323_transpose_y_0 = const()[name = tensor("x_323_transpose_y_0"), val = tensor(false)]; tensor value_31 = transpose(perm = value_31_perm_0, x = v_29)[name = tensor("transpose_189")]; tensor x_323 = matmul(transpose_x = x_323_transpose_x_0, transpose_y = x_323_transpose_y_0, x = input_761, y = value_31)[name = tensor("x_323")]; tensor var_2750_perm_0 = const()[name = tensor("op_2750_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2751 = const()[name = tensor("op_2751"), val = tensor([1, -1, 1024])]; tensor var_2750 = transpose(perm = var_2750_perm_0, x = x_323)[name = tensor("transpose_185")]; tensor input_763 = reshape(shape = var_2751, x = var_2750)[name = tensor("input_763")]; tensor input_765 = linear(bias = encoder_layers_14_self_attn_linear_out_bias, weight = encoder_layers_14_self_attn_linear_out_weight, x = input_763)[name = tensor("linear_133")]; tensor input_767 = add(x = input_759, y = input_765)[name = tensor("input_767")]; tensor x_327_axes_0 = const()[name = tensor("x_327_axes_0"), val = tensor([-1])]; tensor x_327 = layer_norm(axes = x_327_axes_0, beta = encoder_layers_14_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_14_norm_conv_weight, x = input_767)[name = tensor("x_327")]; tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; tensor input_769 = transpose(perm = input_769_perm_0, x = x_327)[name = tensor("transpose_184")]; tensor input_771 = conv(bias = encoder_layers_14_conv_pointwise_conv1_bias, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = encoder_layers_14_conv_pointwise_conv1_weight, x = input_769)[name = tensor("input_771")]; tensor x_329_split_num_splits_0 = const()[name = tensor("x_329_split_num_splits_0"), val = tensor(2)]; tensor x_329_split_axis_0 = const()[name = tensor("x_329_split_axis_0"), val = tensor(1)]; tensor x_329_split_0, tensor x_329_split_1 = split(axis = x_329_split_axis_0, num_splits = x_329_split_num_splits_0, x = input_771)[name = tensor("x_329_split")]; tensor x_329_split_1_sigmoid = sigmoid(x = x_329_split_1)[name = tensor("x_329_split_1_sigmoid")]; tensor x_329 = mul(x = x_329_split_0, y = x_329_split_1_sigmoid)[name = tensor("x_329")]; tensor input_773 = select(a = var_13, b = x_329, cond = var_339)[name = tensor("input_773")]; tensor const_157 = const()[name = tensor("const_157"), val = tensor(0x0p+0)]; tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; tensor input_775 = pad(constant_val = const_157, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773)[name = tensor("input_775")]; tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(1024)]; tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; tensor const_276 = const()[name = tensor("const_276"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2357022400)))]; tensor const_277 = const()[name = tensor("const_277"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2357059328)))]; tensor input_779 = conv(bias = const_277, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_276, x = input_775)[name = tensor("input_779")]; tensor input_781 = silu(x = input_779)[name = tensor("input_781")]; tensor x_331_pad_type_0 = const()[name = tensor("x_331_pad_type_0"), val = tensor("valid")]; tensor x_331_strides_0 = const()[name = tensor("x_331_strides_0"), val = tensor([1])]; tensor x_331_pad_0 = const()[name = tensor("x_331_pad_0"), val = tensor([0, 0])]; tensor x_331_dilations_0 = const()[name = tensor("x_331_dilations_0"), val = tensor([1])]; tensor x_331_groups_0 = const()[name = tensor("x_331_groups_0"), val = tensor(1)]; tensor x_331 = conv(bias = encoder_layers_14_conv_pointwise_conv2_bias, dilations = x_331_dilations_0, groups = x_331_groups_0, pad = x_331_pad_0, pad_type = x_331_pad_type_0, strides = x_331_strides_0, weight = encoder_layers_14_conv_pointwise_conv2_weight, x = input_781)[name = tensor("x_331")]; tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; tensor input_783 = transpose(perm = input_783_perm_0, x = x_331)[name = tensor("transpose_183")]; tensor input_785 = add(x = input_767, y = input_783)[name = tensor("input_785")]; tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; tensor input_787 = layer_norm(axes = input_787_axes_0, beta = encoder_layers_14_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_14_norm_feed_forward2_weight, x = input_785)[name = tensor("input_787")]; tensor input_789 = linear(bias = encoder_layers_14_feed_forward2_linear1_bias, weight = encoder_layers_14_feed_forward2_linear1_weight, x = input_787)[name = tensor("linear_134")]; tensor input_791 = silu(x = input_789)[name = tensor("input_791")]; tensor input_795 = linear(bias = encoder_layers_14_feed_forward2_linear2_bias, weight = encoder_layers_14_feed_forward2_linear2_weight, x = input_791)[name = tensor("linear_135")]; tensor var_2817 = const()[name = tensor("op_2817"), val = tensor(0x1p-1)]; tensor var_2818 = mul(x = input_795, y = var_2817)[name = tensor("op_2818")]; tensor input_797 = add(x = input_785, y = var_2818)[name = tensor("input_797")]; tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; tensor input_799 = layer_norm(axes = input_799_axes_0, beta = encoder_layers_14_norm_out_bias, epsilon = var_11, gamma = encoder_layers_14_norm_out_weight, x = input_797)[name = tensor("input_799")]; tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; tensor input_801 = layer_norm(axes = input_801_axes_0, beta = encoder_layers_15_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_15_norm_feed_forward1_weight, x = input_799)[name = tensor("input_801")]; tensor input_803 = linear(bias = encoder_layers_15_feed_forward1_linear1_bias, weight = encoder_layers_15_feed_forward1_linear1_weight, x = input_801)[name = tensor("linear_136")]; tensor input_805 = silu(x = input_803)[name = tensor("input_805")]; tensor input_809 = linear(bias = encoder_layers_15_feed_forward1_linear2_bias, weight = encoder_layers_15_feed_forward1_linear2_weight, x = input_805)[name = tensor("linear_137")]; tensor var_2848 = const()[name = tensor("op_2848"), val = tensor(0x1p-1)]; tensor var_2849 = mul(x = input_809, y = var_2848)[name = tensor("op_2849")]; tensor input_811 = add(x = input_799, y = var_2849)[name = tensor("input_811")]; tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; tensor query_31 = layer_norm(axes = query_31_axes_0, beta = encoder_layers_15_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_15_norm_self_att_weight, x = input_811)[name = tensor("query_31")]; tensor var_2865 = linear(bias = encoder_layers_15_self_attn_linear_q_bias, weight = encoder_layers_15_self_attn_linear_q_weight, x = query_31)[name = tensor("linear_138")]; tensor var_2866 = const()[name = tensor("op_2866"), val = tensor([1, -1, 8, 128])]; tensor q_91 = reshape(shape = var_2866, x = var_2865)[name = tensor("q_91")]; tensor var_2870 = linear(bias = encoder_layers_15_self_attn_linear_k_bias, weight = encoder_layers_15_self_attn_linear_k_weight, x = query_31)[name = tensor("linear_139")]; tensor var_2871 = const()[name = tensor("op_2871"), val = tensor([1, -1, 8, 128])]; tensor k_61 = reshape(shape = var_2871, x = var_2870)[name = tensor("k_61")]; tensor var_2875 = linear(bias = encoder_layers_15_self_attn_linear_v_bias, weight = encoder_layers_15_self_attn_linear_v_weight, x = query_31)[name = tensor("linear_140")]; tensor var_2876 = const()[name = tensor("op_2876"), val = tensor([1, -1, 8, 128])]; tensor v_31 = reshape(shape = var_2876, x = var_2875)[name = tensor("v_31")]; tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2888 = add(x = q_91, y = encoder_layers_15_self_attn_pos_bias_u)[name = tensor("op_2888")]; tensor var_2890 = add(x = q_91, y = encoder_layers_15_self_attn_pos_bias_v)[name = tensor("op_2890")]; tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2892 = const()[name = tensor("op_2892"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2357063488)))]; tensor x_339_transpose_x_0 = const()[name = tensor("x_339_transpose_x_0"), val = tensor(false)]; tensor x_339_transpose_y_0 = const()[name = tensor("x_339_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_31 = transpose(perm = q_with_bias_v_31_perm_0, x = var_2890)[name = tensor("transpose_181")]; tensor x_339 = matmul(transpose_x = x_339_transpose_x_0, transpose_y = x_339_transpose_y_0, x = q_with_bias_v_31, y = var_2892)[name = tensor("x_339")]; tensor const_164 = const()[name = tensor("const_164"), val = tensor(0x0p+0)]; tensor x_341_pad_0 = const()[name = tensor("x_341_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_341_mode_0 = const()[name = tensor("x_341_mode_0"), val = tensor("constant")]; tensor x_341 = pad(constant_val = const_164, mode = x_341_mode_0, pad = x_341_pad_0, x = x_339)[name = tensor("x_341")]; tensor var_2900 = const()[name = tensor("op_2900"), val = tensor([1, 8, -1, 188])]; tensor x_343 = reshape(shape = var_2900, x = x_341)[name = tensor("x_343")]; tensor var_2904_begin_0 = const()[name = tensor("op_2904_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2904_end_0 = const()[name = tensor("op_2904_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2904_end_mask_0 = const()[name = tensor("op_2904_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2904 = slice_by_index(begin = var_2904_begin_0, end = var_2904_end_0, end_mask = var_2904_end_mask_0, x = x_343)[name = tensor("op_2904")]; tensor var_2905 = const()[name = tensor("op_2905"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_61 = reshape(shape = var_2905, x = var_2904)[name = tensor("matrix_bd_61")]; tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; tensor transpose_102_perm_0 = const()[name = tensor("transpose_102_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_103_perm_0 = const()[name = tensor("transpose_103_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_103 = transpose(perm = transpose_103_perm_0, x = k_61)[name = tensor("transpose_179")]; tensor transpose_102 = transpose(perm = transpose_102_perm_0, x = var_2888)[name = tensor("transpose_180")]; tensor matrix_ac_31 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_102, y = transpose_103)[name = tensor("matrix_ac_31")]; tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_63 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61)[name = tensor("matrix_bd_63")]; tensor var_2914 = add(x = matrix_ac_31, y = matrix_bd_63)[name = tensor("op_2914")]; tensor _inversed_scores_61_y_0 = const()[name = tensor("_inversed_scores_61_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_61 = mul(x = var_2914, y = _inversed_scores_61_y_0)[name = tensor("_inversed_scores_61")]; tensor scores_63 = select(a = var_14, b = _inversed_scores_61, cond = mask_3)[name = tensor("scores_63")]; tensor var_2920 = softmax(axis = var_32, x = scores_63)[name = tensor("op_2920")]; tensor input_813 = select(a = var_13, b = var_2920, cond = mask_3)[name = tensor("input_813")]; tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; tensor value_33 = transpose(perm = value_33_perm_0, x = v_31)[name = tensor("transpose_182")]; tensor x_345 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = input_813, y = value_33)[name = tensor("x_345")]; tensor var_2924_perm_0 = const()[name = tensor("op_2924_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2925 = const()[name = tensor("op_2925"), val = tensor([1, -1, 1024])]; tensor var_2924 = transpose(perm = var_2924_perm_0, x = x_345)[name = tensor("transpose_178")]; tensor input_815 = reshape(shape = var_2925, x = var_2924)[name = tensor("input_815")]; tensor input_817 = linear(bias = encoder_layers_15_self_attn_linear_out_bias, weight = encoder_layers_15_self_attn_linear_out_weight, x = input_815)[name = tensor("linear_142")]; tensor input_819 = add(x = input_811, y = input_817)[name = tensor("input_819")]; tensor x_349_axes_0 = const()[name = tensor("x_349_axes_0"), val = tensor([-1])]; tensor x_349 = layer_norm(axes = x_349_axes_0, beta = encoder_layers_15_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_15_norm_conv_weight, x = input_819)[name = tensor("x_349")]; tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; tensor input_821 = transpose(perm = input_821_perm_0, x = x_349)[name = tensor("transpose_177")]; tensor input_823 = conv(bias = encoder_layers_15_conv_pointwise_conv1_bias, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = encoder_layers_15_conv_pointwise_conv1_weight, x = input_821)[name = tensor("input_823")]; tensor x_351_split_num_splits_0 = const()[name = tensor("x_351_split_num_splits_0"), val = tensor(2)]; tensor x_351_split_axis_0 = const()[name = tensor("x_351_split_axis_0"), val = tensor(1)]; tensor x_351_split_0, tensor x_351_split_1 = split(axis = x_351_split_axis_0, num_splits = x_351_split_num_splits_0, x = input_823)[name = tensor("x_351_split")]; tensor x_351_split_1_sigmoid = sigmoid(x = x_351_split_1)[name = tensor("x_351_split_1_sigmoid")]; tensor x_351 = mul(x = x_351_split_0, y = x_351_split_1_sigmoid)[name = tensor("x_351")]; tensor input_825 = select(a = var_13, b = x_351, cond = var_339)[name = tensor("input_825")]; tensor const_167 = const()[name = tensor("const_167"), val = tensor(0x0p+0)]; tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; tensor input_827 = pad(constant_val = const_167, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825)[name = tensor("input_827")]; tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(1024)]; tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; tensor const_278 = const()[name = tensor("const_278"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2358599552)))]; tensor const_279 = const()[name = tensor("const_279"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2358636480)))]; tensor input_831 = conv(bias = const_279, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_278, x = input_827)[name = tensor("input_831")]; tensor input_833 = silu(x = input_831)[name = tensor("input_833")]; tensor x_353_pad_type_0 = const()[name = tensor("x_353_pad_type_0"), val = tensor("valid")]; tensor x_353_strides_0 = const()[name = tensor("x_353_strides_0"), val = tensor([1])]; tensor x_353_pad_0 = const()[name = tensor("x_353_pad_0"), val = tensor([0, 0])]; tensor x_353_dilations_0 = const()[name = tensor("x_353_dilations_0"), val = tensor([1])]; tensor x_353_groups_0 = const()[name = tensor("x_353_groups_0"), val = tensor(1)]; tensor x_353 = conv(bias = encoder_layers_15_conv_pointwise_conv2_bias, dilations = x_353_dilations_0, groups = x_353_groups_0, pad = x_353_pad_0, pad_type = x_353_pad_type_0, strides = x_353_strides_0, weight = encoder_layers_15_conv_pointwise_conv2_weight, x = input_833)[name = tensor("x_353")]; tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; tensor input_835 = transpose(perm = input_835_perm_0, x = x_353)[name = tensor("transpose_176")]; tensor input_837 = add(x = input_819, y = input_835)[name = tensor("input_837")]; tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; tensor input_839 = layer_norm(axes = input_839_axes_0, beta = encoder_layers_15_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_15_norm_feed_forward2_weight, x = input_837)[name = tensor("input_839")]; tensor input_841 = linear(bias = encoder_layers_15_feed_forward2_linear1_bias, weight = encoder_layers_15_feed_forward2_linear1_weight, x = input_839)[name = tensor("linear_143")]; tensor input_843 = silu(x = input_841)[name = tensor("input_843")]; tensor input_847 = linear(bias = encoder_layers_15_feed_forward2_linear2_bias, weight = encoder_layers_15_feed_forward2_linear2_weight, x = input_843)[name = tensor("linear_144")]; tensor var_2991 = const()[name = tensor("op_2991"), val = tensor(0x1p-1)]; tensor var_2992 = mul(x = input_847, y = var_2991)[name = tensor("op_2992")]; tensor input_849 = add(x = input_837, y = var_2992)[name = tensor("input_849")]; tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; tensor input_851 = layer_norm(axes = input_851_axes_0, beta = encoder_layers_15_norm_out_bias, epsilon = var_11, gamma = encoder_layers_15_norm_out_weight, x = input_849)[name = tensor("input_851")]; tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; tensor input_853 = layer_norm(axes = input_853_axes_0, beta = encoder_layers_16_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_16_norm_feed_forward1_weight, x = input_851)[name = tensor("input_853")]; tensor input_855 = linear(bias = encoder_layers_16_feed_forward1_linear1_bias, weight = encoder_layers_16_feed_forward1_linear1_weight, x = input_853)[name = tensor("linear_145")]; tensor input_857 = silu(x = input_855)[name = tensor("input_857")]; tensor input_861 = linear(bias = encoder_layers_16_feed_forward1_linear2_bias, weight = encoder_layers_16_feed_forward1_linear2_weight, x = input_857)[name = tensor("linear_146")]; tensor var_3022 = const()[name = tensor("op_3022"), val = tensor(0x1p-1)]; tensor var_3023 = mul(x = input_861, y = var_3022)[name = tensor("op_3023")]; tensor input_863 = add(x = input_851, y = var_3023)[name = tensor("input_863")]; tensor query_33_axes_0 = const()[name = tensor("query_33_axes_0"), val = tensor([-1])]; tensor query_33 = layer_norm(axes = query_33_axes_0, beta = encoder_layers_16_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_16_norm_self_att_weight, x = input_863)[name = tensor("query_33")]; tensor var_3039 = linear(bias = encoder_layers_16_self_attn_linear_q_bias, weight = encoder_layers_16_self_attn_linear_q_weight, x = query_33)[name = tensor("linear_147")]; tensor var_3040 = const()[name = tensor("op_3040"), val = tensor([1, -1, 8, 128])]; tensor q_97 = reshape(shape = var_3040, x = var_3039)[name = tensor("q_97")]; tensor var_3044 = linear(bias = encoder_layers_16_self_attn_linear_k_bias, weight = encoder_layers_16_self_attn_linear_k_weight, x = query_33)[name = tensor("linear_148")]; tensor var_3045 = const()[name = tensor("op_3045"), val = tensor([1, -1, 8, 128])]; tensor k_65 = reshape(shape = var_3045, x = var_3044)[name = tensor("k_65")]; tensor var_3049 = linear(bias = encoder_layers_16_self_attn_linear_v_bias, weight = encoder_layers_16_self_attn_linear_v_weight, x = query_33)[name = tensor("linear_149")]; tensor var_3050 = const()[name = tensor("op_3050"), val = tensor([1, -1, 8, 128])]; tensor v_33 = reshape(shape = var_3050, x = var_3049)[name = tensor("v_33")]; tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3062 = add(x = q_97, y = encoder_layers_16_self_attn_pos_bias_u)[name = tensor("op_3062")]; tensor var_3064 = add(x = q_97, y = encoder_layers_16_self_attn_pos_bias_v)[name = tensor("op_3064")]; tensor q_with_bias_v_33_perm_0 = const()[name = tensor("q_with_bias_v_33_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3066 = const()[name = tensor("op_3066"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2358640640)))]; tensor x_361_transpose_x_0 = const()[name = tensor("x_361_transpose_x_0"), val = tensor(false)]; tensor x_361_transpose_y_0 = const()[name = tensor("x_361_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_33 = transpose(perm = q_with_bias_v_33_perm_0, x = var_3064)[name = tensor("transpose_174")]; tensor x_361 = matmul(transpose_x = x_361_transpose_x_0, transpose_y = x_361_transpose_y_0, x = q_with_bias_v_33, y = var_3066)[name = tensor("x_361")]; tensor const_174 = const()[name = tensor("const_174"), val = tensor(0x0p+0)]; tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_363_mode_0 = const()[name = tensor("x_363_mode_0"), val = tensor("constant")]; tensor x_363 = pad(constant_val = const_174, mode = x_363_mode_0, pad = x_363_pad_0, x = x_361)[name = tensor("x_363")]; tensor var_3074 = const()[name = tensor("op_3074"), val = tensor([1, 8, -1, 188])]; tensor x_365 = reshape(shape = var_3074, x = x_363)[name = tensor("x_365")]; tensor var_3078_begin_0 = const()[name = tensor("op_3078_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3078_end_0 = const()[name = tensor("op_3078_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3078_end_mask_0 = const()[name = tensor("op_3078_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3078 = slice_by_index(begin = var_3078_begin_0, end = var_3078_end_0, end_mask = var_3078_end_mask_0, x = x_365)[name = tensor("op_3078")]; tensor var_3079 = const()[name = tensor("op_3079"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_65 = reshape(shape = var_3079, x = var_3078)[name = tensor("matrix_bd_65")]; tensor matrix_ac_33_transpose_x_0 = const()[name = tensor("matrix_ac_33_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_33_transpose_y_0 = const()[name = tensor("matrix_ac_33_transpose_y_0"), val = tensor(false)]; tensor transpose_104_perm_0 = const()[name = tensor("transpose_104_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_105_perm_0 = const()[name = tensor("transpose_105_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_105 = transpose(perm = transpose_105_perm_0, x = k_65)[name = tensor("transpose_172")]; tensor transpose_104 = transpose(perm = transpose_104_perm_0, x = var_3062)[name = tensor("transpose_173")]; tensor matrix_ac_33 = matmul(transpose_x = matrix_ac_33_transpose_x_0, transpose_y = matrix_ac_33_transpose_y_0, x = transpose_104, y = transpose_105)[name = tensor("matrix_ac_33")]; tensor matrix_bd_67_begin_0 = const()[name = tensor("matrix_bd_67_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_67_end_0 = const()[name = tensor("matrix_bd_67_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_67_end_mask_0 = const()[name = tensor("matrix_bd_67_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_67 = slice_by_index(begin = matrix_bd_67_begin_0, end = matrix_bd_67_end_0, end_mask = matrix_bd_67_end_mask_0, x = matrix_bd_65)[name = tensor("matrix_bd_67")]; tensor var_3088 = add(x = matrix_ac_33, y = matrix_bd_67)[name = tensor("op_3088")]; tensor _inversed_scores_65_y_0 = const()[name = tensor("_inversed_scores_65_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_65 = mul(x = var_3088, y = _inversed_scores_65_y_0)[name = tensor("_inversed_scores_65")]; tensor scores_67 = select(a = var_14, b = _inversed_scores_65, cond = mask_3)[name = tensor("scores_67")]; tensor var_3094 = softmax(axis = var_32, x = scores_67)[name = tensor("op_3094")]; tensor input_865 = select(a = var_13, b = var_3094, cond = mask_3)[name = tensor("input_865")]; tensor x_367_transpose_x_0 = const()[name = tensor("x_367_transpose_x_0"), val = tensor(false)]; tensor x_367_transpose_y_0 = const()[name = tensor("x_367_transpose_y_0"), val = tensor(false)]; tensor value_35 = transpose(perm = value_35_perm_0, x = v_33)[name = tensor("transpose_175")]; tensor x_367 = matmul(transpose_x = x_367_transpose_x_0, transpose_y = x_367_transpose_y_0, x = input_865, y = value_35)[name = tensor("x_367")]; tensor var_3098_perm_0 = const()[name = tensor("op_3098_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3099 = const()[name = tensor("op_3099"), val = tensor([1, -1, 1024])]; tensor var_3098 = transpose(perm = var_3098_perm_0, x = x_367)[name = tensor("transpose_171")]; tensor input_867 = reshape(shape = var_3099, x = var_3098)[name = tensor("input_867")]; tensor input_869 = linear(bias = encoder_layers_16_self_attn_linear_out_bias, weight = encoder_layers_16_self_attn_linear_out_weight, x = input_867)[name = tensor("linear_151")]; tensor input_871 = add(x = input_863, y = input_869)[name = tensor("input_871")]; tensor x_371_axes_0 = const()[name = tensor("x_371_axes_0"), val = tensor([-1])]; tensor x_371 = layer_norm(axes = x_371_axes_0, beta = encoder_layers_16_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_16_norm_conv_weight, x = input_871)[name = tensor("x_371")]; tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; tensor input_873 = transpose(perm = input_873_perm_0, x = x_371)[name = tensor("transpose_170")]; tensor input_875 = conv(bias = encoder_layers_16_conv_pointwise_conv1_bias, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = encoder_layers_16_conv_pointwise_conv1_weight, x = input_873)[name = tensor("input_875")]; tensor x_373_split_num_splits_0 = const()[name = tensor("x_373_split_num_splits_0"), val = tensor(2)]; tensor x_373_split_axis_0 = const()[name = tensor("x_373_split_axis_0"), val = tensor(1)]; tensor x_373_split_0, tensor x_373_split_1 = split(axis = x_373_split_axis_0, num_splits = x_373_split_num_splits_0, x = input_875)[name = tensor("x_373_split")]; tensor x_373_split_1_sigmoid = sigmoid(x = x_373_split_1)[name = tensor("x_373_split_1_sigmoid")]; tensor x_373 = mul(x = x_373_split_0, y = x_373_split_1_sigmoid)[name = tensor("x_373")]; tensor input_877 = select(a = var_13, b = x_373, cond = var_339)[name = tensor("input_877")]; tensor const_177 = const()[name = tensor("const_177"), val = tensor(0x0p+0)]; tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; tensor input_879 = pad(constant_val = const_177, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877)[name = tensor("input_879")]; tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(1024)]; tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; tensor const_280 = const()[name = tensor("const_280"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2360176704)))]; tensor const_281 = const()[name = tensor("const_281"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2360213632)))]; tensor input_883 = conv(bias = const_281, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_280, x = input_879)[name = tensor("input_883")]; tensor input_885 = silu(x = input_883)[name = tensor("input_885")]; tensor x_375_pad_type_0 = const()[name = tensor("x_375_pad_type_0"), val = tensor("valid")]; tensor x_375_strides_0 = const()[name = tensor("x_375_strides_0"), val = tensor([1])]; tensor x_375_pad_0 = const()[name = tensor("x_375_pad_0"), val = tensor([0, 0])]; tensor x_375_dilations_0 = const()[name = tensor("x_375_dilations_0"), val = tensor([1])]; tensor x_375_groups_0 = const()[name = tensor("x_375_groups_0"), val = tensor(1)]; tensor x_375 = conv(bias = encoder_layers_16_conv_pointwise_conv2_bias, dilations = x_375_dilations_0, groups = x_375_groups_0, pad = x_375_pad_0, pad_type = x_375_pad_type_0, strides = x_375_strides_0, weight = encoder_layers_16_conv_pointwise_conv2_weight, x = input_885)[name = tensor("x_375")]; tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; tensor input_887 = transpose(perm = input_887_perm_0, x = x_375)[name = tensor("transpose_169")]; tensor input_889 = add(x = input_871, y = input_887)[name = tensor("input_889")]; tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; tensor input_891 = layer_norm(axes = input_891_axes_0, beta = encoder_layers_16_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_16_norm_feed_forward2_weight, x = input_889)[name = tensor("input_891")]; tensor input_893 = linear(bias = encoder_layers_16_feed_forward2_linear1_bias, weight = encoder_layers_16_feed_forward2_linear1_weight, x = input_891)[name = tensor("linear_152")]; tensor input_895 = silu(x = input_893)[name = tensor("input_895")]; tensor input_899 = linear(bias = encoder_layers_16_feed_forward2_linear2_bias, weight = encoder_layers_16_feed_forward2_linear2_weight, x = input_895)[name = tensor("linear_153")]; tensor var_3165 = const()[name = tensor("op_3165"), val = tensor(0x1p-1)]; tensor var_3166 = mul(x = input_899, y = var_3165)[name = tensor("op_3166")]; tensor input_901 = add(x = input_889, y = var_3166)[name = tensor("input_901")]; tensor input_903_axes_0 = const()[name = tensor("input_903_axes_0"), val = tensor([-1])]; tensor input_903 = layer_norm(axes = input_903_axes_0, beta = encoder_layers_16_norm_out_bias, epsilon = var_11, gamma = encoder_layers_16_norm_out_weight, x = input_901)[name = tensor("input_903")]; tensor input_905_axes_0 = const()[name = tensor("input_905_axes_0"), val = tensor([-1])]; tensor input_905 = layer_norm(axes = input_905_axes_0, beta = encoder_layers_17_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_17_norm_feed_forward1_weight, x = input_903)[name = tensor("input_905")]; tensor input_907 = linear(bias = encoder_layers_17_feed_forward1_linear1_bias, weight = encoder_layers_17_feed_forward1_linear1_weight, x = input_905)[name = tensor("linear_154")]; tensor input_909 = silu(x = input_907)[name = tensor("input_909")]; tensor input_913 = linear(bias = encoder_layers_17_feed_forward1_linear2_bias, weight = encoder_layers_17_feed_forward1_linear2_weight, x = input_909)[name = tensor("linear_155")]; tensor var_3196 = const()[name = tensor("op_3196"), val = tensor(0x1p-1)]; tensor var_3197 = mul(x = input_913, y = var_3196)[name = tensor("op_3197")]; tensor input_915 = add(x = input_903, y = var_3197)[name = tensor("input_915")]; tensor query_35_axes_0 = const()[name = tensor("query_35_axes_0"), val = tensor([-1])]; tensor query_35 = layer_norm(axes = query_35_axes_0, beta = encoder_layers_17_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_17_norm_self_att_weight, x = input_915)[name = tensor("query_35")]; tensor var_3213 = linear(bias = encoder_layers_17_self_attn_linear_q_bias, weight = encoder_layers_17_self_attn_linear_q_weight, x = query_35)[name = tensor("linear_156")]; tensor var_3214 = const()[name = tensor("op_3214"), val = tensor([1, -1, 8, 128])]; tensor q_103 = reshape(shape = var_3214, x = var_3213)[name = tensor("q_103")]; tensor var_3218 = linear(bias = encoder_layers_17_self_attn_linear_k_bias, weight = encoder_layers_17_self_attn_linear_k_weight, x = query_35)[name = tensor("linear_157")]; tensor var_3219 = const()[name = tensor("op_3219"), val = tensor([1, -1, 8, 128])]; tensor k_69 = reshape(shape = var_3219, x = var_3218)[name = tensor("k_69")]; tensor var_3223 = linear(bias = encoder_layers_17_self_attn_linear_v_bias, weight = encoder_layers_17_self_attn_linear_v_weight, x = query_35)[name = tensor("linear_158")]; tensor var_3224 = const()[name = tensor("op_3224"), val = tensor([1, -1, 8, 128])]; tensor v_35 = reshape(shape = var_3224, x = var_3223)[name = tensor("v_35")]; tensor value_37_perm_0 = const()[name = tensor("value_37_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3236 = add(x = q_103, y = encoder_layers_17_self_attn_pos_bias_u)[name = tensor("op_3236")]; tensor var_3238 = add(x = q_103, y = encoder_layers_17_self_attn_pos_bias_v)[name = tensor("op_3238")]; tensor q_with_bias_v_35_perm_0 = const()[name = tensor("q_with_bias_v_35_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3240 = const()[name = tensor("op_3240"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2360217792)))]; tensor x_383_transpose_x_0 = const()[name = tensor("x_383_transpose_x_0"), val = tensor(false)]; tensor x_383_transpose_y_0 = const()[name = tensor("x_383_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_35 = transpose(perm = q_with_bias_v_35_perm_0, x = var_3238)[name = tensor("transpose_167")]; tensor x_383 = matmul(transpose_x = x_383_transpose_x_0, transpose_y = x_383_transpose_y_0, x = q_with_bias_v_35, y = var_3240)[name = tensor("x_383")]; tensor const_184 = const()[name = tensor("const_184"), val = tensor(0x0p+0)]; tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_385_mode_0 = const()[name = tensor("x_385_mode_0"), val = tensor("constant")]; tensor x_385 = pad(constant_val = const_184, mode = x_385_mode_0, pad = x_385_pad_0, x = x_383)[name = tensor("x_385")]; tensor var_3248 = const()[name = tensor("op_3248"), val = tensor([1, 8, -1, 188])]; tensor x_387 = reshape(shape = var_3248, x = x_385)[name = tensor("x_387")]; tensor var_3252_begin_0 = const()[name = tensor("op_3252_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3252_end_0 = const()[name = tensor("op_3252_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3252_end_mask_0 = const()[name = tensor("op_3252_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3252 = slice_by_index(begin = var_3252_begin_0, end = var_3252_end_0, end_mask = var_3252_end_mask_0, x = x_387)[name = tensor("op_3252")]; tensor var_3253 = const()[name = tensor("op_3253"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_69 = reshape(shape = var_3253, x = var_3252)[name = tensor("matrix_bd_69")]; tensor matrix_ac_35_transpose_x_0 = const()[name = tensor("matrix_ac_35_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_35_transpose_y_0 = const()[name = tensor("matrix_ac_35_transpose_y_0"), val = tensor(false)]; tensor transpose_106_perm_0 = const()[name = tensor("transpose_106_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_107_perm_0 = const()[name = tensor("transpose_107_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_107 = transpose(perm = transpose_107_perm_0, x = k_69)[name = tensor("transpose_165")]; tensor transpose_106 = transpose(perm = transpose_106_perm_0, x = var_3236)[name = tensor("transpose_166")]; tensor matrix_ac_35 = matmul(transpose_x = matrix_ac_35_transpose_x_0, transpose_y = matrix_ac_35_transpose_y_0, x = transpose_106, y = transpose_107)[name = tensor("matrix_ac_35")]; tensor matrix_bd_71_begin_0 = const()[name = tensor("matrix_bd_71_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_71_end_0 = const()[name = tensor("matrix_bd_71_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_71_end_mask_0 = const()[name = tensor("matrix_bd_71_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_71 = slice_by_index(begin = matrix_bd_71_begin_0, end = matrix_bd_71_end_0, end_mask = matrix_bd_71_end_mask_0, x = matrix_bd_69)[name = tensor("matrix_bd_71")]; tensor var_3262 = add(x = matrix_ac_35, y = matrix_bd_71)[name = tensor("op_3262")]; tensor _inversed_scores_69_y_0 = const()[name = tensor("_inversed_scores_69_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_69 = mul(x = var_3262, y = _inversed_scores_69_y_0)[name = tensor("_inversed_scores_69")]; tensor scores_71 = select(a = var_14, b = _inversed_scores_69, cond = mask_3)[name = tensor("scores_71")]; tensor var_3268 = softmax(axis = var_32, x = scores_71)[name = tensor("op_3268")]; tensor input_917 = select(a = var_13, b = var_3268, cond = mask_3)[name = tensor("input_917")]; tensor x_389_transpose_x_0 = const()[name = tensor("x_389_transpose_x_0"), val = tensor(false)]; tensor x_389_transpose_y_0 = const()[name = tensor("x_389_transpose_y_0"), val = tensor(false)]; tensor value_37 = transpose(perm = value_37_perm_0, x = v_35)[name = tensor("transpose_168")]; tensor x_389 = matmul(transpose_x = x_389_transpose_x_0, transpose_y = x_389_transpose_y_0, x = input_917, y = value_37)[name = tensor("x_389")]; tensor var_3272_perm_0 = const()[name = tensor("op_3272_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3273 = const()[name = tensor("op_3273"), val = tensor([1, -1, 1024])]; tensor var_3272 = transpose(perm = var_3272_perm_0, x = x_389)[name = tensor("transpose_164")]; tensor input_919 = reshape(shape = var_3273, x = var_3272)[name = tensor("input_919")]; tensor input_921 = linear(bias = encoder_layers_17_self_attn_linear_out_bias, weight = encoder_layers_17_self_attn_linear_out_weight, x = input_919)[name = tensor("linear_160")]; tensor input_923 = add(x = input_915, y = input_921)[name = tensor("input_923")]; tensor x_393_axes_0 = const()[name = tensor("x_393_axes_0"), val = tensor([-1])]; tensor x_393 = layer_norm(axes = x_393_axes_0, beta = encoder_layers_17_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_17_norm_conv_weight, x = input_923)[name = tensor("x_393")]; tensor input_925_perm_0 = const()[name = tensor("input_925_perm_0"), val = tensor([0, 2, 1])]; tensor input_927_pad_type_0 = const()[name = tensor("input_927_pad_type_0"), val = tensor("valid")]; tensor input_927_strides_0 = const()[name = tensor("input_927_strides_0"), val = tensor([1])]; tensor input_927_pad_0 = const()[name = tensor("input_927_pad_0"), val = tensor([0, 0])]; tensor input_927_dilations_0 = const()[name = tensor("input_927_dilations_0"), val = tensor([1])]; tensor input_927_groups_0 = const()[name = tensor("input_927_groups_0"), val = tensor(1)]; tensor input_925 = transpose(perm = input_925_perm_0, x = x_393)[name = tensor("transpose_163")]; tensor input_927 = conv(bias = encoder_layers_17_conv_pointwise_conv1_bias, dilations = input_927_dilations_0, groups = input_927_groups_0, pad = input_927_pad_0, pad_type = input_927_pad_type_0, strides = input_927_strides_0, weight = encoder_layers_17_conv_pointwise_conv1_weight, x = input_925)[name = tensor("input_927")]; tensor x_395_split_num_splits_0 = const()[name = tensor("x_395_split_num_splits_0"), val = tensor(2)]; tensor x_395_split_axis_0 = const()[name = tensor("x_395_split_axis_0"), val = tensor(1)]; tensor x_395_split_0, tensor x_395_split_1 = split(axis = x_395_split_axis_0, num_splits = x_395_split_num_splits_0, x = input_927)[name = tensor("x_395_split")]; tensor x_395_split_1_sigmoid = sigmoid(x = x_395_split_1)[name = tensor("x_395_split_1_sigmoid")]; tensor x_395 = mul(x = x_395_split_0, y = x_395_split_1_sigmoid)[name = tensor("x_395")]; tensor input_929 = select(a = var_13, b = x_395, cond = var_339)[name = tensor("input_929")]; tensor const_187 = const()[name = tensor("const_187"), val = tensor(0x0p+0)]; tensor input_931_pad_0 = const()[name = tensor("input_931_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_931_mode_0 = const()[name = tensor("input_931_mode_0"), val = tensor("constant")]; tensor input_931 = pad(constant_val = const_187, mode = input_931_mode_0, pad = input_931_pad_0, x = input_929)[name = tensor("input_931")]; tensor input_933_pad_type_0 = const()[name = tensor("input_933_pad_type_0"), val = tensor("valid")]; tensor input_933_groups_0 = const()[name = tensor("input_933_groups_0"), val = tensor(1024)]; tensor input_933_strides_0 = const()[name = tensor("input_933_strides_0"), val = tensor([1])]; tensor input_933_pad_0 = const()[name = tensor("input_933_pad_0"), val = tensor([0, 0])]; tensor input_933_dilations_0 = const()[name = tensor("input_933_dilations_0"), val = tensor([1])]; tensor const_282 = const()[name = tensor("const_282"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2361753856)))]; tensor const_283 = const()[name = tensor("const_283"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2361790784)))]; tensor input_935 = conv(bias = const_283, dilations = input_933_dilations_0, groups = input_933_groups_0, pad = input_933_pad_0, pad_type = input_933_pad_type_0, strides = input_933_strides_0, weight = const_282, x = input_931)[name = tensor("input_935")]; tensor input_937 = silu(x = input_935)[name = tensor("input_937")]; tensor x_397_pad_type_0 = const()[name = tensor("x_397_pad_type_0"), val = tensor("valid")]; tensor x_397_strides_0 = const()[name = tensor("x_397_strides_0"), val = tensor([1])]; tensor x_397_pad_0 = const()[name = tensor("x_397_pad_0"), val = tensor([0, 0])]; tensor x_397_dilations_0 = const()[name = tensor("x_397_dilations_0"), val = tensor([1])]; tensor x_397_groups_0 = const()[name = tensor("x_397_groups_0"), val = tensor(1)]; tensor x_397 = conv(bias = encoder_layers_17_conv_pointwise_conv2_bias, dilations = x_397_dilations_0, groups = x_397_groups_0, pad = x_397_pad_0, pad_type = x_397_pad_type_0, strides = x_397_strides_0, weight = encoder_layers_17_conv_pointwise_conv2_weight, x = input_937)[name = tensor("x_397")]; tensor input_939_perm_0 = const()[name = tensor("input_939_perm_0"), val = tensor([0, 2, 1])]; tensor input_939 = transpose(perm = input_939_perm_0, x = x_397)[name = tensor("transpose_162")]; tensor input_941 = add(x = input_923, y = input_939)[name = tensor("input_941")]; tensor input_943_axes_0 = const()[name = tensor("input_943_axes_0"), val = tensor([-1])]; tensor input_943 = layer_norm(axes = input_943_axes_0, beta = encoder_layers_17_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_17_norm_feed_forward2_weight, x = input_941)[name = tensor("input_943")]; tensor input_945 = linear(bias = encoder_layers_17_feed_forward2_linear1_bias, weight = encoder_layers_17_feed_forward2_linear1_weight, x = input_943)[name = tensor("linear_161")]; tensor input_947 = silu(x = input_945)[name = tensor("input_947")]; tensor input_951 = linear(bias = encoder_layers_17_feed_forward2_linear2_bias, weight = encoder_layers_17_feed_forward2_linear2_weight, x = input_947)[name = tensor("linear_162")]; tensor var_3339 = const()[name = tensor("op_3339"), val = tensor(0x1p-1)]; tensor var_3340 = mul(x = input_951, y = var_3339)[name = tensor("op_3340")]; tensor input_953 = add(x = input_941, y = var_3340)[name = tensor("input_953")]; tensor input_955_axes_0 = const()[name = tensor("input_955_axes_0"), val = tensor([-1])]; tensor input_955 = layer_norm(axes = input_955_axes_0, beta = encoder_layers_17_norm_out_bias, epsilon = var_11, gamma = encoder_layers_17_norm_out_weight, x = input_953)[name = tensor("input_955")]; tensor input_957_axes_0 = const()[name = tensor("input_957_axes_0"), val = tensor([-1])]; tensor input_957 = layer_norm(axes = input_957_axes_0, beta = encoder_layers_18_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_18_norm_feed_forward1_weight, x = input_955)[name = tensor("input_957")]; tensor input_959 = linear(bias = encoder_layers_18_feed_forward1_linear1_bias, weight = encoder_layers_18_feed_forward1_linear1_weight, x = input_957)[name = tensor("linear_163")]; tensor input_961 = silu(x = input_959)[name = tensor("input_961")]; tensor input_965 = linear(bias = encoder_layers_18_feed_forward1_linear2_bias, weight = encoder_layers_18_feed_forward1_linear2_weight, x = input_961)[name = tensor("linear_164")]; tensor var_3370 = const()[name = tensor("op_3370"), val = tensor(0x1p-1)]; tensor var_3371 = mul(x = input_965, y = var_3370)[name = tensor("op_3371")]; tensor input_967 = add(x = input_955, y = var_3371)[name = tensor("input_967")]; tensor query_37_axes_0 = const()[name = tensor("query_37_axes_0"), val = tensor([-1])]; tensor query_37 = layer_norm(axes = query_37_axes_0, beta = encoder_layers_18_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_18_norm_self_att_weight, x = input_967)[name = tensor("query_37")]; tensor var_3387 = linear(bias = encoder_layers_18_self_attn_linear_q_bias, weight = encoder_layers_18_self_attn_linear_q_weight, x = query_37)[name = tensor("linear_165")]; tensor var_3388 = const()[name = tensor("op_3388"), val = tensor([1, -1, 8, 128])]; tensor q_109 = reshape(shape = var_3388, x = var_3387)[name = tensor("q_109")]; tensor var_3392 = linear(bias = encoder_layers_18_self_attn_linear_k_bias, weight = encoder_layers_18_self_attn_linear_k_weight, x = query_37)[name = tensor("linear_166")]; tensor var_3393 = const()[name = tensor("op_3393"), val = tensor([1, -1, 8, 128])]; tensor k_73 = reshape(shape = var_3393, x = var_3392)[name = tensor("k_73")]; tensor var_3397 = linear(bias = encoder_layers_18_self_attn_linear_v_bias, weight = encoder_layers_18_self_attn_linear_v_weight, x = query_37)[name = tensor("linear_167")]; tensor var_3398 = const()[name = tensor("op_3398"), val = tensor([1, -1, 8, 128])]; tensor v_37 = reshape(shape = var_3398, x = var_3397)[name = tensor("v_37")]; tensor value_39_perm_0 = const()[name = tensor("value_39_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3410 = add(x = q_109, y = encoder_layers_18_self_attn_pos_bias_u)[name = tensor("op_3410")]; tensor var_3412 = add(x = q_109, y = encoder_layers_18_self_attn_pos_bias_v)[name = tensor("op_3412")]; tensor q_with_bias_v_37_perm_0 = const()[name = tensor("q_with_bias_v_37_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3414 = const()[name = tensor("op_3414"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2361794944)))]; tensor x_405_transpose_x_0 = const()[name = tensor("x_405_transpose_x_0"), val = tensor(false)]; tensor x_405_transpose_y_0 = const()[name = tensor("x_405_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_37 = transpose(perm = q_with_bias_v_37_perm_0, x = var_3412)[name = tensor("transpose_160")]; tensor x_405 = matmul(transpose_x = x_405_transpose_x_0, transpose_y = x_405_transpose_y_0, x = q_with_bias_v_37, y = var_3414)[name = tensor("x_405")]; tensor const_194 = const()[name = tensor("const_194"), val = tensor(0x0p+0)]; tensor x_407_pad_0 = const()[name = tensor("x_407_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_407_mode_0 = const()[name = tensor("x_407_mode_0"), val = tensor("constant")]; tensor x_407 = pad(constant_val = const_194, mode = x_407_mode_0, pad = x_407_pad_0, x = x_405)[name = tensor("x_407")]; tensor var_3422 = const()[name = tensor("op_3422"), val = tensor([1, 8, -1, 188])]; tensor x_409 = reshape(shape = var_3422, x = x_407)[name = tensor("x_409")]; tensor var_3426_begin_0 = const()[name = tensor("op_3426_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3426_end_0 = const()[name = tensor("op_3426_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3426_end_mask_0 = const()[name = tensor("op_3426_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3426 = slice_by_index(begin = var_3426_begin_0, end = var_3426_end_0, end_mask = var_3426_end_mask_0, x = x_409)[name = tensor("op_3426")]; tensor var_3427 = const()[name = tensor("op_3427"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_73 = reshape(shape = var_3427, x = var_3426)[name = tensor("matrix_bd_73")]; tensor matrix_ac_37_transpose_x_0 = const()[name = tensor("matrix_ac_37_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_37_transpose_y_0 = const()[name = tensor("matrix_ac_37_transpose_y_0"), val = tensor(false)]; tensor transpose_108_perm_0 = const()[name = tensor("transpose_108_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_109_perm_0 = const()[name = tensor("transpose_109_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_109 = transpose(perm = transpose_109_perm_0, x = k_73)[name = tensor("transpose_158")]; tensor transpose_108 = transpose(perm = transpose_108_perm_0, x = var_3410)[name = tensor("transpose_159")]; tensor matrix_ac_37 = matmul(transpose_x = matrix_ac_37_transpose_x_0, transpose_y = matrix_ac_37_transpose_y_0, x = transpose_108, y = transpose_109)[name = tensor("matrix_ac_37")]; tensor matrix_bd_75_begin_0 = const()[name = tensor("matrix_bd_75_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_75_end_0 = const()[name = tensor("matrix_bd_75_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_75_end_mask_0 = const()[name = tensor("matrix_bd_75_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_75 = slice_by_index(begin = matrix_bd_75_begin_0, end = matrix_bd_75_end_0, end_mask = matrix_bd_75_end_mask_0, x = matrix_bd_73)[name = tensor("matrix_bd_75")]; tensor var_3436 = add(x = matrix_ac_37, y = matrix_bd_75)[name = tensor("op_3436")]; tensor _inversed_scores_73_y_0 = const()[name = tensor("_inversed_scores_73_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_73 = mul(x = var_3436, y = _inversed_scores_73_y_0)[name = tensor("_inversed_scores_73")]; tensor scores_75 = select(a = var_14, b = _inversed_scores_73, cond = mask_3)[name = tensor("scores_75")]; tensor var_3442 = softmax(axis = var_32, x = scores_75)[name = tensor("op_3442")]; tensor input_969 = select(a = var_13, b = var_3442, cond = mask_3)[name = tensor("input_969")]; tensor x_411_transpose_x_0 = const()[name = tensor("x_411_transpose_x_0"), val = tensor(false)]; tensor x_411_transpose_y_0 = const()[name = tensor("x_411_transpose_y_0"), val = tensor(false)]; tensor value_39 = transpose(perm = value_39_perm_0, x = v_37)[name = tensor("transpose_161")]; tensor x_411 = matmul(transpose_x = x_411_transpose_x_0, transpose_y = x_411_transpose_y_0, x = input_969, y = value_39)[name = tensor("x_411")]; tensor var_3446_perm_0 = const()[name = tensor("op_3446_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3447 = const()[name = tensor("op_3447"), val = tensor([1, -1, 1024])]; tensor var_3446 = transpose(perm = var_3446_perm_0, x = x_411)[name = tensor("transpose_157")]; tensor input_971 = reshape(shape = var_3447, x = var_3446)[name = tensor("input_971")]; tensor input_973 = linear(bias = encoder_layers_18_self_attn_linear_out_bias, weight = encoder_layers_18_self_attn_linear_out_weight, x = input_971)[name = tensor("linear_169")]; tensor input_975 = add(x = input_967, y = input_973)[name = tensor("input_975")]; tensor x_415_axes_0 = const()[name = tensor("x_415_axes_0"), val = tensor([-1])]; tensor x_415 = layer_norm(axes = x_415_axes_0, beta = encoder_layers_18_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_18_norm_conv_weight, x = input_975)[name = tensor("x_415")]; tensor input_977_perm_0 = const()[name = tensor("input_977_perm_0"), val = tensor([0, 2, 1])]; tensor input_979_pad_type_0 = const()[name = tensor("input_979_pad_type_0"), val = tensor("valid")]; tensor input_979_strides_0 = const()[name = tensor("input_979_strides_0"), val = tensor([1])]; tensor input_979_pad_0 = const()[name = tensor("input_979_pad_0"), val = tensor([0, 0])]; tensor input_979_dilations_0 = const()[name = tensor("input_979_dilations_0"), val = tensor([1])]; tensor input_979_groups_0 = const()[name = tensor("input_979_groups_0"), val = tensor(1)]; tensor input_977 = transpose(perm = input_977_perm_0, x = x_415)[name = tensor("transpose_156")]; tensor input_979 = conv(bias = encoder_layers_18_conv_pointwise_conv1_bias, dilations = input_979_dilations_0, groups = input_979_groups_0, pad = input_979_pad_0, pad_type = input_979_pad_type_0, strides = input_979_strides_0, weight = encoder_layers_18_conv_pointwise_conv1_weight, x = input_977)[name = tensor("input_979")]; tensor x_417_split_num_splits_0 = const()[name = tensor("x_417_split_num_splits_0"), val = tensor(2)]; tensor x_417_split_axis_0 = const()[name = tensor("x_417_split_axis_0"), val = tensor(1)]; tensor x_417_split_0, tensor x_417_split_1 = split(axis = x_417_split_axis_0, num_splits = x_417_split_num_splits_0, x = input_979)[name = tensor("x_417_split")]; tensor x_417_split_1_sigmoid = sigmoid(x = x_417_split_1)[name = tensor("x_417_split_1_sigmoid")]; tensor x_417 = mul(x = x_417_split_0, y = x_417_split_1_sigmoid)[name = tensor("x_417")]; tensor input_981 = select(a = var_13, b = x_417, cond = var_339)[name = tensor("input_981")]; tensor const_197 = const()[name = tensor("const_197"), val = tensor(0x0p+0)]; tensor input_983_pad_0 = const()[name = tensor("input_983_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_983_mode_0 = const()[name = tensor("input_983_mode_0"), val = tensor("constant")]; tensor input_983 = pad(constant_val = const_197, mode = input_983_mode_0, pad = input_983_pad_0, x = input_981)[name = tensor("input_983")]; tensor input_985_pad_type_0 = const()[name = tensor("input_985_pad_type_0"), val = tensor("valid")]; tensor input_985_groups_0 = const()[name = tensor("input_985_groups_0"), val = tensor(1024)]; tensor input_985_strides_0 = const()[name = tensor("input_985_strides_0"), val = tensor([1])]; tensor input_985_pad_0 = const()[name = tensor("input_985_pad_0"), val = tensor([0, 0])]; tensor input_985_dilations_0 = const()[name = tensor("input_985_dilations_0"), val = tensor([1])]; tensor const_284 = const()[name = tensor("const_284"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2363331008)))]; tensor const_285 = const()[name = tensor("const_285"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2363367936)))]; tensor input_987 = conv(bias = const_285, dilations = input_985_dilations_0, groups = input_985_groups_0, pad = input_985_pad_0, pad_type = input_985_pad_type_0, strides = input_985_strides_0, weight = const_284, x = input_983)[name = tensor("input_987")]; tensor input_989 = silu(x = input_987)[name = tensor("input_989")]; tensor x_419_pad_type_0 = const()[name = tensor("x_419_pad_type_0"), val = tensor("valid")]; tensor x_419_strides_0 = const()[name = tensor("x_419_strides_0"), val = tensor([1])]; tensor x_419_pad_0 = const()[name = tensor("x_419_pad_0"), val = tensor([0, 0])]; tensor x_419_dilations_0 = const()[name = tensor("x_419_dilations_0"), val = tensor([1])]; tensor x_419_groups_0 = const()[name = tensor("x_419_groups_0"), val = tensor(1)]; tensor x_419 = conv(bias = encoder_layers_18_conv_pointwise_conv2_bias, dilations = x_419_dilations_0, groups = x_419_groups_0, pad = x_419_pad_0, pad_type = x_419_pad_type_0, strides = x_419_strides_0, weight = encoder_layers_18_conv_pointwise_conv2_weight, x = input_989)[name = tensor("x_419")]; tensor input_991_perm_0 = const()[name = tensor("input_991_perm_0"), val = tensor([0, 2, 1])]; tensor input_991 = transpose(perm = input_991_perm_0, x = x_419)[name = tensor("transpose_155")]; tensor input_993 = add(x = input_975, y = input_991)[name = tensor("input_993")]; tensor input_995_axes_0 = const()[name = tensor("input_995_axes_0"), val = tensor([-1])]; tensor input_995 = layer_norm(axes = input_995_axes_0, beta = encoder_layers_18_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_18_norm_feed_forward2_weight, x = input_993)[name = tensor("input_995")]; tensor input_997 = linear(bias = encoder_layers_18_feed_forward2_linear1_bias, weight = encoder_layers_18_feed_forward2_linear1_weight, x = input_995)[name = tensor("linear_170")]; tensor input_999 = silu(x = input_997)[name = tensor("input_999")]; tensor input_1003 = linear(bias = encoder_layers_18_feed_forward2_linear2_bias, weight = encoder_layers_18_feed_forward2_linear2_weight, x = input_999)[name = tensor("linear_171")]; tensor var_3513 = const()[name = tensor("op_3513"), val = tensor(0x1p-1)]; tensor var_3514 = mul(x = input_1003, y = var_3513)[name = tensor("op_3514")]; tensor input_1005 = add(x = input_993, y = var_3514)[name = tensor("input_1005")]; tensor input_1007_axes_0 = const()[name = tensor("input_1007_axes_0"), val = tensor([-1])]; tensor input_1007 = layer_norm(axes = input_1007_axes_0, beta = encoder_layers_18_norm_out_bias, epsilon = var_11, gamma = encoder_layers_18_norm_out_weight, x = input_1005)[name = tensor("input_1007")]; tensor input_1009_axes_0 = const()[name = tensor("input_1009_axes_0"), val = tensor([-1])]; tensor input_1009 = layer_norm(axes = input_1009_axes_0, beta = encoder_layers_19_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_19_norm_feed_forward1_weight, x = input_1007)[name = tensor("input_1009")]; tensor input_1011 = linear(bias = encoder_layers_19_feed_forward1_linear1_bias, weight = encoder_layers_19_feed_forward1_linear1_weight, x = input_1009)[name = tensor("linear_172")]; tensor input_1013 = silu(x = input_1011)[name = tensor("input_1013")]; tensor input_1017 = linear(bias = encoder_layers_19_feed_forward1_linear2_bias, weight = encoder_layers_19_feed_forward1_linear2_weight, x = input_1013)[name = tensor("linear_173")]; tensor var_3544 = const()[name = tensor("op_3544"), val = tensor(0x1p-1)]; tensor var_3545 = mul(x = input_1017, y = var_3544)[name = tensor("op_3545")]; tensor input_1019 = add(x = input_1007, y = var_3545)[name = tensor("input_1019")]; tensor query_39_axes_0 = const()[name = tensor("query_39_axes_0"), val = tensor([-1])]; tensor query_39 = layer_norm(axes = query_39_axes_0, beta = encoder_layers_19_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_19_norm_self_att_weight, x = input_1019)[name = tensor("query_39")]; tensor var_3561 = linear(bias = encoder_layers_19_self_attn_linear_q_bias, weight = encoder_layers_19_self_attn_linear_q_weight, x = query_39)[name = tensor("linear_174")]; tensor var_3562 = const()[name = tensor("op_3562"), val = tensor([1, -1, 8, 128])]; tensor q_115 = reshape(shape = var_3562, x = var_3561)[name = tensor("q_115")]; tensor var_3566 = linear(bias = encoder_layers_19_self_attn_linear_k_bias, weight = encoder_layers_19_self_attn_linear_k_weight, x = query_39)[name = tensor("linear_175")]; tensor var_3567 = const()[name = tensor("op_3567"), val = tensor([1, -1, 8, 128])]; tensor k_77 = reshape(shape = var_3567, x = var_3566)[name = tensor("k_77")]; tensor var_3571 = linear(bias = encoder_layers_19_self_attn_linear_v_bias, weight = encoder_layers_19_self_attn_linear_v_weight, x = query_39)[name = tensor("linear_176")]; tensor var_3572 = const()[name = tensor("op_3572"), val = tensor([1, -1, 8, 128])]; tensor v_39 = reshape(shape = var_3572, x = var_3571)[name = tensor("v_39")]; tensor value_41_perm_0 = const()[name = tensor("value_41_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3584 = add(x = q_115, y = encoder_layers_19_self_attn_pos_bias_u)[name = tensor("op_3584")]; tensor var_3586 = add(x = q_115, y = encoder_layers_19_self_attn_pos_bias_v)[name = tensor("op_3586")]; tensor q_with_bias_v_39_perm_0 = const()[name = tensor("q_with_bias_v_39_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3588 = const()[name = tensor("op_3588"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2363372096)))]; tensor x_427_transpose_x_0 = const()[name = tensor("x_427_transpose_x_0"), val = tensor(false)]; tensor x_427_transpose_y_0 = const()[name = tensor("x_427_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_39 = transpose(perm = q_with_bias_v_39_perm_0, x = var_3586)[name = tensor("transpose_153")]; tensor x_427 = matmul(transpose_x = x_427_transpose_x_0, transpose_y = x_427_transpose_y_0, x = q_with_bias_v_39, y = var_3588)[name = tensor("x_427")]; tensor const_204 = const()[name = tensor("const_204"), val = tensor(0x0p+0)]; tensor x_429_pad_0 = const()[name = tensor("x_429_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_429_mode_0 = const()[name = tensor("x_429_mode_0"), val = tensor("constant")]; tensor x_429 = pad(constant_val = const_204, mode = x_429_mode_0, pad = x_429_pad_0, x = x_427)[name = tensor("x_429")]; tensor var_3596 = const()[name = tensor("op_3596"), val = tensor([1, 8, -1, 188])]; tensor x_431 = reshape(shape = var_3596, x = x_429)[name = tensor("x_431")]; tensor var_3600_begin_0 = const()[name = tensor("op_3600_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3600_end_0 = const()[name = tensor("op_3600_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3600_end_mask_0 = const()[name = tensor("op_3600_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3600 = slice_by_index(begin = var_3600_begin_0, end = var_3600_end_0, end_mask = var_3600_end_mask_0, x = x_431)[name = tensor("op_3600")]; tensor var_3601 = const()[name = tensor("op_3601"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_77 = reshape(shape = var_3601, x = var_3600)[name = tensor("matrix_bd_77")]; tensor matrix_ac_39_transpose_x_0 = const()[name = tensor("matrix_ac_39_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_39_transpose_y_0 = const()[name = tensor("matrix_ac_39_transpose_y_0"), val = tensor(false)]; tensor transpose_110_perm_0 = const()[name = tensor("transpose_110_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_111_perm_0 = const()[name = tensor("transpose_111_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_111 = transpose(perm = transpose_111_perm_0, x = k_77)[name = tensor("transpose_151")]; tensor transpose_110 = transpose(perm = transpose_110_perm_0, x = var_3584)[name = tensor("transpose_152")]; tensor matrix_ac_39 = matmul(transpose_x = matrix_ac_39_transpose_x_0, transpose_y = matrix_ac_39_transpose_y_0, x = transpose_110, y = transpose_111)[name = tensor("matrix_ac_39")]; tensor matrix_bd_79_begin_0 = const()[name = tensor("matrix_bd_79_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_79_end_0 = const()[name = tensor("matrix_bd_79_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_79_end_mask_0 = const()[name = tensor("matrix_bd_79_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_79 = slice_by_index(begin = matrix_bd_79_begin_0, end = matrix_bd_79_end_0, end_mask = matrix_bd_79_end_mask_0, x = matrix_bd_77)[name = tensor("matrix_bd_79")]; tensor var_3610 = add(x = matrix_ac_39, y = matrix_bd_79)[name = tensor("op_3610")]; tensor _inversed_scores_77_y_0 = const()[name = tensor("_inversed_scores_77_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_77 = mul(x = var_3610, y = _inversed_scores_77_y_0)[name = tensor("_inversed_scores_77")]; tensor scores_79 = select(a = var_14, b = _inversed_scores_77, cond = mask_3)[name = tensor("scores_79")]; tensor var_3616 = softmax(axis = var_32, x = scores_79)[name = tensor("op_3616")]; tensor input_1021 = select(a = var_13, b = var_3616, cond = mask_3)[name = tensor("input_1021")]; tensor x_433_transpose_x_0 = const()[name = tensor("x_433_transpose_x_0"), val = tensor(false)]; tensor x_433_transpose_y_0 = const()[name = tensor("x_433_transpose_y_0"), val = tensor(false)]; tensor value_41 = transpose(perm = value_41_perm_0, x = v_39)[name = tensor("transpose_154")]; tensor x_433 = matmul(transpose_x = x_433_transpose_x_0, transpose_y = x_433_transpose_y_0, x = input_1021, y = value_41)[name = tensor("x_433")]; tensor var_3620_perm_0 = const()[name = tensor("op_3620_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3621 = const()[name = tensor("op_3621"), val = tensor([1, -1, 1024])]; tensor var_3620 = transpose(perm = var_3620_perm_0, x = x_433)[name = tensor("transpose_150")]; tensor input_1023 = reshape(shape = var_3621, x = var_3620)[name = tensor("input_1023")]; tensor input_1025 = linear(bias = encoder_layers_19_self_attn_linear_out_bias, weight = encoder_layers_19_self_attn_linear_out_weight, x = input_1023)[name = tensor("linear_178")]; tensor input_1027 = add(x = input_1019, y = input_1025)[name = tensor("input_1027")]; tensor x_437_axes_0 = const()[name = tensor("x_437_axes_0"), val = tensor([-1])]; tensor x_437 = layer_norm(axes = x_437_axes_0, beta = encoder_layers_19_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_19_norm_conv_weight, x = input_1027)[name = tensor("x_437")]; tensor input_1029_perm_0 = const()[name = tensor("input_1029_perm_0"), val = tensor([0, 2, 1])]; tensor input_1031_pad_type_0 = const()[name = tensor("input_1031_pad_type_0"), val = tensor("valid")]; tensor input_1031_strides_0 = const()[name = tensor("input_1031_strides_0"), val = tensor([1])]; tensor input_1031_pad_0 = const()[name = tensor("input_1031_pad_0"), val = tensor([0, 0])]; tensor input_1031_dilations_0 = const()[name = tensor("input_1031_dilations_0"), val = tensor([1])]; tensor input_1031_groups_0 = const()[name = tensor("input_1031_groups_0"), val = tensor(1)]; tensor input_1029 = transpose(perm = input_1029_perm_0, x = x_437)[name = tensor("transpose_149")]; tensor input_1031 = conv(bias = encoder_layers_19_conv_pointwise_conv1_bias, dilations = input_1031_dilations_0, groups = input_1031_groups_0, pad = input_1031_pad_0, pad_type = input_1031_pad_type_0, strides = input_1031_strides_0, weight = encoder_layers_19_conv_pointwise_conv1_weight, x = input_1029)[name = tensor("input_1031")]; tensor x_439_split_num_splits_0 = const()[name = tensor("x_439_split_num_splits_0"), val = tensor(2)]; tensor x_439_split_axis_0 = const()[name = tensor("x_439_split_axis_0"), val = tensor(1)]; tensor x_439_split_0, tensor x_439_split_1 = split(axis = x_439_split_axis_0, num_splits = x_439_split_num_splits_0, x = input_1031)[name = tensor("x_439_split")]; tensor x_439_split_1_sigmoid = sigmoid(x = x_439_split_1)[name = tensor("x_439_split_1_sigmoid")]; tensor x_439 = mul(x = x_439_split_0, y = x_439_split_1_sigmoid)[name = tensor("x_439")]; tensor input_1033 = select(a = var_13, b = x_439, cond = var_339)[name = tensor("input_1033")]; tensor const_207 = const()[name = tensor("const_207"), val = tensor(0x0p+0)]; tensor input_1035_pad_0 = const()[name = tensor("input_1035_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1035_mode_0 = const()[name = tensor("input_1035_mode_0"), val = tensor("constant")]; tensor input_1035 = pad(constant_val = const_207, mode = input_1035_mode_0, pad = input_1035_pad_0, x = input_1033)[name = tensor("input_1035")]; tensor input_1037_pad_type_0 = const()[name = tensor("input_1037_pad_type_0"), val = tensor("valid")]; tensor input_1037_groups_0 = const()[name = tensor("input_1037_groups_0"), val = tensor(1024)]; tensor input_1037_strides_0 = const()[name = tensor("input_1037_strides_0"), val = tensor([1])]; tensor input_1037_pad_0 = const()[name = tensor("input_1037_pad_0"), val = tensor([0, 0])]; tensor input_1037_dilations_0 = const()[name = tensor("input_1037_dilations_0"), val = tensor([1])]; tensor const_286 = const()[name = tensor("const_286"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2364908160)))]; tensor const_287 = const()[name = tensor("const_287"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2364945088)))]; tensor input_1039 = conv(bias = const_287, dilations = input_1037_dilations_0, groups = input_1037_groups_0, pad = input_1037_pad_0, pad_type = input_1037_pad_type_0, strides = input_1037_strides_0, weight = const_286, x = input_1035)[name = tensor("input_1039")]; tensor input_1041 = silu(x = input_1039)[name = tensor("input_1041")]; tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; tensor x_441 = conv(bias = encoder_layers_19_conv_pointwise_conv2_bias, dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = encoder_layers_19_conv_pointwise_conv2_weight, x = input_1041)[name = tensor("x_441")]; tensor input_1043_perm_0 = const()[name = tensor("input_1043_perm_0"), val = tensor([0, 2, 1])]; tensor input_1043 = transpose(perm = input_1043_perm_0, x = x_441)[name = tensor("transpose_148")]; tensor input_1045 = add(x = input_1027, y = input_1043)[name = tensor("input_1045")]; tensor input_1047_axes_0 = const()[name = tensor("input_1047_axes_0"), val = tensor([-1])]; tensor input_1047 = layer_norm(axes = input_1047_axes_0, beta = encoder_layers_19_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_19_norm_feed_forward2_weight, x = input_1045)[name = tensor("input_1047")]; tensor input_1049 = linear(bias = encoder_layers_19_feed_forward2_linear1_bias, weight = encoder_layers_19_feed_forward2_linear1_weight, x = input_1047)[name = tensor("linear_179")]; tensor input_1051 = silu(x = input_1049)[name = tensor("input_1051")]; tensor input_1055 = linear(bias = encoder_layers_19_feed_forward2_linear2_bias, weight = encoder_layers_19_feed_forward2_linear2_weight, x = input_1051)[name = tensor("linear_180")]; tensor var_3687 = const()[name = tensor("op_3687"), val = tensor(0x1p-1)]; tensor var_3688 = mul(x = input_1055, y = var_3687)[name = tensor("op_3688")]; tensor input_1057 = add(x = input_1045, y = var_3688)[name = tensor("input_1057")]; tensor input_1059_axes_0 = const()[name = tensor("input_1059_axes_0"), val = tensor([-1])]; tensor input_1059 = layer_norm(axes = input_1059_axes_0, beta = encoder_layers_19_norm_out_bias, epsilon = var_11, gamma = encoder_layers_19_norm_out_weight, x = input_1057)[name = tensor("input_1059")]; tensor input_1061_axes_0 = const()[name = tensor("input_1061_axes_0"), val = tensor([-1])]; tensor input_1061 = layer_norm(axes = input_1061_axes_0, beta = encoder_layers_20_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_20_norm_feed_forward1_weight, x = input_1059)[name = tensor("input_1061")]; tensor input_1063 = linear(bias = encoder_layers_20_feed_forward1_linear1_bias, weight = encoder_layers_20_feed_forward1_linear1_weight, x = input_1061)[name = tensor("linear_181")]; tensor input_1065 = silu(x = input_1063)[name = tensor("input_1065")]; tensor input_1069 = linear(bias = encoder_layers_20_feed_forward1_linear2_bias, weight = encoder_layers_20_feed_forward1_linear2_weight, x = input_1065)[name = tensor("linear_182")]; tensor var_3718 = const()[name = tensor("op_3718"), val = tensor(0x1p-1)]; tensor var_3719 = mul(x = input_1069, y = var_3718)[name = tensor("op_3719")]; tensor input_1071 = add(x = input_1059, y = var_3719)[name = tensor("input_1071")]; tensor query_41_axes_0 = const()[name = tensor("query_41_axes_0"), val = tensor([-1])]; tensor query_41 = layer_norm(axes = query_41_axes_0, beta = encoder_layers_20_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_20_norm_self_att_weight, x = input_1071)[name = tensor("query_41")]; tensor var_3735 = linear(bias = encoder_layers_20_self_attn_linear_q_bias, weight = encoder_layers_20_self_attn_linear_q_weight, x = query_41)[name = tensor("linear_183")]; tensor var_3736 = const()[name = tensor("op_3736"), val = tensor([1, -1, 8, 128])]; tensor q_121 = reshape(shape = var_3736, x = var_3735)[name = tensor("q_121")]; tensor var_3740 = linear(bias = encoder_layers_20_self_attn_linear_k_bias, weight = encoder_layers_20_self_attn_linear_k_weight, x = query_41)[name = tensor("linear_184")]; tensor var_3741 = const()[name = tensor("op_3741"), val = tensor([1, -1, 8, 128])]; tensor k_81 = reshape(shape = var_3741, x = var_3740)[name = tensor("k_81")]; tensor var_3745 = linear(bias = encoder_layers_20_self_attn_linear_v_bias, weight = encoder_layers_20_self_attn_linear_v_weight, x = query_41)[name = tensor("linear_185")]; tensor var_3746 = const()[name = tensor("op_3746"), val = tensor([1, -1, 8, 128])]; tensor v_41 = reshape(shape = var_3746, x = var_3745)[name = tensor("v_41")]; tensor value_43_perm_0 = const()[name = tensor("value_43_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3758 = add(x = q_121, y = encoder_layers_20_self_attn_pos_bias_u)[name = tensor("op_3758")]; tensor var_3760 = add(x = q_121, y = encoder_layers_20_self_attn_pos_bias_v)[name = tensor("op_3760")]; tensor q_with_bias_v_41_perm_0 = const()[name = tensor("q_with_bias_v_41_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3762 = const()[name = tensor("op_3762"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2364949248)))]; tensor x_449_transpose_x_0 = const()[name = tensor("x_449_transpose_x_0"), val = tensor(false)]; tensor x_449_transpose_y_0 = const()[name = tensor("x_449_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_41 = transpose(perm = q_with_bias_v_41_perm_0, x = var_3760)[name = tensor("transpose_146")]; tensor x_449 = matmul(transpose_x = x_449_transpose_x_0, transpose_y = x_449_transpose_y_0, x = q_with_bias_v_41, y = var_3762)[name = tensor("x_449")]; tensor const_214 = const()[name = tensor("const_214"), val = tensor(0x0p+0)]; tensor x_451_pad_0 = const()[name = tensor("x_451_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_451_mode_0 = const()[name = tensor("x_451_mode_0"), val = tensor("constant")]; tensor x_451 = pad(constant_val = const_214, mode = x_451_mode_0, pad = x_451_pad_0, x = x_449)[name = tensor("x_451")]; tensor var_3770 = const()[name = tensor("op_3770"), val = tensor([1, 8, -1, 188])]; tensor x_453 = reshape(shape = var_3770, x = x_451)[name = tensor("x_453")]; tensor var_3774_begin_0 = const()[name = tensor("op_3774_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3774_end_0 = const()[name = tensor("op_3774_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3774_end_mask_0 = const()[name = tensor("op_3774_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3774 = slice_by_index(begin = var_3774_begin_0, end = var_3774_end_0, end_mask = var_3774_end_mask_0, x = x_453)[name = tensor("op_3774")]; tensor var_3775 = const()[name = tensor("op_3775"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_81 = reshape(shape = var_3775, x = var_3774)[name = tensor("matrix_bd_81")]; tensor matrix_ac_41_transpose_x_0 = const()[name = tensor("matrix_ac_41_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_41_transpose_y_0 = const()[name = tensor("matrix_ac_41_transpose_y_0"), val = tensor(false)]; tensor transpose_112_perm_0 = const()[name = tensor("transpose_112_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_113_perm_0 = const()[name = tensor("transpose_113_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_113 = transpose(perm = transpose_113_perm_0, x = k_81)[name = tensor("transpose_144")]; tensor transpose_112 = transpose(perm = transpose_112_perm_0, x = var_3758)[name = tensor("transpose_145")]; tensor matrix_ac_41 = matmul(transpose_x = matrix_ac_41_transpose_x_0, transpose_y = matrix_ac_41_transpose_y_0, x = transpose_112, y = transpose_113)[name = tensor("matrix_ac_41")]; tensor matrix_bd_83_begin_0 = const()[name = tensor("matrix_bd_83_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_83_end_0 = const()[name = tensor("matrix_bd_83_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_83_end_mask_0 = const()[name = tensor("matrix_bd_83_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_83 = slice_by_index(begin = matrix_bd_83_begin_0, end = matrix_bd_83_end_0, end_mask = matrix_bd_83_end_mask_0, x = matrix_bd_81)[name = tensor("matrix_bd_83")]; tensor var_3784 = add(x = matrix_ac_41, y = matrix_bd_83)[name = tensor("op_3784")]; tensor _inversed_scores_81_y_0 = const()[name = tensor("_inversed_scores_81_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_81 = mul(x = var_3784, y = _inversed_scores_81_y_0)[name = tensor("_inversed_scores_81")]; tensor scores_83 = select(a = var_14, b = _inversed_scores_81, cond = mask_3)[name = tensor("scores_83")]; tensor var_3790 = softmax(axis = var_32, x = scores_83)[name = tensor("op_3790")]; tensor input_1073 = select(a = var_13, b = var_3790, cond = mask_3)[name = tensor("input_1073")]; tensor x_455_transpose_x_0 = const()[name = tensor("x_455_transpose_x_0"), val = tensor(false)]; tensor x_455_transpose_y_0 = const()[name = tensor("x_455_transpose_y_0"), val = tensor(false)]; tensor value_43 = transpose(perm = value_43_perm_0, x = v_41)[name = tensor("transpose_147")]; tensor x_455 = matmul(transpose_x = x_455_transpose_x_0, transpose_y = x_455_transpose_y_0, x = input_1073, y = value_43)[name = tensor("x_455")]; tensor var_3794_perm_0 = const()[name = tensor("op_3794_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3795 = const()[name = tensor("op_3795"), val = tensor([1, -1, 1024])]; tensor var_3794 = transpose(perm = var_3794_perm_0, x = x_455)[name = tensor("transpose_143")]; tensor input_1075 = reshape(shape = var_3795, x = var_3794)[name = tensor("input_1075")]; tensor input_1077 = linear(bias = encoder_layers_20_self_attn_linear_out_bias, weight = encoder_layers_20_self_attn_linear_out_weight, x = input_1075)[name = tensor("linear_187")]; tensor input_1079 = add(x = input_1071, y = input_1077)[name = tensor("input_1079")]; tensor x_459_axes_0 = const()[name = tensor("x_459_axes_0"), val = tensor([-1])]; tensor x_459 = layer_norm(axes = x_459_axes_0, beta = encoder_layers_20_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_20_norm_conv_weight, x = input_1079)[name = tensor("x_459")]; tensor input_1081_perm_0 = const()[name = tensor("input_1081_perm_0"), val = tensor([0, 2, 1])]; tensor input_1083_pad_type_0 = const()[name = tensor("input_1083_pad_type_0"), val = tensor("valid")]; tensor input_1083_strides_0 = const()[name = tensor("input_1083_strides_0"), val = tensor([1])]; tensor input_1083_pad_0 = const()[name = tensor("input_1083_pad_0"), val = tensor([0, 0])]; tensor input_1083_dilations_0 = const()[name = tensor("input_1083_dilations_0"), val = tensor([1])]; tensor input_1083_groups_0 = const()[name = tensor("input_1083_groups_0"), val = tensor(1)]; tensor input_1081 = transpose(perm = input_1081_perm_0, x = x_459)[name = tensor("transpose_142")]; tensor input_1083 = conv(bias = encoder_layers_20_conv_pointwise_conv1_bias, dilations = input_1083_dilations_0, groups = input_1083_groups_0, pad = input_1083_pad_0, pad_type = input_1083_pad_type_0, strides = input_1083_strides_0, weight = encoder_layers_20_conv_pointwise_conv1_weight, x = input_1081)[name = tensor("input_1083")]; tensor x_461_split_num_splits_0 = const()[name = tensor("x_461_split_num_splits_0"), val = tensor(2)]; tensor x_461_split_axis_0 = const()[name = tensor("x_461_split_axis_0"), val = tensor(1)]; tensor x_461_split_0, tensor x_461_split_1 = split(axis = x_461_split_axis_0, num_splits = x_461_split_num_splits_0, x = input_1083)[name = tensor("x_461_split")]; tensor x_461_split_1_sigmoid = sigmoid(x = x_461_split_1)[name = tensor("x_461_split_1_sigmoid")]; tensor x_461 = mul(x = x_461_split_0, y = x_461_split_1_sigmoid)[name = tensor("x_461")]; tensor input_1085 = select(a = var_13, b = x_461, cond = var_339)[name = tensor("input_1085")]; tensor const_217 = const()[name = tensor("const_217"), val = tensor(0x0p+0)]; tensor input_1087_pad_0 = const()[name = tensor("input_1087_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1087_mode_0 = const()[name = tensor("input_1087_mode_0"), val = tensor("constant")]; tensor input_1087 = pad(constant_val = const_217, mode = input_1087_mode_0, pad = input_1087_pad_0, x = input_1085)[name = tensor("input_1087")]; tensor input_1089_pad_type_0 = const()[name = tensor("input_1089_pad_type_0"), val = tensor("valid")]; tensor input_1089_groups_0 = const()[name = tensor("input_1089_groups_0"), val = tensor(1024)]; tensor input_1089_strides_0 = const()[name = tensor("input_1089_strides_0"), val = tensor([1])]; tensor input_1089_pad_0 = const()[name = tensor("input_1089_pad_0"), val = tensor([0, 0])]; tensor input_1089_dilations_0 = const()[name = tensor("input_1089_dilations_0"), val = tensor([1])]; tensor const_288 = const()[name = tensor("const_288"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2366485312)))]; tensor const_289 = const()[name = tensor("const_289"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2366522240)))]; tensor input_1091 = conv(bias = const_289, dilations = input_1089_dilations_0, groups = input_1089_groups_0, pad = input_1089_pad_0, pad_type = input_1089_pad_type_0, strides = input_1089_strides_0, weight = const_288, x = input_1087)[name = tensor("input_1091")]; tensor input_1093 = silu(x = input_1091)[name = tensor("input_1093")]; tensor x_463_pad_type_0 = const()[name = tensor("x_463_pad_type_0"), val = tensor("valid")]; tensor x_463_strides_0 = const()[name = tensor("x_463_strides_0"), val = tensor([1])]; tensor x_463_pad_0 = const()[name = tensor("x_463_pad_0"), val = tensor([0, 0])]; tensor x_463_dilations_0 = const()[name = tensor("x_463_dilations_0"), val = tensor([1])]; tensor x_463_groups_0 = const()[name = tensor("x_463_groups_0"), val = tensor(1)]; tensor x_463 = conv(bias = encoder_layers_20_conv_pointwise_conv2_bias, dilations = x_463_dilations_0, groups = x_463_groups_0, pad = x_463_pad_0, pad_type = x_463_pad_type_0, strides = x_463_strides_0, weight = encoder_layers_20_conv_pointwise_conv2_weight, x = input_1093)[name = tensor("x_463")]; tensor input_1095_perm_0 = const()[name = tensor("input_1095_perm_0"), val = tensor([0, 2, 1])]; tensor input_1095 = transpose(perm = input_1095_perm_0, x = x_463)[name = tensor("transpose_141")]; tensor input_1097 = add(x = input_1079, y = input_1095)[name = tensor("input_1097")]; tensor input_1099_axes_0 = const()[name = tensor("input_1099_axes_0"), val = tensor([-1])]; tensor input_1099 = layer_norm(axes = input_1099_axes_0, beta = encoder_layers_20_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_20_norm_feed_forward2_weight, x = input_1097)[name = tensor("input_1099")]; tensor input_1101 = linear(bias = encoder_layers_20_feed_forward2_linear1_bias, weight = encoder_layers_20_feed_forward2_linear1_weight, x = input_1099)[name = tensor("linear_188")]; tensor input_1103 = silu(x = input_1101)[name = tensor("input_1103")]; tensor input_1107 = linear(bias = encoder_layers_20_feed_forward2_linear2_bias, weight = encoder_layers_20_feed_forward2_linear2_weight, x = input_1103)[name = tensor("linear_189")]; tensor var_3861 = const()[name = tensor("op_3861"), val = tensor(0x1p-1)]; tensor var_3862 = mul(x = input_1107, y = var_3861)[name = tensor("op_3862")]; tensor input_1109 = add(x = input_1097, y = var_3862)[name = tensor("input_1109")]; tensor input_1111_axes_0 = const()[name = tensor("input_1111_axes_0"), val = tensor([-1])]; tensor input_1111 = layer_norm(axes = input_1111_axes_0, beta = encoder_layers_20_norm_out_bias, epsilon = var_11, gamma = encoder_layers_20_norm_out_weight, x = input_1109)[name = tensor("input_1111")]; tensor input_1113_axes_0 = const()[name = tensor("input_1113_axes_0"), val = tensor([-1])]; tensor input_1113 = layer_norm(axes = input_1113_axes_0, beta = encoder_layers_21_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_21_norm_feed_forward1_weight, x = input_1111)[name = tensor("input_1113")]; tensor input_1115 = linear(bias = encoder_layers_21_feed_forward1_linear1_bias, weight = encoder_layers_21_feed_forward1_linear1_weight, x = input_1113)[name = tensor("linear_190")]; tensor input_1117 = silu(x = input_1115)[name = tensor("input_1117")]; tensor input_1121 = linear(bias = encoder_layers_21_feed_forward1_linear2_bias, weight = encoder_layers_21_feed_forward1_linear2_weight, x = input_1117)[name = tensor("linear_191")]; tensor var_3892 = const()[name = tensor("op_3892"), val = tensor(0x1p-1)]; tensor var_3893 = mul(x = input_1121, y = var_3892)[name = tensor("op_3893")]; tensor input_1123 = add(x = input_1111, y = var_3893)[name = tensor("input_1123")]; tensor query_43_axes_0 = const()[name = tensor("query_43_axes_0"), val = tensor([-1])]; tensor query_43 = layer_norm(axes = query_43_axes_0, beta = encoder_layers_21_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_21_norm_self_att_weight, x = input_1123)[name = tensor("query_43")]; tensor var_3909 = linear(bias = encoder_layers_21_self_attn_linear_q_bias, weight = encoder_layers_21_self_attn_linear_q_weight, x = query_43)[name = tensor("linear_192")]; tensor var_3910 = const()[name = tensor("op_3910"), val = tensor([1, -1, 8, 128])]; tensor q_127 = reshape(shape = var_3910, x = var_3909)[name = tensor("q_127")]; tensor var_3914 = linear(bias = encoder_layers_21_self_attn_linear_k_bias, weight = encoder_layers_21_self_attn_linear_k_weight, x = query_43)[name = tensor("linear_193")]; tensor var_3915 = const()[name = tensor("op_3915"), val = tensor([1, -1, 8, 128])]; tensor k_85 = reshape(shape = var_3915, x = var_3914)[name = tensor("k_85")]; tensor var_3919 = linear(bias = encoder_layers_21_self_attn_linear_v_bias, weight = encoder_layers_21_self_attn_linear_v_weight, x = query_43)[name = tensor("linear_194")]; tensor var_3920 = const()[name = tensor("op_3920"), val = tensor([1, -1, 8, 128])]; tensor v_43 = reshape(shape = var_3920, x = var_3919)[name = tensor("v_43")]; tensor value_45_perm_0 = const()[name = tensor("value_45_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3932 = add(x = q_127, y = encoder_layers_21_self_attn_pos_bias_u)[name = tensor("op_3932")]; tensor var_3934 = add(x = q_127, y = encoder_layers_21_self_attn_pos_bias_v)[name = tensor("op_3934")]; tensor q_with_bias_v_43_perm_0 = const()[name = tensor("q_with_bias_v_43_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3936 = const()[name = tensor("op_3936"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2366526400)))]; tensor x_471_transpose_x_0 = const()[name = tensor("x_471_transpose_x_0"), val = tensor(false)]; tensor x_471_transpose_y_0 = const()[name = tensor("x_471_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_43 = transpose(perm = q_with_bias_v_43_perm_0, x = var_3934)[name = tensor("transpose_139")]; tensor x_471 = matmul(transpose_x = x_471_transpose_x_0, transpose_y = x_471_transpose_y_0, x = q_with_bias_v_43, y = var_3936)[name = tensor("x_471")]; tensor const_224 = const()[name = tensor("const_224"), val = tensor(0x0p+0)]; tensor x_473_pad_0 = const()[name = tensor("x_473_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_473_mode_0 = const()[name = tensor("x_473_mode_0"), val = tensor("constant")]; tensor x_473 = pad(constant_val = const_224, mode = x_473_mode_0, pad = x_473_pad_0, x = x_471)[name = tensor("x_473")]; tensor var_3944 = const()[name = tensor("op_3944"), val = tensor([1, 8, -1, 188])]; tensor x_475 = reshape(shape = var_3944, x = x_473)[name = tensor("x_475")]; tensor var_3948_begin_0 = const()[name = tensor("op_3948_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3948_end_0 = const()[name = tensor("op_3948_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3948_end_mask_0 = const()[name = tensor("op_3948_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3948 = slice_by_index(begin = var_3948_begin_0, end = var_3948_end_0, end_mask = var_3948_end_mask_0, x = x_475)[name = tensor("op_3948")]; tensor var_3949 = const()[name = tensor("op_3949"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_85 = reshape(shape = var_3949, x = var_3948)[name = tensor("matrix_bd_85")]; tensor matrix_ac_43_transpose_x_0 = const()[name = tensor("matrix_ac_43_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_43_transpose_y_0 = const()[name = tensor("matrix_ac_43_transpose_y_0"), val = tensor(false)]; tensor transpose_114_perm_0 = const()[name = tensor("transpose_114_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_115_perm_0 = const()[name = tensor("transpose_115_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_115 = transpose(perm = transpose_115_perm_0, x = k_85)[name = tensor("transpose_137")]; tensor transpose_114 = transpose(perm = transpose_114_perm_0, x = var_3932)[name = tensor("transpose_138")]; tensor matrix_ac_43 = matmul(transpose_x = matrix_ac_43_transpose_x_0, transpose_y = matrix_ac_43_transpose_y_0, x = transpose_114, y = transpose_115)[name = tensor("matrix_ac_43")]; tensor matrix_bd_87_begin_0 = const()[name = tensor("matrix_bd_87_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_87_end_0 = const()[name = tensor("matrix_bd_87_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_87_end_mask_0 = const()[name = tensor("matrix_bd_87_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_87 = slice_by_index(begin = matrix_bd_87_begin_0, end = matrix_bd_87_end_0, end_mask = matrix_bd_87_end_mask_0, x = matrix_bd_85)[name = tensor("matrix_bd_87")]; tensor var_3958 = add(x = matrix_ac_43, y = matrix_bd_87)[name = tensor("op_3958")]; tensor _inversed_scores_85_y_0 = const()[name = tensor("_inversed_scores_85_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_85 = mul(x = var_3958, y = _inversed_scores_85_y_0)[name = tensor("_inversed_scores_85")]; tensor scores_87 = select(a = var_14, b = _inversed_scores_85, cond = mask_3)[name = tensor("scores_87")]; tensor var_3964 = softmax(axis = var_32, x = scores_87)[name = tensor("op_3964")]; tensor input_1125 = select(a = var_13, b = var_3964, cond = mask_3)[name = tensor("input_1125")]; tensor x_477_transpose_x_0 = const()[name = tensor("x_477_transpose_x_0"), val = tensor(false)]; tensor x_477_transpose_y_0 = const()[name = tensor("x_477_transpose_y_0"), val = tensor(false)]; tensor value_45 = transpose(perm = value_45_perm_0, x = v_43)[name = tensor("transpose_140")]; tensor x_477 = matmul(transpose_x = x_477_transpose_x_0, transpose_y = x_477_transpose_y_0, x = input_1125, y = value_45)[name = tensor("x_477")]; tensor var_3968_perm_0 = const()[name = tensor("op_3968_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3969 = const()[name = tensor("op_3969"), val = tensor([1, -1, 1024])]; tensor var_3968 = transpose(perm = var_3968_perm_0, x = x_477)[name = tensor("transpose_136")]; tensor input_1127 = reshape(shape = var_3969, x = var_3968)[name = tensor("input_1127")]; tensor input_1129 = linear(bias = encoder_layers_21_self_attn_linear_out_bias, weight = encoder_layers_21_self_attn_linear_out_weight, x = input_1127)[name = tensor("linear_196")]; tensor input_1131 = add(x = input_1123, y = input_1129)[name = tensor("input_1131")]; tensor x_481_axes_0 = const()[name = tensor("x_481_axes_0"), val = tensor([-1])]; tensor x_481 = layer_norm(axes = x_481_axes_0, beta = encoder_layers_21_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_21_norm_conv_weight, x = input_1131)[name = tensor("x_481")]; tensor input_1133_perm_0 = const()[name = tensor("input_1133_perm_0"), val = tensor([0, 2, 1])]; tensor input_1135_pad_type_0 = const()[name = tensor("input_1135_pad_type_0"), val = tensor("valid")]; tensor input_1135_strides_0 = const()[name = tensor("input_1135_strides_0"), val = tensor([1])]; tensor input_1135_pad_0 = const()[name = tensor("input_1135_pad_0"), val = tensor([0, 0])]; tensor input_1135_dilations_0 = const()[name = tensor("input_1135_dilations_0"), val = tensor([1])]; tensor input_1135_groups_0 = const()[name = tensor("input_1135_groups_0"), val = tensor(1)]; tensor input_1133 = transpose(perm = input_1133_perm_0, x = x_481)[name = tensor("transpose_135")]; tensor input_1135 = conv(bias = encoder_layers_21_conv_pointwise_conv1_bias, dilations = input_1135_dilations_0, groups = input_1135_groups_0, pad = input_1135_pad_0, pad_type = input_1135_pad_type_0, strides = input_1135_strides_0, weight = encoder_layers_21_conv_pointwise_conv1_weight, x = input_1133)[name = tensor("input_1135")]; tensor x_483_split_num_splits_0 = const()[name = tensor("x_483_split_num_splits_0"), val = tensor(2)]; tensor x_483_split_axis_0 = const()[name = tensor("x_483_split_axis_0"), val = tensor(1)]; tensor x_483_split_0, tensor x_483_split_1 = split(axis = x_483_split_axis_0, num_splits = x_483_split_num_splits_0, x = input_1135)[name = tensor("x_483_split")]; tensor x_483_split_1_sigmoid = sigmoid(x = x_483_split_1)[name = tensor("x_483_split_1_sigmoid")]; tensor x_483 = mul(x = x_483_split_0, y = x_483_split_1_sigmoid)[name = tensor("x_483")]; tensor input_1137 = select(a = var_13, b = x_483, cond = var_339)[name = tensor("input_1137")]; tensor const_227 = const()[name = tensor("const_227"), val = tensor(0x0p+0)]; tensor input_1139_pad_0 = const()[name = tensor("input_1139_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1139_mode_0 = const()[name = tensor("input_1139_mode_0"), val = tensor("constant")]; tensor input_1139 = pad(constant_val = const_227, mode = input_1139_mode_0, pad = input_1139_pad_0, x = input_1137)[name = tensor("input_1139")]; tensor input_1141_pad_type_0 = const()[name = tensor("input_1141_pad_type_0"), val = tensor("valid")]; tensor input_1141_groups_0 = const()[name = tensor("input_1141_groups_0"), val = tensor(1024)]; tensor input_1141_strides_0 = const()[name = tensor("input_1141_strides_0"), val = tensor([1])]; tensor input_1141_pad_0 = const()[name = tensor("input_1141_pad_0"), val = tensor([0, 0])]; tensor input_1141_dilations_0 = const()[name = tensor("input_1141_dilations_0"), val = tensor([1])]; tensor const_290 = const()[name = tensor("const_290"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2368062464)))]; tensor const_291 = const()[name = tensor("const_291"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2368099392)))]; tensor input_1143 = conv(bias = const_291, dilations = input_1141_dilations_0, groups = input_1141_groups_0, pad = input_1141_pad_0, pad_type = input_1141_pad_type_0, strides = input_1141_strides_0, weight = const_290, x = input_1139)[name = tensor("input_1143")]; tensor input_1145 = silu(x = input_1143)[name = tensor("input_1145")]; tensor x_485_pad_type_0 = const()[name = tensor("x_485_pad_type_0"), val = tensor("valid")]; tensor x_485_strides_0 = const()[name = tensor("x_485_strides_0"), val = tensor([1])]; tensor x_485_pad_0 = const()[name = tensor("x_485_pad_0"), val = tensor([0, 0])]; tensor x_485_dilations_0 = const()[name = tensor("x_485_dilations_0"), val = tensor([1])]; tensor x_485_groups_0 = const()[name = tensor("x_485_groups_0"), val = tensor(1)]; tensor x_485 = conv(bias = encoder_layers_21_conv_pointwise_conv2_bias, dilations = x_485_dilations_0, groups = x_485_groups_0, pad = x_485_pad_0, pad_type = x_485_pad_type_0, strides = x_485_strides_0, weight = encoder_layers_21_conv_pointwise_conv2_weight, x = input_1145)[name = tensor("x_485")]; tensor input_1147_perm_0 = const()[name = tensor("input_1147_perm_0"), val = tensor([0, 2, 1])]; tensor input_1147 = transpose(perm = input_1147_perm_0, x = x_485)[name = tensor("transpose_134")]; tensor input_1149 = add(x = input_1131, y = input_1147)[name = tensor("input_1149")]; tensor input_1151_axes_0 = const()[name = tensor("input_1151_axes_0"), val = tensor([-1])]; tensor input_1151 = layer_norm(axes = input_1151_axes_0, beta = encoder_layers_21_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_21_norm_feed_forward2_weight, x = input_1149)[name = tensor("input_1151")]; tensor input_1153 = linear(bias = encoder_layers_21_feed_forward2_linear1_bias, weight = encoder_layers_21_feed_forward2_linear1_weight, x = input_1151)[name = tensor("linear_197")]; tensor input_1155 = silu(x = input_1153)[name = tensor("input_1155")]; tensor input_1159 = linear(bias = encoder_layers_21_feed_forward2_linear2_bias, weight = encoder_layers_21_feed_forward2_linear2_weight, x = input_1155)[name = tensor("linear_198")]; tensor var_4035 = const()[name = tensor("op_4035"), val = tensor(0x1p-1)]; tensor var_4036 = mul(x = input_1159, y = var_4035)[name = tensor("op_4036")]; tensor input_1161 = add(x = input_1149, y = var_4036)[name = tensor("input_1161")]; tensor input_1163_axes_0 = const()[name = tensor("input_1163_axes_0"), val = tensor([-1])]; tensor input_1163 = layer_norm(axes = input_1163_axes_0, beta = encoder_layers_21_norm_out_bias, epsilon = var_11, gamma = encoder_layers_21_norm_out_weight, x = input_1161)[name = tensor("input_1163")]; tensor input_1165_axes_0 = const()[name = tensor("input_1165_axes_0"), val = tensor([-1])]; tensor input_1165 = layer_norm(axes = input_1165_axes_0, beta = encoder_layers_22_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_22_norm_feed_forward1_weight, x = input_1163)[name = tensor("input_1165")]; tensor input_1167 = linear(bias = encoder_layers_22_feed_forward1_linear1_bias, weight = encoder_layers_22_feed_forward1_linear1_weight, x = input_1165)[name = tensor("linear_199")]; tensor input_1169 = silu(x = input_1167)[name = tensor("input_1169")]; tensor input_1173 = linear(bias = encoder_layers_22_feed_forward1_linear2_bias, weight = encoder_layers_22_feed_forward1_linear2_weight, x = input_1169)[name = tensor("linear_200")]; tensor var_4066 = const()[name = tensor("op_4066"), val = tensor(0x1p-1)]; tensor var_4067 = mul(x = input_1173, y = var_4066)[name = tensor("op_4067")]; tensor input_1175 = add(x = input_1163, y = var_4067)[name = tensor("input_1175")]; tensor query_45_axes_0 = const()[name = tensor("query_45_axes_0"), val = tensor([-1])]; tensor query_45 = layer_norm(axes = query_45_axes_0, beta = encoder_layers_22_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_22_norm_self_att_weight, x = input_1175)[name = tensor("query_45")]; tensor var_4083 = linear(bias = encoder_layers_22_self_attn_linear_q_bias, weight = encoder_layers_22_self_attn_linear_q_weight, x = query_45)[name = tensor("linear_201")]; tensor var_4084 = const()[name = tensor("op_4084"), val = tensor([1, -1, 8, 128])]; tensor q_133 = reshape(shape = var_4084, x = var_4083)[name = tensor("q_133")]; tensor var_4088 = linear(bias = encoder_layers_22_self_attn_linear_k_bias, weight = encoder_layers_22_self_attn_linear_k_weight, x = query_45)[name = tensor("linear_202")]; tensor var_4089 = const()[name = tensor("op_4089"), val = tensor([1, -1, 8, 128])]; tensor k_89 = reshape(shape = var_4089, x = var_4088)[name = tensor("k_89")]; tensor var_4093 = linear(bias = encoder_layers_22_self_attn_linear_v_bias, weight = encoder_layers_22_self_attn_linear_v_weight, x = query_45)[name = tensor("linear_203")]; tensor var_4094 = const()[name = tensor("op_4094"), val = tensor([1, -1, 8, 128])]; tensor v_45 = reshape(shape = var_4094, x = var_4093)[name = tensor("v_45")]; tensor value_47_perm_0 = const()[name = tensor("value_47_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4106 = add(x = q_133, y = encoder_layers_22_self_attn_pos_bias_u)[name = tensor("op_4106")]; tensor var_4108 = add(x = q_133, y = encoder_layers_22_self_attn_pos_bias_v)[name = tensor("op_4108")]; tensor q_with_bias_v_45_perm_0 = const()[name = tensor("q_with_bias_v_45_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4110 = const()[name = tensor("op_4110"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2368103552)))]; tensor x_493_transpose_x_0 = const()[name = tensor("x_493_transpose_x_0"), val = tensor(false)]; tensor x_493_transpose_y_0 = const()[name = tensor("x_493_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v_45 = transpose(perm = q_with_bias_v_45_perm_0, x = var_4108)[name = tensor("transpose_132")]; tensor x_493 = matmul(transpose_x = x_493_transpose_x_0, transpose_y = x_493_transpose_y_0, x = q_with_bias_v_45, y = var_4110)[name = tensor("x_493")]; tensor const_234 = const()[name = tensor("const_234"), val = tensor(0x0p+0)]; tensor x_495_pad_0 = const()[name = tensor("x_495_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_495_mode_0 = const()[name = tensor("x_495_mode_0"), val = tensor("constant")]; tensor x_495 = pad(constant_val = const_234, mode = x_495_mode_0, pad = x_495_pad_0, x = x_493)[name = tensor("x_495")]; tensor var_4118 = const()[name = tensor("op_4118"), val = tensor([1, 8, -1, 188])]; tensor x_497 = reshape(shape = var_4118, x = x_495)[name = tensor("x_497")]; tensor var_4122_begin_0 = const()[name = tensor("op_4122_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_4122_end_0 = const()[name = tensor("op_4122_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_4122_end_mask_0 = const()[name = tensor("op_4122_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_4122 = slice_by_index(begin = var_4122_begin_0, end = var_4122_end_0, end_mask = var_4122_end_mask_0, x = x_497)[name = tensor("op_4122")]; tensor var_4123 = const()[name = tensor("op_4123"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_89 = reshape(shape = var_4123, x = var_4122)[name = tensor("matrix_bd_89")]; tensor matrix_ac_45_transpose_x_0 = const()[name = tensor("matrix_ac_45_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_45_transpose_y_0 = const()[name = tensor("matrix_ac_45_transpose_y_0"), val = tensor(false)]; tensor transpose_116_perm_0 = const()[name = tensor("transpose_116_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_117_perm_0 = const()[name = tensor("transpose_117_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_117 = transpose(perm = transpose_117_perm_0, x = k_89)[name = tensor("transpose_130")]; tensor transpose_116 = transpose(perm = transpose_116_perm_0, x = var_4106)[name = tensor("transpose_131")]; tensor matrix_ac_45 = matmul(transpose_x = matrix_ac_45_transpose_x_0, transpose_y = matrix_ac_45_transpose_y_0, x = transpose_116, y = transpose_117)[name = tensor("matrix_ac_45")]; tensor matrix_bd_91_begin_0 = const()[name = tensor("matrix_bd_91_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_91_end_0 = const()[name = tensor("matrix_bd_91_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_91_end_mask_0 = const()[name = tensor("matrix_bd_91_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_91 = slice_by_index(begin = matrix_bd_91_begin_0, end = matrix_bd_91_end_0, end_mask = matrix_bd_91_end_mask_0, x = matrix_bd_89)[name = tensor("matrix_bd_91")]; tensor var_4132 = add(x = matrix_ac_45, y = matrix_bd_91)[name = tensor("op_4132")]; tensor _inversed_scores_89_y_0 = const()[name = tensor("_inversed_scores_89_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_89 = mul(x = var_4132, y = _inversed_scores_89_y_0)[name = tensor("_inversed_scores_89")]; tensor scores_91 = select(a = var_14, b = _inversed_scores_89, cond = mask_3)[name = tensor("scores_91")]; tensor var_4138 = softmax(axis = var_32, x = scores_91)[name = tensor("op_4138")]; tensor input_1177 = select(a = var_13, b = var_4138, cond = mask_3)[name = tensor("input_1177")]; tensor x_499_transpose_x_0 = const()[name = tensor("x_499_transpose_x_0"), val = tensor(false)]; tensor x_499_transpose_y_0 = const()[name = tensor("x_499_transpose_y_0"), val = tensor(false)]; tensor value_47 = transpose(perm = value_47_perm_0, x = v_45)[name = tensor("transpose_133")]; tensor x_499 = matmul(transpose_x = x_499_transpose_x_0, transpose_y = x_499_transpose_y_0, x = input_1177, y = value_47)[name = tensor("x_499")]; tensor var_4142_perm_0 = const()[name = tensor("op_4142_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4143 = const()[name = tensor("op_4143"), val = tensor([1, -1, 1024])]; tensor var_4142 = transpose(perm = var_4142_perm_0, x = x_499)[name = tensor("transpose_129")]; tensor input_1179 = reshape(shape = var_4143, x = var_4142)[name = tensor("input_1179")]; tensor input_1181 = linear(bias = encoder_layers_22_self_attn_linear_out_bias, weight = encoder_layers_22_self_attn_linear_out_weight, x = input_1179)[name = tensor("linear_205")]; tensor input_1183 = add(x = input_1175, y = input_1181)[name = tensor("input_1183")]; tensor x_503_axes_0 = const()[name = tensor("x_503_axes_0"), val = tensor([-1])]; tensor x_503 = layer_norm(axes = x_503_axes_0, beta = encoder_layers_22_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_22_norm_conv_weight, x = input_1183)[name = tensor("x_503")]; tensor input_1185_perm_0 = const()[name = tensor("input_1185_perm_0"), val = tensor([0, 2, 1])]; tensor input_1187_pad_type_0 = const()[name = tensor("input_1187_pad_type_0"), val = tensor("valid")]; tensor input_1187_strides_0 = const()[name = tensor("input_1187_strides_0"), val = tensor([1])]; tensor input_1187_pad_0 = const()[name = tensor("input_1187_pad_0"), val = tensor([0, 0])]; tensor input_1187_dilations_0 = const()[name = tensor("input_1187_dilations_0"), val = tensor([1])]; tensor input_1187_groups_0 = const()[name = tensor("input_1187_groups_0"), val = tensor(1)]; tensor input_1185 = transpose(perm = input_1185_perm_0, x = x_503)[name = tensor("transpose_128")]; tensor input_1187 = conv(bias = encoder_layers_22_conv_pointwise_conv1_bias, dilations = input_1187_dilations_0, groups = input_1187_groups_0, pad = input_1187_pad_0, pad_type = input_1187_pad_type_0, strides = input_1187_strides_0, weight = encoder_layers_22_conv_pointwise_conv1_weight, x = input_1185)[name = tensor("input_1187")]; tensor x_505_split_num_splits_0 = const()[name = tensor("x_505_split_num_splits_0"), val = tensor(2)]; tensor x_505_split_axis_0 = const()[name = tensor("x_505_split_axis_0"), val = tensor(1)]; tensor x_505_split_0, tensor x_505_split_1 = split(axis = x_505_split_axis_0, num_splits = x_505_split_num_splits_0, x = input_1187)[name = tensor("x_505_split")]; tensor x_505_split_1_sigmoid = sigmoid(x = x_505_split_1)[name = tensor("x_505_split_1_sigmoid")]; tensor x_505 = mul(x = x_505_split_0, y = x_505_split_1_sigmoid)[name = tensor("x_505")]; tensor input_1189 = select(a = var_13, b = x_505, cond = var_339)[name = tensor("input_1189")]; tensor const_237 = const()[name = tensor("const_237"), val = tensor(0x0p+0)]; tensor input_1191_pad_0 = const()[name = tensor("input_1191_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1191_mode_0 = const()[name = tensor("input_1191_mode_0"), val = tensor("constant")]; tensor input_1191 = pad(constant_val = const_237, mode = input_1191_mode_0, pad = input_1191_pad_0, x = input_1189)[name = tensor("input_1191")]; tensor input_1193_pad_type_0 = const()[name = tensor("input_1193_pad_type_0"), val = tensor("valid")]; tensor input_1193_groups_0 = const()[name = tensor("input_1193_groups_0"), val = tensor(1024)]; tensor input_1193_strides_0 = const()[name = tensor("input_1193_strides_0"), val = tensor([1])]; tensor input_1193_pad_0 = const()[name = tensor("input_1193_pad_0"), val = tensor([0, 0])]; tensor input_1193_dilations_0 = const()[name = tensor("input_1193_dilations_0"), val = tensor([1])]; tensor const_292 = const()[name = tensor("const_292"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2369639616)))]; tensor const_293 = const()[name = tensor("const_293"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2369676544)))]; tensor input_1195 = conv(bias = const_293, dilations = input_1193_dilations_0, groups = input_1193_groups_0, pad = input_1193_pad_0, pad_type = input_1193_pad_type_0, strides = input_1193_strides_0, weight = const_292, x = input_1191)[name = tensor("input_1195")]; tensor input_1197 = silu(x = input_1195)[name = tensor("input_1197")]; tensor x_507_pad_type_0 = const()[name = tensor("x_507_pad_type_0"), val = tensor("valid")]; tensor x_507_strides_0 = const()[name = tensor("x_507_strides_0"), val = tensor([1])]; tensor x_507_pad_0 = const()[name = tensor("x_507_pad_0"), val = tensor([0, 0])]; tensor x_507_dilations_0 = const()[name = tensor("x_507_dilations_0"), val = tensor([1])]; tensor x_507_groups_0 = const()[name = tensor("x_507_groups_0"), val = tensor(1)]; tensor x_507 = conv(bias = encoder_layers_22_conv_pointwise_conv2_bias, dilations = x_507_dilations_0, groups = x_507_groups_0, pad = x_507_pad_0, pad_type = x_507_pad_type_0, strides = x_507_strides_0, weight = encoder_layers_22_conv_pointwise_conv2_weight, x = input_1197)[name = tensor("x_507")]; tensor input_1199_perm_0 = const()[name = tensor("input_1199_perm_0"), val = tensor([0, 2, 1])]; tensor input_1199 = transpose(perm = input_1199_perm_0, x = x_507)[name = tensor("transpose_127")]; tensor input_1201 = add(x = input_1183, y = input_1199)[name = tensor("input_1201")]; tensor input_1203_axes_0 = const()[name = tensor("input_1203_axes_0"), val = tensor([-1])]; tensor input_1203 = layer_norm(axes = input_1203_axes_0, beta = encoder_layers_22_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_22_norm_feed_forward2_weight, x = input_1201)[name = tensor("input_1203")]; tensor input_1205 = linear(bias = encoder_layers_22_feed_forward2_linear1_bias, weight = encoder_layers_22_feed_forward2_linear1_weight, x = input_1203)[name = tensor("linear_206")]; tensor input_1207 = silu(x = input_1205)[name = tensor("input_1207")]; tensor input_1211 = linear(bias = encoder_layers_22_feed_forward2_linear2_bias, weight = encoder_layers_22_feed_forward2_linear2_weight, x = input_1207)[name = tensor("linear_207")]; tensor var_4209 = const()[name = tensor("op_4209"), val = tensor(0x1p-1)]; tensor var_4210 = mul(x = input_1211, y = var_4209)[name = tensor("op_4210")]; tensor input_1213 = add(x = input_1201, y = var_4210)[name = tensor("input_1213")]; tensor input_1215_axes_0 = const()[name = tensor("input_1215_axes_0"), val = tensor([-1])]; tensor input_1215 = layer_norm(axes = input_1215_axes_0, beta = encoder_layers_22_norm_out_bias, epsilon = var_11, gamma = encoder_layers_22_norm_out_weight, x = input_1213)[name = tensor("input_1215")]; tensor input_1217_axes_0 = const()[name = tensor("input_1217_axes_0"), val = tensor([-1])]; tensor input_1217 = layer_norm(axes = input_1217_axes_0, beta = encoder_layers_23_norm_feed_forward1_bias, epsilon = var_11, gamma = encoder_layers_23_norm_feed_forward1_weight, x = input_1215)[name = tensor("input_1217")]; tensor input_1219 = linear(bias = encoder_layers_23_feed_forward1_linear1_bias, weight = encoder_layers_23_feed_forward1_linear1_weight, x = input_1217)[name = tensor("linear_208")]; tensor input_1221 = silu(x = input_1219)[name = tensor("input_1221")]; tensor input_1225 = linear(bias = encoder_layers_23_feed_forward1_linear2_bias, weight = encoder_layers_23_feed_forward1_linear2_weight, x = input_1221)[name = tensor("linear_209")]; tensor var_4240 = const()[name = tensor("op_4240"), val = tensor(0x1p-1)]; tensor var_4241 = mul(x = input_1225, y = var_4240)[name = tensor("op_4241")]; tensor input_1227 = add(x = input_1215, y = var_4241)[name = tensor("input_1227")]; tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; tensor query = layer_norm(axes = query_axes_0, beta = encoder_layers_23_norm_self_att_bias, epsilon = var_11, gamma = encoder_layers_23_norm_self_att_weight, x = input_1227)[name = tensor("query")]; tensor var_4257 = linear(bias = encoder_layers_23_self_attn_linear_q_bias, weight = encoder_layers_23_self_attn_linear_q_weight, x = query)[name = tensor("linear_210")]; tensor var_4258 = const()[name = tensor("op_4258"), val = tensor([1, -1, 8, 128])]; tensor q_139 = reshape(shape = var_4258, x = var_4257)[name = tensor("q_139")]; tensor var_4262 = linear(bias = encoder_layers_23_self_attn_linear_k_bias, weight = encoder_layers_23_self_attn_linear_k_weight, x = query)[name = tensor("linear_211")]; tensor var_4263 = const()[name = tensor("op_4263"), val = tensor([1, -1, 8, 128])]; tensor k_93 = reshape(shape = var_4263, x = var_4262)[name = tensor("k_93")]; tensor var_4267 = linear(bias = encoder_layers_23_self_attn_linear_v_bias, weight = encoder_layers_23_self_attn_linear_v_weight, x = query)[name = tensor("linear_212")]; tensor var_4268 = const()[name = tensor("op_4268"), val = tensor([1, -1, 8, 128])]; tensor v = reshape(shape = var_4268, x = var_4267)[name = tensor("v")]; tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4280 = add(x = q_139, y = encoder_layers_23_self_attn_pos_bias_u)[name = tensor("op_4280")]; tensor var_4282 = add(x = q_139, y = encoder_layers_23_self_attn_pos_bias_v)[name = tensor("op_4282")]; tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4284 = const()[name = tensor("op_4284"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2369680704)))]; tensor x_515_transpose_x_0 = const()[name = tensor("x_515_transpose_x_0"), val = tensor(false)]; tensor x_515_transpose_y_0 = const()[name = tensor("x_515_transpose_y_0"), val = tensor(false)]; tensor q_with_bias_v = transpose(perm = q_with_bias_v_perm_0, x = var_4282)[name = tensor("transpose_125")]; tensor x_515 = matmul(transpose_x = x_515_transpose_x_0, transpose_y = x_515_transpose_y_0, x = q_with_bias_v, y = var_4284)[name = tensor("x_515")]; tensor const_244 = const()[name = tensor("const_244"), val = tensor(0x0p+0)]; tensor x_517_pad_0 = const()[name = tensor("x_517_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_517_mode_0 = const()[name = tensor("x_517_mode_0"), val = tensor("constant")]; tensor x_517 = pad(constant_val = const_244, mode = x_517_mode_0, pad = x_517_pad_0, x = x_515)[name = tensor("x_517")]; tensor var_4292 = const()[name = tensor("op_4292"), val = tensor([1, 8, -1, 188])]; tensor x_519 = reshape(shape = var_4292, x = x_517)[name = tensor("x_519")]; tensor var_4296_begin_0 = const()[name = tensor("op_4296_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_4296_end_0 = const()[name = tensor("op_4296_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_4296_end_mask_0 = const()[name = tensor("op_4296_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_4296 = slice_by_index(begin = var_4296_begin_0, end = var_4296_end_0, end_mask = var_4296_end_mask_0, x = x_519)[name = tensor("op_4296")]; tensor var_4297 = const()[name = tensor("op_4297"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_93 = reshape(shape = var_4297, x = var_4296)[name = tensor("matrix_bd_93")]; tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; tensor transpose_118_perm_0 = const()[name = tensor("transpose_118_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_119_perm_0 = const()[name = tensor("transpose_119_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_119 = transpose(perm = transpose_119_perm_0, x = k_93)[name = tensor("transpose_123")]; tensor transpose_118 = transpose(perm = transpose_118_perm_0, x = var_4280)[name = tensor("transpose_124")]; tensor matrix_ac = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_118, y = transpose_119)[name = tensor("matrix_ac")]; tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_93)[name = tensor("matrix_bd")]; tensor var_4306 = add(x = matrix_ac, y = matrix_bd)[name = tensor("op_4306")]; tensor _inversed_scores_93_y_0 = const()[name = tensor("_inversed_scores_93_y_0"), val = tensor(0x1.6a09e6p-4)]; tensor _inversed_scores_93 = mul(x = var_4306, y = _inversed_scores_93_y_0)[name = tensor("_inversed_scores_93")]; tensor scores = select(a = var_14, b = _inversed_scores_93, cond = mask_3)[name = tensor("scores")]; tensor var_4312 = softmax(axis = var_32, x = scores)[name = tensor("op_4312")]; tensor input_1229 = select(a = var_13, b = var_4312, cond = mask_3)[name = tensor("input_1229")]; tensor x_521_transpose_x_0 = const()[name = tensor("x_521_transpose_x_0"), val = tensor(false)]; tensor x_521_transpose_y_0 = const()[name = tensor("x_521_transpose_y_0"), val = tensor(false)]; tensor value = transpose(perm = value_perm_0, x = v)[name = tensor("transpose_126")]; tensor x_521 = matmul(transpose_x = x_521_transpose_x_0, transpose_y = x_521_transpose_y_0, x = input_1229, y = value)[name = tensor("x_521")]; tensor var_4316_perm_0 = const()[name = tensor("op_4316_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_4317 = const()[name = tensor("op_4317"), val = tensor([1, -1, 1024])]; tensor var_4316 = transpose(perm = var_4316_perm_0, x = x_521)[name = tensor("transpose_122")]; tensor input_1231 = reshape(shape = var_4317, x = var_4316)[name = tensor("input_1231")]; tensor input_1233 = linear(bias = encoder_layers_23_self_attn_linear_out_bias, weight = encoder_layers_23_self_attn_linear_out_weight, x = input_1231)[name = tensor("linear_214")]; tensor input_1235 = add(x = input_1227, y = input_1233)[name = tensor("input_1235")]; tensor x_525_axes_0 = const()[name = tensor("x_525_axes_0"), val = tensor([-1])]; tensor x_525 = layer_norm(axes = x_525_axes_0, beta = encoder_layers_23_norm_conv_bias, epsilon = var_11, gamma = encoder_layers_23_norm_conv_weight, x = input_1235)[name = tensor("x_525")]; tensor input_1237_perm_0 = const()[name = tensor("input_1237_perm_0"), val = tensor([0, 2, 1])]; tensor input_1239_pad_type_0 = const()[name = tensor("input_1239_pad_type_0"), val = tensor("valid")]; tensor input_1239_strides_0 = const()[name = tensor("input_1239_strides_0"), val = tensor([1])]; tensor input_1239_pad_0 = const()[name = tensor("input_1239_pad_0"), val = tensor([0, 0])]; tensor input_1239_dilations_0 = const()[name = tensor("input_1239_dilations_0"), val = tensor([1])]; tensor input_1239_groups_0 = const()[name = tensor("input_1239_groups_0"), val = tensor(1)]; tensor input_1237 = transpose(perm = input_1237_perm_0, x = x_525)[name = tensor("transpose_121")]; tensor input_1239 = conv(bias = encoder_layers_23_conv_pointwise_conv1_bias, dilations = input_1239_dilations_0, groups = input_1239_groups_0, pad = input_1239_pad_0, pad_type = input_1239_pad_type_0, strides = input_1239_strides_0, weight = encoder_layers_23_conv_pointwise_conv1_weight, x = input_1237)[name = tensor("input_1239")]; tensor x_527_split_num_splits_0 = const()[name = tensor("x_527_split_num_splits_0"), val = tensor(2)]; tensor x_527_split_axis_0 = const()[name = tensor("x_527_split_axis_0"), val = tensor(1)]; tensor x_527_split_0, tensor x_527_split_1 = split(axis = x_527_split_axis_0, num_splits = x_527_split_num_splits_0, x = input_1239)[name = tensor("x_527_split")]; tensor x_527_split_1_sigmoid = sigmoid(x = x_527_split_1)[name = tensor("x_527_split_1_sigmoid")]; tensor x_527 = mul(x = x_527_split_0, y = x_527_split_1_sigmoid)[name = tensor("x_527")]; tensor input_1241 = select(a = var_13, b = x_527, cond = var_339)[name = tensor("input_1241")]; tensor const_247 = const()[name = tensor("const_247"), val = tensor(0x0p+0)]; tensor input_1243_pad_0 = const()[name = tensor("input_1243_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_1243_mode_0 = const()[name = tensor("input_1243_mode_0"), val = tensor("constant")]; tensor input_1243 = pad(constant_val = const_247, mode = input_1243_mode_0, pad = input_1243_pad_0, x = input_1241)[name = tensor("input_1243")]; tensor input_1245_pad_type_0 = const()[name = tensor("input_1245_pad_type_0"), val = tensor("valid")]; tensor input_1245_groups_0 = const()[name = tensor("input_1245_groups_0"), val = tensor(1024)]; tensor input_1245_strides_0 = const()[name = tensor("input_1245_strides_0"), val = tensor([1])]; tensor input_1245_pad_0 = const()[name = tensor("input_1245_pad_0"), val = tensor([0, 0])]; tensor input_1245_dilations_0 = const()[name = tensor("input_1245_dilations_0"), val = tensor([1])]; tensor const_294 = const()[name = tensor("const_294"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2371216768)))]; tensor const_295 = const()[name = tensor("const_295"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2371253696)))]; tensor input_1247 = conv(bias = const_295, dilations = input_1245_dilations_0, groups = input_1245_groups_0, pad = input_1245_pad_0, pad_type = input_1245_pad_type_0, strides = input_1245_strides_0, weight = const_294, x = input_1243)[name = tensor("input_1247")]; tensor input_1249 = silu(x = input_1247)[name = tensor("input_1249")]; tensor x_529_pad_type_0 = const()[name = tensor("x_529_pad_type_0"), val = tensor("valid")]; tensor x_529_strides_0 = const()[name = tensor("x_529_strides_0"), val = tensor([1])]; tensor x_529_pad_0 = const()[name = tensor("x_529_pad_0"), val = tensor([0, 0])]; tensor x_529_dilations_0 = const()[name = tensor("x_529_dilations_0"), val = tensor([1])]; tensor x_529_groups_0 = const()[name = tensor("x_529_groups_0"), val = tensor(1)]; tensor x_529 = conv(bias = encoder_layers_23_conv_pointwise_conv2_bias, dilations = x_529_dilations_0, groups = x_529_groups_0, pad = x_529_pad_0, pad_type = x_529_pad_type_0, strides = x_529_strides_0, weight = encoder_layers_23_conv_pointwise_conv2_weight, x = input_1249)[name = tensor("x_529")]; tensor input_1251_perm_0 = const()[name = tensor("input_1251_perm_0"), val = tensor([0, 2, 1])]; tensor input_1251 = transpose(perm = input_1251_perm_0, x = x_529)[name = tensor("transpose_120")]; tensor input_1253 = add(x = input_1235, y = input_1251)[name = tensor("input_1253")]; tensor input_1255_axes_0 = const()[name = tensor("input_1255_axes_0"), val = tensor([-1])]; tensor input_1255 = layer_norm(axes = input_1255_axes_0, beta = encoder_layers_23_norm_feed_forward2_bias, epsilon = var_11, gamma = encoder_layers_23_norm_feed_forward2_weight, x = input_1253)[name = tensor("input_1255")]; tensor input_1257 = linear(bias = encoder_layers_23_feed_forward2_linear1_bias, weight = encoder_layers_23_feed_forward2_linear1_weight, x = input_1255)[name = tensor("linear_215")]; tensor input_1259 = silu(x = input_1257)[name = tensor("input_1259")]; tensor input_1263 = linear(bias = encoder_layers_23_feed_forward2_linear2_bias, weight = encoder_layers_23_feed_forward2_linear2_weight, x = input_1259)[name = tensor("linear_216")]; tensor var_4383 = const()[name = tensor("op_4383"), val = tensor(0x1p-1)]; tensor var_4384 = mul(x = input_1263, y = var_4383)[name = tensor("op_4384")]; tensor input = add(x = input_1253, y = var_4384)[name = tensor("input")]; tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; tensor audio_signal = layer_norm(axes = audio_signal_axes_0, beta = encoder_layers_23_norm_out_bias, epsilon = var_11, gamma = encoder_layers_23_norm_out_weight, x = input)[name = tensor("audio_signal")]; tensor ctc_head_raw_output = linear(bias = proj_bias, weight = proj_weight, x = audio_signal)[name = tensor("linear_217")]; } -> (ctc_head_raw_output, encoder_length); }